1
|
Tong R, Su S, Liang Y, Li C, Sun L, Zhang X. Functional Connectivity Encodes Sound Locations by Lateralization Angles. Neurosci Bull 2025; 41:261-271. [PMID: 39470972 PMCID: PMC11794782 DOI: 10.1007/s12264-024-01312-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2024] [Accepted: 06/16/2024] [Indexed: 11/01/2024] Open
Abstract
The ability to localize sound sources rapidly allows human beings to efficiently understand the surrounding environment. Previous studies have suggested that there is an auditory "where" pathway in the cortex for processing sound locations. The neural activation in regions along this pathway encodes sound locations by opponent hemifield coding, in which each unilateral region is activated by sounds coming from the contralateral hemifield. However, it is still unclear how these regions interact with each other to form a unified representation of the auditory space. In the present study, we investigated whether functional connectivity in the auditory "where" pathway encoded sound locations during passive listening. Participants underwent functional magnetic resonance imaging while passively listening to sounds from five distinct horizontal locations (-90°, -45°, 0°, 45°, 90°). We were able to decode sound locations from the functional connectivity patterns of the "where" pathway. Furthermore, we found that such neural representation of sound locations was primarily based on the coding of sound lateralization angles to the frontal midline. In addition, whole-brain analysis indicated that functional connectivity between occipital regions and the primary auditory cortex also encoded sound locations by lateralization angles. Overall, our results reveal a lateralization-angle-based representation of sound locations encoded by functional connectivity patterns, which could add on the activation-based opponent hemifield coding to provide a more precise representation of the auditory space.
Collapse
Affiliation(s)
- Renjie Tong
- School of Biomedical Engineering, Capital Medical University, Beijing, 100069, China
- Beijing Key Laboratory of Fundamental Research on Biomechanics in Clinical Application, Beijing, 100069, China
| | - Shaoyi Su
- Institute of Medical Technology, Peking University Health Science Center, Beijing, 100191, China
| | - Ying Liang
- School of Biomedical Engineering, Capital Medical University, Beijing, 100069, China
- Beijing Key Laboratory of Fundamental Research on Biomechanics in Clinical Application, Beijing, 100069, China
| | - Chunlin Li
- School of Biomedical Engineering, Capital Medical University, Beijing, 100069, China
- Beijing Key Laboratory of Fundamental Research on Biomechanics in Clinical Application, Beijing, 100069, China
| | - Liwei Sun
- School of Biomedical Engineering, Capital Medical University, Beijing, 100069, China.
- Beijing Key Laboratory of Fundamental Research on Biomechanics in Clinical Application, Beijing, 100069, China.
| | - Xu Zhang
- School of Biomedical Engineering, Capital Medical University, Beijing, 100069, China.
- Beijing Key Laboratory of Fundamental Research on Biomechanics in Clinical Application, Beijing, 100069, China.
| |
Collapse
|
2
|
McMullin MA, Kumar R, Higgins NC, Gygi B, Elhilali M, Snyder JS. Preliminary Evidence for Global Properties in Human Listeners During Natural Auditory Scene Perception. Open Mind (Camb) 2024; 8:333-365. [PMID: 38571530 PMCID: PMC10990578 DOI: 10.1162/opmi_a_00131] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2023] [Accepted: 02/10/2024] [Indexed: 04/05/2024] Open
Abstract
Theories of auditory and visual scene analysis suggest the perception of scenes relies on the identification and segregation of objects within it, resembling a detail-oriented processing style. However, a more global process may occur while analyzing scenes, which has been evidenced in the visual domain. It is our understanding that a similar line of research has not been explored in the auditory domain; therefore, we evaluated the contributions of high-level global and low-level acoustic information to auditory scene perception. An additional aim was to increase the field's ecological validity by using and making available a new collection of high-quality auditory scenes. Participants rated scenes on 8 global properties (e.g., open vs. enclosed) and an acoustic analysis evaluated which low-level features predicted the ratings. We submitted the acoustic measures and average ratings of the global properties to separate exploratory factor analyses (EFAs). The EFA of the acoustic measures revealed a seven-factor structure explaining 57% of the variance in the data, while the EFA of the global property measures revealed a two-factor structure explaining 64% of the variance in the data. Regression analyses revealed each global property was predicted by at least one acoustic variable (R2 = 0.33-0.87). These findings were extended using deep neural network models where we examined correlations between human ratings of global properties and deep embeddings of two computational models: an object-based model and a scene-based model. The results support that participants' ratings are more strongly explained by a global analysis of the scene setting, though the relationship between scene perception and auditory perception is multifaceted, with differing correlation patterns evident between the two models. Taken together, our results provide evidence for the ability to perceive auditory scenes from a global perspective. Some of the acoustic measures predicted ratings of global scene perception, suggesting representations of auditory objects may be transformed through many stages of processing in the ventral auditory stream, similar to what has been proposed in the ventral visual stream. These findings and the open availability of our scene collection will make future studies on perception, attention, and memory for natural auditory scenes possible.
Collapse
Affiliation(s)
| | - Rohit Kumar
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD, USA
| | - Nathan C. Higgins
- Department of Communication Sciences & Disorders, University of South Florida, Tampa, FL, USA
| | - Brian Gygi
- East Bay Institute for Research and Education, Martinez, CA, USA
| | - Mounya Elhilali
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD, USA
| | - Joel S. Snyder
- Department of Psychology, University of Nevada, Las Vegas, Las Vegas, NV, USA
| |
Collapse
|
3
|
Wu YY, Hu YS, Wang J, Zang YF, Zhang Y. Toward Precise Localization of Abnormal Brain Activity: 1D CNN on Single Voxel fMRI Time-Series. Front Comput Neurosci 2022; 16:822237. [PMID: 35573265 PMCID: PMC9094401 DOI: 10.3389/fncom.2022.822237] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2021] [Accepted: 03/31/2022] [Indexed: 11/13/2022] Open
Abstract
Functional magnetic resonance imaging (fMRI) is one of the best techniques for precise localization of abnormal brain activity non-invasively. Machine-learning approaches have been widely used in neuroimaging studies; however, few studies have investigated the single-voxel modeling of fMRI data under cognitive tasks. We proposed a hybrid one-dimensional (1D) convolutional neural network (1D-CNN) based on the temporal dynamics of single-voxel fMRI time-series and successfully differentiated two continuous task states, namely, self-initiated (SI) and visually guided (VG) motor tasks. First, 25 activation peaks were identified from the contrast maps of SI and VG tasks in a blocked design. Then, the fMRI time-series of each peak voxel was transformed into a temporal-frequency domain by using continuous wavelet transform across a broader frequency range (0.003–0.313 Hz, with a step of 0.01 Hz). The transformed time-series was inputted into a 1D-CNN model for the binary classification of SI and VG continuous tasks. Compared with the univariate analysis, e.g., amplitude of low-frequency fluctuation (ALFF) at each frequency band, including, wavelet-ALFF, the 1D-CNN model highly outperformed wavelet-ALFF, with more efficient decoding models [46% of 800 models showing area under the curve (AUC) > 0.61] and higher decoding accuracies (94% of the efficient models), especially on the high-frequency bands (>0.1 Hz). Moreover, our results also demonstrated the advantages of wavelet decompositions over the original fMRI series by showing higher decoding performance on all peak voxels. Overall, this study suggests a great potential of single-voxel analysis using 1D-CNN and wavelet transformation of fMRI series with continuous, naturalistic, steady-state task design or resting-state design. It opens new avenues to precise localization of abnormal brain activity and fMRI-guided precision brain stimulation therapy.
Collapse
Affiliation(s)
- Yun-Ying Wu
- Center for Cognition and Brain Disorders and the Affiliated Hospital, Hangzhou Normal University, Hangzhou, China
- Zhejiang Key Laboratory for Research in Assessment of Cognitive Impairments, Hangzhou, China
- Institutes of Psychological Sciences, Hangzhou Normal University, Hangzhou, China
| | - Yun-Song Hu
- Center for Cognition and Brain Disorders and the Affiliated Hospital, Hangzhou Normal University, Hangzhou, China
- Zhejiang Key Laboratory for Research in Assessment of Cognitive Impairments, Hangzhou, China
- Institutes of Psychological Sciences, Hangzhou Normal University, Hangzhou, China
| | - Jue Wang
- Institute of Sports Medicine and Health, Chengdu Sport University, Chengdu, China
| | - Yu-Feng Zang
- Center for Cognition and Brain Disorders and the Affiliated Hospital, Hangzhou Normal University, Hangzhou, China
- Zhejiang Key Laboratory for Research in Assessment of Cognitive Impairments, Hangzhou, China
- Institutes of Psychological Sciences, Hangzhou Normal University, Hangzhou, China
- Transcranial Magnetic Stimulation Center, Deqing Hospital of Hangzhou Normal University, Huzhou, China
- *Correspondence: Yu-Feng Zang
| | - Yu Zhang
- Research Center for Healthcare Data Science, Zhejiang Lab, Hangzhou, China
- Yu Zhang
| |
Collapse
|
4
|
Liang Y, Liu B, Ji J, Li X. Network Representations of Facial and Bodily Expressions: Evidence From Multivariate Connectivity Pattern Classification. Front Neurosci 2019; 13:1111. [PMID: 31736683 PMCID: PMC6828617 DOI: 10.3389/fnins.2019.01111] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2019] [Accepted: 10/02/2019] [Indexed: 01/21/2023] Open
Abstract
Emotions can be perceived from both facial and bodily expressions. Our previous study has found the successful decoding of facial expressions based on the functional connectivity (FC) patterns. However, the role of the FC patterns in the recognition of bodily expressions remained unclear, and no neuroimaging studies have adequately addressed the question of whether emotions perceiving from facial and bodily expressions are processed rely upon common or different neural networks. To address this, the present study collected functional magnetic resonance imaging (fMRI) data from a block design experiment with facial and bodily expression videos as stimuli (three emotions: anger, fear, and joy), and conducted multivariate pattern classification analysis based on the estimated FC patterns. We found that in addition to the facial expressions, bodily expressions could also be successfully decoded based on the large-scale FC patterns. The emotion classification accuracies for the facial expressions were higher than that for the bodily expressions. Further contributive FC analysis showed that emotion-discriminative networks were widely distributed in both hemispheres, containing regions that ranged from primary visual areas to higher-level cognitive areas. Moreover, for a particular emotion, discriminative FCs for facial and bodily expressions were distinct. Together, our findings highlight the key role of the FC patterns in the emotion processing, indicating how large-scale FC patterns reconfigure in processing of facial and bodily expressions, and suggest the distributed neural representation for the emotion recognition. Furthermore, our results also suggest that the human brain employs separate network representations for facial and bodily expressions of the same emotions. This study provides new evidence for the network representations for emotion perception and may further our understanding of the potential mechanisms underlying body language emotion recognition.
Collapse
Affiliation(s)
- Yin Liang
- Faculty of Information Technology, Beijing Artificial Intelligence Institute, Beijing University of Technology, Beijing, China
| | - Baolin Liu
- Tianjin Key Laboratory of Cognitive Computing and Application, School of Computer Science and Technology, Tianjin University, Tianjin, China.,School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, China.,State Key Laboratory of Intelligent Technology and Systems, National Laboratory for Information Science and Technology, Tsinghua University, Beijing, China
| | - Junzhong Ji
- Faculty of Information Technology, Beijing Artificial Intelligence Institute, Beijing University of Technology, Beijing, China
| | - Xianglin Li
- Medical Imaging Research Institute, Binzhou Medical University, Yantai, China
| |
Collapse
|