1
|
Hakonen M, Dahmani L, Lankinen K, Ren J, Barbaro J, Blazejewska A, Cui W, Kotlarz P, Li M, Polimeni JR, Turpin T, Uluç I, Wang D, Liu H, Ahveninen J. Individual connectivity-based parcellations reflect functional properties of human auditory cortex. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.01.20.576475. [PMID: 38293021 PMCID: PMC10827228 DOI: 10.1101/2024.01.20.576475] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/01/2024]
Abstract
Neuroimaging studies of the functional organization of human auditory cortex have focused on group-level analyses to identify tendencies that represent the typical brain. Here, we mapped auditory areas of the human superior temporal cortex (STC) in 30 participants by combining functional network analysis and 1-mm isotropic resolution 7T functional magnetic resonance imaging (fMRI). Two resting-state fMRI sessions, and one or two auditory and audiovisual speech localizer sessions, were collected on 3-4 separate days. We generated a set of functional network-based parcellations from these data. Solutions with 4, 6, and 11 networks were selected for closer examination based on local maxima of Dice and Silhouette values. The resulting parcellation of auditory cortices showed high intraindividual reproducibility both between resting state sessions (Dice coefficient: 69-78%) and between resting state and task sessions (Dice coefficient: 62-73%). This demonstrates that auditory areas in STC can be reliably segmented into functional subareas. The interindividual variability was significantly larger than intraindividual variability (Dice coefficient: 57%-68%, p<0.001), indicating that the parcellations also captured meaningful interindividual variability. The individual-specific parcellations yielded the highest alignment with task response topographies, suggesting that individual variability in parcellations reflects individual variability in auditory function. Connectional homogeneity within networks was also highest for the individual-specific parcellations. Furthermore, the similarity in the functional parcellations was not explainable by the similarity of macroanatomical properties of auditory cortex. Our findings suggest that individual-level parcellations capture meaningful idiosyncrasies in auditory cortex organization.
Collapse
Affiliation(s)
- M Hakonen
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital Charlestown, MA, USA
- Department of Radiology, Harvard Medical School, Boston, MA, USA
| | - L Dahmani
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital Charlestown, MA, USA
- Department of Radiology, Harvard Medical School, Boston, MA, USA
| | - K Lankinen
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital Charlestown, MA, USA
- Department of Radiology, Harvard Medical School, Boston, MA, USA
| | - J Ren
- Division of Brain Sciences, Changping Laboratory, Beijing, China
| | - J Barbaro
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital Charlestown, MA, USA
| | - A Blazejewska
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital Charlestown, MA, USA
- Department of Radiology, Harvard Medical School, Boston, MA, USA
| | - W Cui
- Division of Brain Sciences, Changping Laboratory, Beijing, China
| | - P Kotlarz
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital Charlestown, MA, USA
| | - M Li
- Division of Brain Sciences, Changping Laboratory, Beijing, China
| | - J R Polimeni
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital Charlestown, MA, USA
- Department of Radiology, Harvard Medical School, Boston, MA, USA
- Harvard-MIT Program in Health Sciences and Technology, Massachusetts Institute of Technology, Cambridge, MA, USA
| | - T Turpin
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital Charlestown, MA, USA
| | - I Uluç
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital Charlestown, MA, USA
- Department of Radiology, Harvard Medical School, Boston, MA, USA
| | - D Wang
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital Charlestown, MA, USA
- Department of Radiology, Harvard Medical School, Boston, MA, USA
| | - H Liu
- Division of Brain Sciences, Changping Laboratory, Beijing, China
- Biomedical Pioneering Innovation Center (BIOPIC), Peking University, Beijing, China
| | - J Ahveninen
- Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital Charlestown, MA, USA
- Department of Radiology, Harvard Medical School, Boston, MA, USA
| |
Collapse
|
2
|
Zhang Y, Shen SX, Bibic A, Wang X. Evolutionary continuity and divergence of auditory dorsal and ventral pathways in primates revealed by ultra-high field diffusion MRI. Proc Natl Acad Sci U S A 2024; 121:e2313831121. [PMID: 38377216 PMCID: PMC10907247 DOI: 10.1073/pnas.2313831121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2023] [Accepted: 01/22/2024] [Indexed: 02/22/2024] Open
Abstract
Auditory dorsal and ventral pathways in the human brain play important roles in supporting speech and language processing. However, the evolutionary root of the dual auditory pathways in the primate brain is unclear. By parcellating the auditory cortex of marmosets (a New World monkey species), macaques (an Old World monkey species), and humans using the same individual-based analysis method and tracking the pathways from the auditory cortex based on multi-shell diffusion-weighted MRI (dMRI), homologous auditory dorsal and ventral fiber tracks were identified in these primate species. The ventral pathway was found to be well conserved in all three primate species analyzed but extend to more anterior temporal regions in humans. In contrast, the dorsal pathway showed a divergence between monkey and human brains. First, frontal regions in the human brain have stronger connections to the higher-level auditory regions than to the lower-level auditory regions along the dorsal pathway, while frontal regions in the monkey brain show opposite connection patterns along the dorsal pathway. Second, the left lateralization of the dorsal pathway is only found in humans. Moreover, the connectivity strength of the dorsal pathway in marmosets is more similar to that of humans than macaques. These results demonstrate the continuity and divergence of the dual auditory pathways in the primate brains along the evolutionary path, suggesting that the putative neural networks supporting human speech and language processing might have emerged early in primate evolution.
Collapse
Affiliation(s)
- Yang Zhang
- Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD21205
| | - Sherry Xinyi Shen
- Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD21205
| | - Adnan Bibic
- Department of Radiology, Johns Hopkins University School of Medicine, Baltimore, MD21205
- Kirby Research Center for Functional Brain Imaging, Kennedy Krieger Institute, F. M. Kirby Center, Baltimore, MD21205
| | - Xiaoqin Wang
- Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD21205
| |
Collapse
|
3
|
Bianco R, Zuk NJ, Bigand F, Quarta E, Grasso S, Arnese F, Ravignani A, Battaglia-Mayer A, Novembre G. Neural encoding of musical expectations in a non-human primate. Curr Biol 2024; 34:444-450.e5. [PMID: 38176416 DOI: 10.1016/j.cub.2023.12.019] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2023] [Revised: 10/26/2023] [Accepted: 12/07/2023] [Indexed: 01/06/2024]
Abstract
The appreciation of music is a universal trait of humankind.1,2,3 Evidence supporting this notion includes the ubiquity of music across cultures4,5,6,7 and the natural predisposition toward music that humans display early in development.8,9,10 Are we musical animals because of species-specific predispositions? This question cannot be answered by relying on cross-cultural or developmental studies alone, as these cannot rule out enculturation.11 Instead, it calls for cross-species experiments testing whether homologous neural mechanisms underlying music perception are present in non-human primates. We present music to two rhesus monkeys, reared without musical exposure, while recording electroencephalography (EEG) and pupillometry. Monkeys exhibit higher engagement and neural encoding of expectations based on the previously seeded musical context when passively listening to real music as opposed to shuffled controls. We then compare human and monkey neural responses to the same stimuli and find a species-dependent contribution of two fundamental musical features-pitch and timing12-in generating expectations: while timing- and pitch-based expectations13 are similarly weighted in humans, monkeys rely on timing rather than pitch. Together, these results shed light on the phylogeny of music perception. They highlight monkeys' capacity for processing temporal structures beyond plain acoustic processing, and they identify a species-dependent contribution of time- and pitch-related features to the neural encoding of musical expectations.
Collapse
Affiliation(s)
- Roberta Bianco
- Neuroscience of Perception & Action Lab, Italian Institute of Technology, Viale Regina Elena 291, 00161 Rome, Italy.
| | - Nathaniel J Zuk
- Department of Psychology, Nottingham Trent University, 50 Shakespeare Street, Nottingham NG1 4FQ, UK
| | - Félix Bigand
- Neuroscience of Perception & Action Lab, Italian Institute of Technology, Viale Regina Elena 291, 00161 Rome, Italy
| | - Eros Quarta
- Department of Physiology and Pharmacology, Sapienza University of Rome, Piazzale Aldo Moro 5, 00185 Rome, Italy
| | - Stefano Grasso
- Department of Physiology and Pharmacology, Sapienza University of Rome, Piazzale Aldo Moro 5, 00185 Rome, Italy
| | - Flavia Arnese
- Neuroscience of Perception & Action Lab, Italian Institute of Technology, Viale Regina Elena 291, 00161 Rome, Italy
| | - Andrea Ravignani
- Comparative Bioacoustics Group, Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525 XD Nijmegen, the Netherlands; Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music, Universitetsbyen 3, 8000 Aarhus, Denmark; Department of Human Neurosciences, Sapienza University of Rome, Piazzale Aldo Moro 5, 00185 Rome, Italy
| | - Alexandra Battaglia-Mayer
- Department of Physiology and Pharmacology, Sapienza University of Rome, Piazzale Aldo Moro 5, 00185 Rome, Italy
| | - Giacomo Novembre
- Neuroscience of Perception & Action Lab, Italian Institute of Technology, Viale Regina Elena 291, 00161 Rome, Italy.
| |
Collapse
|
4
|
Kim G, Kim DK, Jeong H. Spontaneous emergence of rudimentary music detectors in deep neural networks. Nat Commun 2024; 15:148. [PMID: 38168097 PMCID: PMC10761941 DOI: 10.1038/s41467-023-44516-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2021] [Accepted: 12/15/2023] [Indexed: 01/05/2024] Open
Abstract
Music exists in almost every society, has universal acoustic features, and is processed by distinct neural circuits in humans even with no experience of musical training. However, it remains unclear how these innate characteristics emerge and what functions they serve. Here, using an artificial deep neural network that models the auditory information processing of the brain, we show that units tuned to music can spontaneously emerge by learning natural sound detection, even without learning music. The music-selective units encoded the temporal structure of music in multiple timescales, following the population-level response characteristics observed in the brain. We found that the process of generalization is critical for the emergence of music-selectivity and that music-selectivity can work as a functional basis for the generalization of natural sound, thereby elucidating its origin. These findings suggest that evolutionary adaptation to process natural sounds can provide an initial blueprint for our sense of music.
Collapse
Affiliation(s)
- Gwangsu Kim
- Department of Physics, Korea Advanced Institute of Science and Technology, Daejeon, 34141, Korea
| | - Dong-Kyum Kim
- Department of Physics, Korea Advanced Institute of Science and Technology, Daejeon, 34141, Korea
| | - Hawoong Jeong
- Department of Physics, Korea Advanced Institute of Science and Technology, Daejeon, 34141, Korea.
- Center for Complex Systems, Korea Advanced Institute of Science and Technology, Daejeon, 34141, Korea.
| |
Collapse
|
5
|
Brattico E, Delussi M. Making sense of music: Insights from neurophysiology and connectivity analyses in naturalistic listening conditions. Hear Res 2024; 441:108923. [PMID: 38091866 DOI: 10.1016/j.heares.2023.108923] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/01/2023] [Revised: 11/08/2023] [Accepted: 11/22/2023] [Indexed: 12/27/2023]
Abstract
According to the latest frameworks, auditory perception and memory involve the constant prediction of future sound events by the brain, based on the continuous extraction of feature regularities from the environment. The neural hierarchical mechanisms for predictive processes in perception and memory for sounds are typically studied in relation to simple acoustic features in isolated sounds or sound patterns inserted in highly certain contexts. Such studies have identified reliable prediction formation and error signals, e.g., the N100 or the mismatch negativity (MMN) evoked responses. In real life, though, individuals often face situations in which uncertainty prevails and where making sense of sounds becomes a hard challenge. In music, not only deviations from predictions are masterly set up by composers to induce emotions but sometimes the sheer uncertainty of sound scenes is exploited for aesthetic purposes, especially in compositional styles such as Western atonal classical music. In very recent magnetoencephalography (MEG) and electroencephalography (EEG) studies, experimental and technical advances in stimulation paradigms and analysis approaches have permitted the identification of prediction-error responses from highly uncertain, atonal contexts and the extraction of prediction-related responses from real, continuous music. Moreover, functional connectivity analyses revealed the emergence of cortico-hippocampal interactions during the formation of auditory memories for more predictable vs. less predictable patterns. These findings contribute to understanding the general brain mechanisms that enable us to predict even highly uncertain sound environments and to possibly make sense of and appreciate even atonal music.
Collapse
Affiliation(s)
- Elvira Brattico
- Centre for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg, Aarhus, Denmark; Department of Education, Psychology, Communication, University of Bari Aldo Moro, Italy.
| | - Marianna Delussi
- Department of Education, Psychology, Communication, University of Bari Aldo Moro, Italy
| |
Collapse
|
6
|
Harris I, Niven EC, Griffin A, Scott SK. Is song processing distinct and special in the auditory cortex? Nat Rev Neurosci 2023; 24:711-722. [PMID: 37783820 DOI: 10.1038/s41583-023-00743-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 08/30/2023] [Indexed: 10/04/2023]
Abstract
Is the singing voice processed distinctively in the human brain? In this Perspective, we discuss what might distinguish song processing from speech processing in light of recent work suggesting that some cortical neuronal populations respond selectively to song and we outline the implications for our understanding of auditory processing. We review the literature regarding the neural and physiological mechanisms of song production and perception and show that this provides evidence for key differences between song and speech processing. We conclude by discussing the significance of the notion that song processing is special in terms of how this might contribute to theories of the neurobiological origins of vocal communication and to our understanding of the neural circuitry underlying sound processing in the human cortex.
Collapse
Affiliation(s)
- Ilana Harris
- Institute of Cognitive Neuroscience, University College London, London, UK
| | - Efe C Niven
- Institute of Cognitive Neuroscience, University College London, London, UK
| | - Alex Griffin
- Department of Psychology, University of Cambridge, Cambridge, UK
| | - Sophie K Scott
- Institute of Cognitive Neuroscience, University College London, London, UK.
| |
Collapse
|
7
|
Berger JI, Gander PE, Kikuchi Y, Petkov CI, Kumar S, Kovach C, Oya H, Kawasaki H, Howard MA, Griffiths TD. Distribution of multiunit pitch responses recorded intracranially from human auditory cortex. Cereb Cortex 2023:7180374. [PMID: 37246155 DOI: 10.1093/cercor/bhad186] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2022] [Revised: 05/05/2023] [Accepted: 05/06/2023] [Indexed: 05/30/2023] Open
Abstract
The perception of pitch is a fundamental percept, which is mediated by the auditory system, requiring the abstraction of stimulus properties related to the spectro-temporal structure of sound. Despite its importance, there is still debate as to the precise areas responsible for its encoding, which may be due to species differences or differences in the recording measures and choices of stimuli used in previous studies. Moreover, it was unknown whether the human brain contains pitch neurons and how distributed such neurons might be. Here, we present the first study to measure multiunit neural activity in response to pitch stimuli in the auditory cortex of intracranially implanted humans. The stimulus sets were regular-interval noise with a pitch strength that is related to the temporal regularity and a pitch value determined by the repetition rate and harmonic complexes. Specifically, we demonstrate reliable responses to these different pitch-inducing paradigms that are distributed throughout Heschl's gyrus, rather than being localized to a particular region, and this finding was evident regardless of the stimulus presented. These data provide a bridge across animal and human studies and aid our understanding of the processing of a critical percept associated with acoustic stimuli.
Collapse
Affiliation(s)
- Joel I Berger
- Department of Neurosurgery, 1800 JPP, 200 Hawkins Drive, University of Iowa, Iowa City, IA 52242, United States
| | - Phillip E Gander
- Department of Neurosurgery, 1800 JPP, 200 Hawkins Drive, University of Iowa, Iowa City, IA 52242, United States
| | - Yukiko Kikuchi
- Biosciences Institute, Newcastle upon Tyne NE2 4HH, United Kingdom
| | - Christopher I Petkov
- Department of Neurosurgery, 1800 JPP, 200 Hawkins Drive, University of Iowa, Iowa City, IA 52242, United States
- Biosciences Institute, Newcastle upon Tyne NE2 4HH, United Kingdom
| | - Sukhbinder Kumar
- Department of Neurosurgery, 1800 JPP, 200 Hawkins Drive, University of Iowa, Iowa City, IA 52242, United States
| | - Christopher Kovach
- Department of Neurosurgery, 1800 JPP, 200 Hawkins Drive, University of Iowa, Iowa City, IA 52242, United States
| | - Hiroyuki Oya
- Department of Neurosurgery, 1800 JPP, 200 Hawkins Drive, University of Iowa, Iowa City, IA 52242, United States
| | - Hiroto Kawasaki
- Department of Neurosurgery, 1800 JPP, 200 Hawkins Drive, University of Iowa, Iowa City, IA 52242, United States
| | - Matthew A Howard
- Department of Neurosurgery, 1800 JPP, 200 Hawkins Drive, University of Iowa, Iowa City, IA 52242, United States
| | | |
Collapse
|
8
|
Papatzikis E, Agapaki M, Selvan RN, Pandey V, Zeba F. Quality standards and recommendations for research in music and neuroplasticity. Ann N Y Acad Sci 2023; 1520:20-33. [PMID: 36478395 DOI: 10.1111/nyas.14944] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/12/2022]
Abstract
Research on how music influences brain plasticity has gained momentum in recent years. Considering, however, the nonuniform methodological standards implemented, the findings end up being nonreplicable and less generalizable. To address the need for a standardized baseline of research quality, we gathered all the studies in the music and neuroplasticity field in 2019 and appraised their methodological rigor systematically and critically. The aim was to provide a preliminary and, at the minimum, acceptable quality threshold-and, ipso facto, suggested recommendations-whereupon further discussion and development may take place. Quality appraisal was performed on 89 articles by three independent raters, following a standardized scoring system. The raters' scoring was cross-referenced following an inter-rater reliability measure, and further studied by performing multiple ratings comparisons and matrix analyses. The results for methodological quality were at a quite good level (quantitative articles: mean = 0.737, SD = 0.084; qualitative articles: mean = 0.677, SD = 0.144), following a moderate but statistically significant level of agreement between the raters (W = 0.44, χ2 = 117.249, p = 0.020). We conclude that the standards for implementation and reporting are of high quality; however, certain improvements are needed to reach the stringent levels presumed for such an influential interdisciplinary scientific field.
Collapse
Affiliation(s)
- Efthymios Papatzikis
- Department of Early Childhood Education and Care, Oslo Metropolitan University, Oslo, Norway
| | - Maria Agapaki
- Department of Early Childhood Education and Care, Oslo Metropolitan University, Oslo, Norway
| | - Rosari Naveena Selvan
- Institute for Physics 3 - Biophysics and Bernstein Center for Computational Neuroscience (BCCN), University of Göttingen, Göttingen, Germany.,Department of Psychology, University of Münster, Münster, Germany
| | | | - Fathima Zeba
- School of Humanities and Social Sciences, Manipal Academy of Higher Education Dubai, Dubai, United Arab Emirates
| |
Collapse
|
9
|
Basiński K, Quiroga-Martinez DR, Vuust P. Temporal hierarchies in the predictive processing of melody - From pure tones to songs. Neurosci Biobehav Rev 2023; 145:105007. [PMID: 36535375 DOI: 10.1016/j.neubiorev.2022.105007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2022] [Revised: 11/30/2022] [Accepted: 12/14/2022] [Indexed: 12/23/2022]
Abstract
Listening to musical melodies is a complex task that engages perceptual and memoryrelated processes. The processes underlying melody cognition happen simultaneously on different timescales, ranging from milliseconds to minutes. Although attempts have been made, research on melody perception is yet to produce a unified framework of how melody processing is achieved in the brain. This may in part be due to the difficulty of integrating concepts such as perception, attention and memory, which pertain to different temporal scales. Recent theories on brain processing, which hold prediction as a fundamental principle, offer potential solutions to this problem and may provide a unifying framework for explaining the neural processes that enable melody perception on multiple temporal levels. In this article, we review empirical evidence for predictive coding on the levels of pitch formation, basic pitch-related auditory patterns,more complex regularity processing extracted from basic patterns and long-term expectations related to musical syntax. We also identify areas that would benefit from further inquiry and suggest future directions in research on musical melody perception.
Collapse
Affiliation(s)
- Krzysztof Basiński
- Division of Quality of Life Research, Medical University of Gdańsk, Poland
| | - David Ricardo Quiroga-Martinez
- Helen Wills Neuroscience Institute & Department of Psychology, University of California Berkeley, USA; Center for Music in the Brain, Aarhus University & The Royal Academy of Music, Denmark
| | - Peter Vuust
- Center for Music in the Brain, Aarhus University & The Royal Academy of Music, Denmark
| |
Collapse
|
10
|
Axelrod V, Rozier C, Lehongre K, Adam C, Lambrecq V, Navarro V, Naccache L. Neural modulations in the auditory cortex during internal and external attention tasks: A single-patient intracranial recording study. Cortex 2022; 157:211-230. [PMID: 36335821 DOI: 10.1016/j.cortex.2022.09.011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2021] [Revised: 05/12/2022] [Accepted: 09/27/2022] [Indexed: 12/15/2022]
Abstract
Brain sensory processing is not passive, but is rather modulated by our internal state. Different research methods such as non-invasive imaging methods and intracranial recording of the local field potential (LFP) have been used to study to what extent sensory processing and the auditory cortex in particular are modulated by selective attention. However, at the level of the single- or multi-units the selective attention in humans has not been tested. In addition, most previous research on selective attention has explored externally-oriented attention, but attention can be also directed inward (i.e., internal attention), like spontaneous self-generated thoughts and mind-wandering. In the present study we had a rare opportunity to record multi-unit activity (MUA) in the auditory cortex of a patient. To complement, we also analyzed the LFP signal of the macro-contact in the auditory cortex. Our experiment consisted of two conditions with periodic beeping sounds. The participants were asked either to count the beeps (i.e., an "external attention" condition) or to recall the events of the previous day (i.e., an "internal attention" condition). We found that the four out of seven recorded units in the auditory cortex showed increased firing rates in "external attention" compared to "internal attention" condition. The beginning of this attentional modulation varied across multi-units between 30-50 msec and 130-150 msec from stimulus onset, a result that is compatible with an early selection view. The LFP evoked potential and induced high gamma activity both showed attentional modulation starting at about 70-80 msec. As the control, for the same experiment we recorded MUA activity in the amygdala and hippocampus of two additional patients. No major attentional modulation was found in the control regions. Overall, we believe that our results provide new empirical information and support for existing theoretical views on selective attention and spontaneous self-generated cognition.
Collapse
Affiliation(s)
- Vadim Axelrod
- The Gonda Multidisciplinary Brain Research Center, Bar Ilan University, Ramat Gan, Israel.
| | - Camille Rozier
- Sorbonne Université, Institut du Cerveau - Paris Brain Institute, ICM, INSERM U1127, CNRS UMR 7225, Paris, France
| | - Katia Lehongre
- Sorbonne Université, Institut du Cerveau - Paris Brain Institute, ICM, INSERM U1127, CNRS UMR 7225, Paris, France; Centre de NeuroImagerie de Recherche-CENIR, Paris Brain Institute, UMRS 1127, CNRS UMR 7225, Pitié-Salpêtriere Hospital, Paris, France
| | - Claude Adam
- AP-HP, GH Pitie-Salpêtrière-Charles Foix, Epilepsy Unit, Neurology Department, Paris, France
| | - Virginie Lambrecq
- Sorbonne Université, Institut du Cerveau - Paris Brain Institute, ICM, INSERM U1127, CNRS UMR 7225, Paris, France; AP-HP, Groupe hospitalier Pitié-Salpêtrière, Department of Neurophysiology, Paris, France; Sorbonne Université, UMR S1127, Paris, France
| | - Vincent Navarro
- Sorbonne Université, Institut du Cerveau - Paris Brain Institute, ICM, INSERM U1127, CNRS UMR 7225, Paris, France; AP-HP, GH Pitie-Salpêtrière-Charles Foix, Epilepsy Unit, Neurology Department, Paris, France; Sorbonne Université, UMR S1127, Paris, France
| | - Lionel Naccache
- Sorbonne Université, Institut du Cerveau - Paris Brain Institute, ICM, INSERM U1127, CNRS UMR 7225, Paris, France; AP-HP, Groupe hospitalier Pitié-Salpêtrière, Department of Neurophysiology, Paris, France
| |
Collapse
|
11
|
Fernández-Rubio G, Brattico E, Kotz SA, Kringelbach ML, Vuust P, Bonetti L. Magnetoencephalography recordings reveal the spatiotemporal dynamics of recognition memory for complex versus simple auditory sequences. Commun Biol 2022; 5:1272. [PMID: 36402843 PMCID: PMC9675809 DOI: 10.1038/s42003-022-04217-8] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2022] [Accepted: 11/02/2022] [Indexed: 11/21/2022] Open
Abstract
Auditory recognition is a crucial cognitive process that relies on the organization of single elements over time. However, little is known about the spatiotemporal dynamics underlying the conscious recognition of auditory sequences varying in complexity. To study this, we asked 71 participants to learn and recognize simple tonal musical sequences and matched complex atonal sequences while their brain activity was recorded using magnetoencephalography (MEG). Results reveal qualitative changes in neural activity dependent on stimulus complexity: recognition of tonal sequences engages hippocampal and cingulate areas, whereas recognition of atonal sequences mainly activates the auditory processing network. Our findings reveal the involvement of a cortico-subcortical brain network for auditory recognition and support the idea that stimulus complexity qualitatively alters the neural pathways of recognition memory.
Collapse
Affiliation(s)
- Gemma Fernández-Rubio
- grid.7048.b0000 0001 1956 2722Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music, Aarhus/Aalborg, Aarhus, Denmark ,grid.5012.60000 0001 0481 6099Department of Neuropsychology and Psychopharmacology, Faculty of Psychology and Neuroscience, Maastricht University, Maastricht, The Netherlands
| | - Elvira Brattico
- grid.7048.b0000 0001 1956 2722Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music, Aarhus/Aalborg, Aarhus, Denmark ,grid.7644.10000 0001 0120 3326Department of Education, Psychology, Communication, University of Bari Aldo Moro, Bari, Italy
| | - Sonja A. Kotz
- grid.5012.60000 0001 0481 6099Department of Neuropsychology and Psychopharmacology, Faculty of Psychology and Neuroscience, Maastricht University, Maastricht, The Netherlands
| | - Morten L. Kringelbach
- grid.7048.b0000 0001 1956 2722Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music, Aarhus/Aalborg, Aarhus, Denmark ,grid.4991.50000 0004 1936 8948Centre for Eudaimonia and Human Flourishing, Linacre College, University of Oxford, Oxford, United Kingdom ,grid.4991.50000 0004 1936 8948Department of Psychiatry, University of Oxford, Oxford, United Kingdom
| | - Peter Vuust
- grid.7048.b0000 0001 1956 2722Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music, Aarhus/Aalborg, Aarhus, Denmark
| | - Leonardo Bonetti
- grid.7048.b0000 0001 1956 2722Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music, Aarhus/Aalborg, Aarhus, Denmark ,grid.4991.50000 0004 1936 8948Centre for Eudaimonia and Human Flourishing, Linacre College, University of Oxford, Oxford, United Kingdom ,grid.4991.50000 0004 1936 8948Department of Psychiatry, University of Oxford, Oxford, United Kingdom
| |
Collapse
|
12
|
Comparison of non-invasive, scalp-recorded auditory steady-state responses in humans, rhesus monkeys, and common marmosets. Sci Rep 2022; 12:9210. [PMID: 35654875 PMCID: PMC9163194 DOI: 10.1038/s41598-022-13228-8] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2022] [Accepted: 05/23/2022] [Indexed: 12/27/2022] Open
Abstract
Auditory steady-state responses (ASSRs) are basic neural responses used to probe the ability of auditory circuits to produce synchronous activity to repetitive external stimulation. Reduced ASSR has been observed in patients with schizophrenia, especially at 40 Hz. Although ASSR is a translatable biomarker with a potential both in animal models and patients with schizophrenia, little is known about the features of ASSR in monkeys. Herein, we recorded the ASSR from humans, rhesus monkeys, and marmosets using the same method to directly compare the characteristics of ASSRs among the species. We used auditory trains on a wide range of frequencies to investigate the suitable frequency for ASSRs induction, because monkeys usually use stimulus frequency ranges different from humans for vocalization. We found that monkeys and marmosets also show auditory event-related potentials and phase-locking activity in gamma-frequency trains, although the optimal frequency with the best synchronization differed among these species. These results suggest that the ASSR could be a useful translational, cross-species biomarker to examine the generation of gamma-band synchronization in nonhuman primate models of schizophrenia.
Collapse
|
13
|
Abstract
Hearing in noise is a core problem in audition, and a challenge for hearing-impaired listeners, yet the underlying mechanisms are poorly understood. We explored whether harmonic frequency relations, a signature property of many communication sounds, aid hearing in noise for normal hearing listeners. We measured detection thresholds in noise for tones and speech synthesized to have harmonic or inharmonic spectra. Harmonic signals were consistently easier to detect than otherwise identical inharmonic signals. Harmonicity also improved discrimination of sounds in noise. The largest benefits were observed for two-note up-down "pitch" discrimination and melodic contour discrimination, both of which could be performed equally well with harmonic and inharmonic tones in quiet, but which showed large harmonic advantages in noise. The results show that harmonicity facilitates hearing in noise, plausibly by providing a noise-robust pitch cue that aids detection and discrimination.
Collapse
|
14
|
Cortical activity evoked by voice pitch changes: a combined fNIRS and EEG study. Hear Res 2022; 420:108483. [DOI: 10.1016/j.heares.2022.108483] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/08/2021] [Revised: 03/02/2022] [Accepted: 03/10/2022] [Indexed: 11/22/2022]
|
15
|
Hermann KL, Singh SR, Rosenthal IA, Pantazis D, Conway BR. Temporal dynamics of the neural representation of hue and luminance polarity. Nat Commun 2022; 13:661. [PMID: 35115511 PMCID: PMC8814185 DOI: 10.1038/s41467-022-28249-0] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/16/2021] [Accepted: 01/12/2022] [Indexed: 11/09/2022] Open
Abstract
Hue and luminance contrast are basic visual features. Here we use multivariate analyses of magnetoencephalography data to investigate the timing of the neural computations that extract them, and whether they depend on common neural circuits. We show that hue and luminance-contrast polarity can be decoded from MEG data and, with lower accuracy, both features can be decoded across changes in the other feature. These results are consistent with the existence of both common and separable neural mechanisms. The decoding time course is earlier and more temporally precise for luminance polarity than hue, a result that does not depend on task, suggesting that luminance contrast is an updating signal that separates visual events. Meanwhile, cross-temporal generalization is slightly greater for representations of hue compared to luminance polarity, providing a neural correlate of the preeminence of hue in perceptual grouping and memory. Finally, decoding of luminance polarity varies depending on the hues used to obtain training and testing data. The pattern of results is consistent with observations that luminance contrast is mediated by both L-M and S cone sub-cortical mechanisms.
Collapse
Affiliation(s)
- Katherine L Hermann
- Laboratory of Sensorimotor Research, National Eye Institute, Bethesda, MD, 20892, USA
- Department of Psychology, Stanford University, Stanford, CA, 94305, USA
| | - Shridhar R Singh
- Laboratory of Sensorimotor Research, National Eye Institute, Bethesda, MD, 20892, USA
| | - Isabelle A Rosenthal
- Laboratory of Sensorimotor Research, National Eye Institute, Bethesda, MD, 20892, USA
- Division of Biology and Biological Engineering, California Institute of Technology, Pasadena, CA, 91125, USA
| | - Dimitrios Pantazis
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA, 02139, USA
| | - Bevil R Conway
- Laboratory of Sensorimotor Research, National Eye Institute, Bethesda, MD, 20892, USA.
- National Institute of Mental Health, Bethesda, MD, 20892, USA.
| |
Collapse
|
16
|
Shahdloo M, Schüffelgen U, Papp D, Miller KL, Chiew M. Model-based dynamic off-resonance correction for improved accelerated fMRI in awake behaving nonhuman primates. Magn Reson Med 2022; 87:2922-2932. [PMID: 35081259 PMCID: PMC9306555 DOI: 10.1002/mrm.29167] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2021] [Revised: 11/26/2021] [Accepted: 01/03/2022] [Indexed: 11/18/2022]
Abstract
Purpose To estimate dynamic off‐resonance due to vigorous body motion in accelerated fMRI of awake behaving nonhuman primates (NHPs) using the echo‐planar imaging reference navigator, in order to attenuate the effects of time‐varying off‐resonance on the reconstruction. Methods In NHP fMRI, the animal’s head is usually head‐posted, and the dynamic off‐resonance is mainly caused by motion in body parts that are distant from the brain and have low spatial frequency. Hence, off‐resonance at each frame can be approximated as a spatially linear perturbation of the off‐resonance at a reference frame, and is manifested as a relative linear shift in k‐space. Using GRAPPA operators, we estimated these shifts by comparing the navigator at each time frame with that at the reference frame. Estimated shifts were then used to correct the data at each frame. The proposed method was evaluated in phantom scans, simulations, and in vivo data. Results The proposed method is shown to successfully estimate spatially low‐order dynamic off‐resonance perturbations, including induced linear off‐resonance perturbations in phantoms, and is able to correct retrospectively corrupted data in simulations. Finally, it is shown to reduce ghosting artifacts and geometric distortions by up to 20% in simultaneous multislice in vivo acquisitions in awake‐behaving NHPs. Conclusion A method is proposed that does not need sequence modification or extra acquisitions and makes accelerated awake behaving NHP imaging more robust and reliable, reducing the gap between what is possible with NHP protocols and state‐of‐the‐art human imaging.
Collapse
Affiliation(s)
- Mo Shahdloo
- Wellcome Centre for Integrative Neuroimaging, Department of Experimental Psychology, University of Oxford, Oxford, UK
| | - Urs Schüffelgen
- Wellcome Centre for Integrative Neuroimaging, Department of Experimental Psychology, University of Oxford, Oxford, UK
| | - Daniel Papp
- Wellcome Centre for Integrative Neuroimaging, FMRIB, Nuffield Department of Clinical Neurosciences, University of Oxford, Oxford, UK.,NeuroPoly Lab, Electrical Engineering Department, Polytechnique Montréal, Montreal, Canada
| | - Karla L Miller
- Wellcome Centre for Integrative Neuroimaging, FMRIB, Nuffield Department of Clinical Neurosciences, University of Oxford, Oxford, UK
| | - Mark Chiew
- Wellcome Centre for Integrative Neuroimaging, FMRIB, Nuffield Department of Clinical Neurosciences, University of Oxford, Oxford, UK
| |
Collapse
|
17
|
Landemard A, Bimbard C, Demené C, Shamma S, Norman-Haignere S, Boubenec Y. Distinct higher-order representations of natural sounds in human and ferret auditory cortex. eLife 2021; 10:e65566. [PMID: 34792467 PMCID: PMC8601661 DOI: 10.7554/elife.65566] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2020] [Accepted: 10/22/2021] [Indexed: 11/29/2022] Open
Abstract
Little is known about how neural representations of natural sounds differ across species. For example, speech and music play a unique role in human hearing, yet it is unclear how auditory representations of speech and music differ between humans and other animals. Using functional ultrasound imaging, we measured responses in ferrets to a set of natural and spectrotemporally matched synthetic sounds previously tested in humans. Ferrets showed similar lower-level frequency and modulation tuning to that observed in humans. But while humans showed substantially larger responses to natural vs. synthetic speech and music in non-primary regions, ferret responses to natural and synthetic sounds were closely matched throughout primary and non-primary auditory cortex, even when tested with ferret vocalizations. This finding reveals that auditory representations in humans and ferrets diverge sharply at late stages of cortical processing, potentially driven by higher-order processing demands in speech and music.
Collapse
Affiliation(s)
- Agnès Landemard
- Laboratoire des Systèmes Perceptifs, Département d’Études Cognitives, École Normale Supérieure PSL Research University, CNRSParisFrance
| | - Célian Bimbard
- Laboratoire des Systèmes Perceptifs, Département d’Études Cognitives, École Normale Supérieure PSL Research University, CNRSParisFrance
- University College LondonLondonUnited Kingdom
| | - Charlie Demené
- Physics for Medicine Paris, Inserm, ESPCI Paris, PSL Research University, CNRSParisFrance
| | - Shihab Shamma
- Laboratoire des Systèmes Perceptifs, Département d’Études Cognitives, École Normale Supérieure PSL Research University, CNRSParisFrance
- Institute for Systems Research, Department of Electrical and Computer Engineering, University of MarylandCollege ParkUnited States
| | - Sam Norman-Haignere
- Laboratoire des Systèmes Perceptifs, Département d’Études Cognitives, École Normale Supérieure PSL Research University, CNRSParisFrance
- HHMI Postdoctoral Fellow of the Life Sciences Research FoundationBaltimoreUnited States
- Zuckerman Mind Brain Behavior Institute, Columbia UniversityNew YorkUnited States
| | - Yves Boubenec
- Laboratoire des Systèmes Perceptifs, Département d’Études Cognitives, École Normale Supérieure PSL Research University, CNRSParisFrance
| |
Collapse
|
18
|
Functionally homologous representation of vocalizations in the auditory cortex of humans and macaques. Curr Biol 2021; 31:4839-4844.e4. [PMID: 34506729 PMCID: PMC8585503 DOI: 10.1016/j.cub.2021.08.043] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2021] [Revised: 07/08/2021] [Accepted: 08/13/2021] [Indexed: 11/21/2022]
Abstract
How the evolution of speech has transformed the human auditory cortex compared to other primates remains largely unknown. While primary auditory cortex is organized largely similarly in humans and macaques,1 the picture is much less clear at higher levels of the anterior auditory pathway,2 particularly regarding the processing of conspecific vocalizations (CVs). A "voice region" similar to the human voice-selective areas3,4 has been identified in the macaque right anterior temporal lobe with functional MRI;5 however, its anatomical localization, seemingly inconsistent with that of the human temporal voice areas (TVAs), has suggested a "repositioning of the voice area" in recent human evolution.6 Here we report a functional homology in the cerebral processing of vocalizations by macaques and humans, using comparative fMRI and a condition-rich auditory stimulation paradigm. We find that the anterior temporal lobe of both species possesses cortical voice areas that are bilateral and not only prefer conspecific vocalizations but also implement a representational geometry categorizing them apart from all other sounds in a species-specific but homologous manner. These results reveal a more similar functional organization of higher-level auditory cortex in macaques and humans than currently known.
Collapse
|
19
|
Lowe MX, Mohsenzadeh Y, Lahner B, Charest I, Oliva A, Teng S. Cochlea to categories: The spatiotemporal dynamics of semantic auditory representations. Cogn Neuropsychol 2021; 38:468-489. [PMID: 35729704 PMCID: PMC10589059 DOI: 10.1080/02643294.2022.2085085] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2021] [Revised: 03/31/2022] [Accepted: 05/25/2022] [Indexed: 10/17/2022]
Abstract
How does the auditory system categorize natural sounds? Here we apply multimodal neuroimaging to illustrate the progression from acoustic to semantically dominated representations. Combining magnetoencephalographic (MEG) and functional magnetic resonance imaging (fMRI) scans of observers listening to naturalistic sounds, we found superior temporal responses beginning ∼55 ms post-stimulus onset, spreading to extratemporal cortices by ∼100 ms. Early regions were distinguished less by onset/peak latency than by functional properties and overall temporal response profiles. Early acoustically-dominated representations trended systematically toward category dominance over time (after ∼200 ms) and space (beyond primary cortex). Semantic category representation was spatially specific: Vocalizations were preferentially distinguished in frontotemporal voice-selective regions and the fusiform; scenes and objects were distinguished in parahippocampal and medial place areas. Our results are consistent with real-world events coded via an extended auditory processing hierarchy, in which acoustic representations rapidly enter multiple streams specialized by category, including areas typically considered visual cortex.
Collapse
Affiliation(s)
- Matthew X. Lowe
- Computer Science and Artificial Intelligence Lab (CSAIL), MIT, Cambridge, MA
- Unlimited Sciences, Colorado Springs, CO
| | - Yalda Mohsenzadeh
- Computer Science and Artificial Intelligence Lab (CSAIL), MIT, Cambridge, MA
- The Brain and Mind Institute, The University of Western Ontario, London, ON, Canada
- Department of Computer Science, The University of Western Ontario, London, ON, Canada
| | - Benjamin Lahner
- Computer Science and Artificial Intelligence Lab (CSAIL), MIT, Cambridge, MA
| | - Ian Charest
- Département de Psychologie, Université de Montréal, Montréal, Québec, Canada
- Center for Human Brain Health, University of Birmingham, UK
| | - Aude Oliva
- Computer Science and Artificial Intelligence Lab (CSAIL), MIT, Cambridge, MA
| | - Santani Teng
- Computer Science and Artificial Intelligence Lab (CSAIL), MIT, Cambridge, MA
- Smith-Kettlewell Eye Research Institute (SKERI), San Francisco, CA
| |
Collapse
|
20
|
Is the MSB hypothesis (music as a coevolved system for social bonding) testable in the Popperian sense? Behav Brain Sci 2021; 44:e70. [PMID: 34588070 DOI: 10.1017/s0140525x20001739] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
"Music As a Coevolved System for Social Bonding" (MSB) is a brilliant synthesis and appealing hypothesis offering insights into the evolution and social bonding of musicality, but is so broad and sweeping it will be challenging to test, prove or falsify in the Popperian sense (Popper, 1959). After general comments, I focus my critique on underlying neurobiological mechanisms, and offer some suggestions for experimental tests of MSB.
Collapse
|
21
|
Arcaro MJ, Livingstone MS. On the relationship between maps and domains in inferotemporal cortex. Nat Rev Neurosci 2021; 22:573-583. [PMID: 34345018 PMCID: PMC8865285 DOI: 10.1038/s41583-021-00490-4] [Citation(s) in RCA: 32] [Impact Index Per Article: 10.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 06/24/2021] [Indexed: 02/07/2023]
Abstract
How does the brain encode information about the environment? Decades of research have led to the pervasive notion that the object-processing pathway in primate cortex consists of multiple areas that are each specialized to process different object categories (such as faces, bodies, hands, non-face objects and scenes). The anatomical consistency and modularity of these regions have been interpreted as evidence that these regions are innately specialized. Here, we propose that ventral-stream modules do not represent clusters of circuits that each evolved to process some specific object category particularly important for survival, but instead reflect the effects of experience on a domain-general architecture that evolved to be able to adapt, within a lifetime, to its particular environment. Furthermore, we propose that the mechanisms underlying the development of domains are both evolutionarily old and universal across cortex. Topographic maps are fundamental, governing the development of specializations across systems, providing a framework for brain organization.
Collapse
|
22
|
Hayashi T, Hou Y, Glasser MF, Autio JA, Knoblauch K, Inoue-Murayama M, Coalson T, Yacoub E, Smith S, Kennedy H, Van Essen DC. The nonhuman primate neuroimaging and neuroanatomy project. Neuroimage 2021; 229:117726. [PMID: 33484849 PMCID: PMC8079967 DOI: 10.1016/j.neuroimage.2021.117726] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2020] [Revised: 12/13/2020] [Accepted: 01/02/2021] [Indexed: 11/29/2022] Open
Abstract
Multi-modal neuroimaging projects such as the Human Connectome Project (HCP) and UK Biobank are advancing our understanding of human brain architecture, function, connectivity, and their variability across individuals using high-quality non-invasive data from many subjects. Such efforts depend upon the accuracy of non-invasive brain imaging measures. However, 'ground truth' validation of connectivity using invasive tracers is not feasible in humans. Studies using nonhuman primates (NHPs) enable comparisons between invasive and non-invasive measures, including exploration of how "functional connectivity" from fMRI and "tractographic connectivity" from diffusion MRI compare with long-distance connections measured using tract tracing. Our NonHuman Primate Neuroimaging & Neuroanatomy Project (NHP_NNP) is an international effort (6 laboratories in 5 countries) to: (i) acquire and analyze high-quality multi-modal brain imaging data of macaque and marmoset monkeys using protocols and methods adapted from the HCP; (ii) acquire quantitative invasive tract-tracing data for cortical and subcortical projections to cortical areas; and (iii) map the distributions of different brain cell types with immunocytochemical stains to better define brain areal boundaries. We are acquiring high-resolution structural, functional, and diffusion MRI data together with behavioral measures from over 100 individual macaques and marmosets in order to generate non-invasive measures of brain architecture such as myelin and cortical thickness maps, as well as functional and diffusion tractography-based connectomes. We are using classical and next-generation anatomical tracers to generate quantitative connectivity maps based on brain-wide counting of labeled cortical and subcortical neurons, providing ground truth measures of connectivity. Advanced statistical modeling techniques address the consistency of both kinds of data across individuals, allowing comparison of tracer-based and non-invasive MRI-based connectivity measures. We aim to develop improved cortical and subcortical areal atlases by combining histological and imaging methods. Finally, we are collecting genetic and sociality-associated behavioral data in all animals in an effort to understand how genetic variation shapes the connectome and behavior.
Collapse
Affiliation(s)
- Takuya Hayashi
- Laboratory for Brain Connectomics Imaging, RIKEN Center for Biosystems Dynamics Research, 6-7-3 MI R&D Center 3F, Minatojima-minamimachi, Chuo-ku, Kobe 650-0047, Japan; Department of Neurobiology, Kyoto University Graduate School of Medicine, Kyoto, Japan
| | - Yujie Hou
- Inserm, Stem Cell and Brain Research Institute U1208, Univ Lyon, Université Claude Bernard Lyon 1, Bron, France
| | - Matthew F Glasser
- Department of Neuroscience, Washington University Medical School, St Louis, MO USA; Department of Neuroscience and Radiology, Washington University Medical School, St Louis, MO USA
| | - Joonas A Autio
- Laboratory for Brain Connectomics Imaging, RIKEN Center for Biosystems Dynamics Research, 6-7-3 MI R&D Center 3F, Minatojima-minamimachi, Chuo-ku, Kobe 650-0047, Japan
| | - Kenneth Knoblauch
- Inserm, Stem Cell and Brain Research Institute U1208, Univ Lyon, Université Claude Bernard Lyon 1, Bron, France
| | | | - Tim Coalson
- Department of Neuroscience, Washington University Medical School, St Louis, MO USA
| | - Essa Yacoub
- Center for Magnetic Resonance Research, Department of Radiology, University of Minnesota, Minneapolis, USA
| | - Stephen Smith
- Oxford Centre for Functional Magnetic Resonance Imaging of the Brain (FMRIB), Wellcome Centre for Integrative Neuroimaging (WIN), Nuffield Department of Clinical Neurosciences, Oxford University, Oxford, UK
| | - Henry Kennedy
- Inserm, Stem Cell and Brain Research Institute U1208, Univ Lyon, Université Claude Bernard Lyon 1, Bron, France; Institute of Neuroscience, State Key Laboratory of Neuroscience, Chinese Academy of Sciences (CAS) Key Laboratory of Primate Neurobiology, CAS, Shanghai, China
| | - David C Van Essen
- Department of Neuroscience, Washington University Medical School, St Louis, MO USA
| |
Collapse
|
23
|
Hajizadeh A, Matysiak A, Brechmann A, König R, May PJC. Why do humans have unique auditory event-related fields? Evidence from computational modeling and MEG experiments. Psychophysiology 2021; 58:e13769. [PMID: 33475173 DOI: 10.1111/psyp.13769] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/20/2020] [Revised: 12/04/2020] [Accepted: 12/20/2020] [Indexed: 11/28/2022]
Abstract
Auditory event-related fields (ERFs) measured with magnetoencephalography (MEG) are useful for studying the neuronal underpinnings of auditory cognition in human cortex. They have a highly subject-specific morphology, albeit certain characteristic deflections (e.g., P1m, N1m, and P2m) can be identified in most subjects. Here, we explore the reason for this subject-specificity through a combination of MEG measurements and computational modeling of auditory cortex. We test whether ERF subject-specificity can predominantly be explained in terms of each subject having an individual cortical gross anatomy, which modulates the MEG signal, or whether individual cortical dynamics is also at play. To our knowledge, this is the first time that tools to address this question are being presented. The effects of anatomical and dynamical variation on the MEG signal is simulated in a model describing the core-belt-parabelt structure of the auditory cortex, and with the dynamics based on the leaky-integrator neuron model. The experimental and simulated ERFs are characterized in terms of the N1m amplitude, latency, and width. Also, we examine the waveform grand-averaged across subjects, and the standard deviation of this grand average. The results show that the intersubject variability of the ERF arises out of both the anatomy and the dynamics of auditory cortex being specific to each subject. Moreover, our results suggest that the latency variation of the N1m is largely related to subject-specific dynamics. The findings are discussed in terms of how learning, plasticity, and sound detection are reflected in the auditory ERFs. The notion of the grand-averaged ERF is critically evaluated.
Collapse
Affiliation(s)
- Aida Hajizadeh
- Leibniz Institute for Neurobiology, Research Group Comparative Neuroscience, Magdeburg, Germany
| | - Artur Matysiak
- Leibniz Institute for Neurobiology, Research Group Comparative Neuroscience, Magdeburg, Germany
| | - André Brechmann
- Leibniz Institute for Neurobiology, Combinatorial NeuroImaging Core Facility, Magdeburg, Germany
| | - Reinhard König
- Leibniz Institute for Neurobiology, Research Group Comparative Neuroscience, Magdeburg, Germany
| | - Patrick J C May
- Leibniz Institute for Neurobiology, Research Group Comparative Neuroscience, Magdeburg, Germany.,Department of Psychology, Lancaster University, Lancaster, UK
| |
Collapse
|
24
|
McPherson MJ, McDermott JH. Time-dependent discrimination advantages for harmonic sounds suggest efficient coding for memory. Proc Natl Acad Sci U S A 2020; 117:32169-32180. [PMID: 33262275 PMCID: PMC7749397 DOI: 10.1073/pnas.2008956117] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Perceptual systems have finite memory resources and must store incoming signals in compressed formats. To explore whether representations of a sound's pitch might derive from this need for compression, we compared discrimination of harmonic and inharmonic sounds across delays. In contrast to inharmonic spectra, harmonic spectra can be summarized, and thus compressed, using their fundamental frequency (f0). Participants heard two sounds and judged which was higher. Despite being comparable for sounds presented back-to-back, discrimination was better for harmonic than inharmonic stimuli when sounds were separated in time, implicating memory representations unique to harmonic sounds. Patterns of individual differences (correlations between thresholds in different conditions) indicated that listeners use different representations depending on the time delay between sounds, directly comparing the spectra of temporally adjacent sounds, but transitioning to comparing f0s across delays. The need to store sound in memory appears to determine reliance on f0-based pitch and may explain its importance in music, in which listeners must extract relationships between notes separated in time.
Collapse
Affiliation(s)
- Malinda J McPherson
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139;
- Program in Speech and Hearing Bioscience and Technology, Harvard University, Boston, MA 02115
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139
| | - Josh H McDermott
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139;
- Program in Speech and Hearing Bioscience and Technology, Harvard University, Boston, MA 02115
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139
- Center for Brains, Minds and Machines, Massachusetts Institute of Technology, Cambridge, MA 02139
| |
Collapse
|
25
|
Steinmetzger K, Shen Z, Riedel H, Rupp A. Auditory cortex activity measured using functional near-infrared spectroscopy (fNIRS) appears to be susceptible to masking by cortical blood stealing. Hear Res 2020; 396:108069. [DOI: 10.1016/j.heares.2020.108069] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/05/2020] [Revised: 08/24/2020] [Accepted: 08/31/2020] [Indexed: 01/21/2023]
|
26
|
Letailleur A, Bisesi E, Legrain P. Strategies Used by Musicians to Identify Notes' Pitch: Cognitive Bricks and Mental Representations. Front Psychol 2020; 11:1480. [PMID: 32733333 PMCID: PMC7358308 DOI: 10.3389/fpsyg.2020.01480] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2019] [Accepted: 06/02/2020] [Indexed: 12/20/2022] Open
Abstract
To this day, the study of the substratum of thought and its implied mechanisms is rarely directly addressed. Nowadays, systemic approaches based on introspective methodologies are no longer fashionable and are often overlooked or ignored. Most frequently, reductionist approaches are followed for deciphering the neuronal circuits functionally associated with cognitive processes. However, we argue that systemic studies of individual thought may still contribute to a useful and complementary description of the multimodal nature of perception, because they can take into account individual diversity while still identifying the common features of perceptual processes. We propose to address this question by looking at one possible task for recognition of a "signifying sound", as an example of conceptual grasping of a perceptual response. By adopting a mixed approach combining qualitative analyses of interviews based on introspection with quantitative statistical analyses carried out on the resulting categorization, this study describes a variety of mental strategies used by musicians to identify notes' pitch. Sixty-seven musicians (music students and professionals) were interviewed, revealing that musicians utilize intermediate steps during note identification by selecting or activating cognitive bricks that help construct and reach the correct decision. We named these elements "mental anchorpoints" (MA). Although the anchorpoints are not universal, and differ between individuals, they can be grouped into categories related to three main sensory modalities - auditory, visual and kinesthetic. Such categorization enabled us to characterize the mental representations (MR) that allow musicians to name notes in relationship to eleven basic typologies of anchorpoints. We propose a conceptual framework which summarizes the process of note identification in five steps, starting from sensory detection and ending with the verbalization of the note pitch, passing through the pivotal role of MAs and MRs. We found that musicians use multiple strategies and select individual combinations of MAs belonging to these three different sensory modalities, both in isolation and in combination.
Collapse
Affiliation(s)
- Alain Letailleur
- CNRS UMR 8131, Centre Georg Simmel Recherches Franco-Allemandes en Sciences Sociales, École des Hautes Études en Sciences Sociales (EHESS), Paris, France
| | - Erica Bisesi
- CNRS UMR 3571, Paris, France.,Unité Perception et Mémoire, Institut Pasteur, Paris, France
| | - Pierre Legrain
- CNRS UMR 3571, Paris, France.,Unité Perception et Mémoire, Institut Pasteur, Paris, France
| |
Collapse
|
27
|
Perceptual fusion of musical notes by native Amazonians suggests universal representations of musical intervals. Nat Commun 2020; 11:2786. [PMID: 32493923 PMCID: PMC7270137 DOI: 10.1038/s41467-020-16448-6] [Citation(s) in RCA: 24] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2019] [Accepted: 04/23/2020] [Indexed: 01/31/2023] Open
Abstract
Music perception is plausibly constrained by universal perceptual mechanisms adapted to natural sounds. Such constraints could arise from our dependence on harmonic frequency spectra for segregating concurrent sounds, but evidence has been circumstantial. We measured the extent to which concurrent musical notes are misperceived as a single sound, testing Westerners as well as native Amazonians with limited exposure to Western music. Both groups were more likely to mistake note combinations related by simple integer ratios as single sounds (‘fusion’). Thus, even with little exposure to Western harmony, acoustic constraints on sound segregation appear to induce perceptual structure on note combinations. However, fusion did not predict aesthetic judgments of intervals in Westerners, or in Amazonians, who were indifferent to consonance/dissonance. The results suggest universal perceptual mechanisms that could help explain cross-cultural regularities in musical systems, but indicate that these mechanisms interact with culture-specific influences to produce musical phenomena such as consonance. Music varies across cultures, but some features are widespread, consistent with biological constraints. Here, the authors report that both Western and native Amazonian listeners perceptually fuse concurrent notes related by simple-integer ratios, suggestive of one such biological constraint.
Collapse
|
28
|
Towards HCP-Style macaque connectomes: 24-Channel 3T multi-array coil, MRI sequences and preprocessing. Neuroimage 2020; 215:116800. [PMID: 32276072 DOI: 10.1016/j.neuroimage.2020.116800] [Citation(s) in RCA: 41] [Impact Index Per Article: 10.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/25/2019] [Revised: 03/16/2020] [Accepted: 03/23/2020] [Indexed: 11/23/2022] Open
Abstract
Macaque monkeys are an important animal model where invasive investigations can lead to a better understanding of the cortical organization of primates including humans. However, the tools and methods for noninvasive image acquisition (e.g. MRI RF coils and pulse sequence protocols) and image data preprocessing have lagged behind those developed for humans. To resolve the structural and functional characteristics of the smaller macaque brain, high spatial, temporal, and angular resolutions combined with high signal-to-noise ratio are required to ensure good image quality. To address these challenges, we developed a macaque 24-channel receive coil for 3-T MRI with parallel imaging capabilities. This coil enables adaptation of the Human Connectome Project (HCP) image acquisition protocols to the in-vivo macaque brain. In addition, we adapted HCP preprocessing methods to the macaque brain, including spatial minimal preprocessing of structural, functional MRI (fMRI), and diffusion MRI (dMRI). The coil provides the necessary high signal-to-noise ratio and high efficiency in data acquisition, allowing four- and five-fold accelerations for dMRI and fMRI. Automated FreeSurfer segmentation of cortex, reconstruction of cortical surface, removal of artefacts and nuisance signals in fMRI, and distortion correction of dMRI all performed well, and the overall quality of basic neurobiological measures was comparable with those for the HCP. Analyses of functional connectivity in fMRI revealed high sensitivity as compared with those from publicly shared datasets. Tractography-based connectivity estimates correlated with tracer connectivity similarly to that achieved using ex-vivo dMRI. The resulting HCP-style in vivo macaque MRI data show considerable promise for analyzing cortical architecture and functional and structural connectivity using advanced methods that have previously only been available in studies of the human brain.
Collapse
|
29
|
Steel MM, Polonenko MJ, Giannantonio S, Hopyan T, Papsin BC, Gordon KA. Music Perception Testing Reveals Advantages and Continued Challenges for Children Using Bilateral Cochlear Implants. Front Psychol 2020; 10:3015. [PMID: 32038391 PMCID: PMC6985588 DOI: 10.3389/fpsyg.2019.03015] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2019] [Accepted: 12/19/2019] [Indexed: 11/25/2022] Open
Abstract
A modified version of the child’s Montreal Battery of Evaluation of Amusia (cMBEA) was used to assess music perception in children using bilateral cochlear implants. Our overall aim was to promote better performance by children with CIs on the cMBEA by modifying the complement of instruments used in the test and adding pieces transposed in frequency. The 10 test trials played by piano were removed and two high and two low frequency trials added to each of five subtests (20 additional). The modified cMBEA was completed by 14 children using bilateral cochlear implants and 23 peers with normal hearing. Results were compared with performance on the original version of the cMBEA previously reported in groups of similar aged children: 2 groups with normal hearing (n = 23: Hopyan et al., 2012; n = 16: Polonenko et al., 2017), 1 group using bilateral cochlear implants (CIs) (n = 26: Polonenko et al., 2017), 1 group using bimodal (hearing aid and CI) devices (n = 8: Polonenko et al., 2017), and 1 group using unilateral CI (n = 23: Hopyan et al., 2012). Children with normal hearing had high scores on the modified version of the cMBEA and there were no significant score differences from children with normal hearing who completed the original cMBEA. Children with CIs showed no significant improvement in scores on the modified cMBEA compared to peers with CIs who completed the original version of the test. The group with bilateral CIs who completed the modified cMBEA showed a trend toward better abilities to remember music compared to children listening through a unilateral CI but effects were smaller than in previous cohorts of children with bilateral CIs and bimodal devices who completed the original cMBEA. Results confirmed that musical perception changes with the type of instrument and is better for music transposed to higher rather than lower frequencies for children with normal hearing but not for children using bilateral CIs. Overall, the modified version of the cMBEA revealed that modifications to music do not overcome the limitations of the CI to improve music perception for children.
Collapse
Affiliation(s)
- Morrison M Steel
- Archie's Cochlear Implant Laboratory, The Hospital for Sick Children, Toronto, ON, Canada
| | - Melissa J Polonenko
- Archie's Cochlear Implant Laboratory, The Hospital for Sick Children, Toronto, ON, Canada
| | - Sara Giannantonio
- Archie's Cochlear Implant Laboratory, The Hospital for Sick Children, Toronto, ON, Canada
| | - Talar Hopyan
- Archie's Cochlear Implant Laboratory, The Hospital for Sick Children, Toronto, ON, Canada
| | - Blake C Papsin
- Archie's Cochlear Implant Laboratory, The Hospital for Sick Children, Toronto, ON, Canada.,Department of Otolaryngology-Head and Neck Surgery, University of Toronto, Toronto, ON, Canada
| | - Karen A Gordon
- Archie's Cochlear Implant Laboratory, The Hospital for Sick Children, Toronto, ON, Canada.,Department of Otolaryngology-Head and Neck Surgery, University of Toronto, Toronto, ON, Canada
| |
Collapse
|