1
|
Wang M, Wang Z, Yu Y, Zhao D, Shen Z, Wei F. From teeth to brain: dental caries causally affects the cortical thickness of the banks of the superior temporal sulcus. BMC Oral Health 2024; 24:124. [PMID: 38263072 PMCID: PMC10807149 DOI: 10.1186/s12903-024-03899-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2023] [Accepted: 01/15/2024] [Indexed: 01/25/2024] Open
Abstract
OBJECTIVES Dental caries is one of the most prevalent oral diseases and causes of tooth loss. Cross-sectional studies observed epidemiological associations between dental caries and brain degeneration disorders, while it is unknown whether dental caries causally affect the cerebral structures. This study tested whether genetically proxied DMFS (the sum of Decayed, Missing, and Filled tooth Surfaces) causally impacts the brain cortical structure using Mendelian randomization (MR). METHODS The summary-level GWAS meta-analysis data from the GLIDE consortium were used for DMFS, including 26,792 participants. ENIGMA (Enhancing NeuroImaging Genetics through Meta Analysis) consortium GWAS summary data of 51,665 patients were used for brain structure. This study estimated the causal effects of DMFS on the surface area (SA) and thickness (TH) of the global cortex and functional cortical regions accessed by magnetic resonance imaging (MRI). Inverse-variance weighted (IVW) was used as the primary estimate, the MR pleiotropy residual sum and outlier (MR-PRESSO), the MR-Egger intercept test, and leave-one-out analyses were used to examine the potential horizontal pleiotropy. RESULTS Genetically proxied DMFS decreases the TH of the banks of the superior temporal sulcus (BANSSTS) with or without global weighted (weighted, β = - 0.0277 mm, 95% CI: - 0.0470 mm to - 0.0085 mm, P = 0.0047; unweighted, β = - 0.0311 mm, 95% CI: - 0.0609 mm to - 0.0012 mm, P = 0.0412). The causal associations were robust in various sensitivity analyses. CONCLUSIONS Dental caries causally decrease the cerebral cortical thickness of the BANKSSTS, a cerebral cortical region crucial for language-related functions, and is the most affected brain region in Alzheimer's disease. This investigation provides the first evidence that dental caries causally affects brain structure, proving the existence of teeth-brain axes. This study also suggested that clinicians should highlight the causal effects of dental caries on brain disorders during the diagnosis and treatments, the cortical thickness of BANKSSTS is a promising diagnostic measurement for dental caries-related brain degeneration.
Collapse
Affiliation(s)
- Mengqiao Wang
- Department of Orthodontics, School and Hospital of Stomatology, Cheeloo College of Medicine, Shandong University, Jinan, China
- Shandong Key Laboratory of Oral Tissue Regeneration & Shandong Engineering Research Center of Dental Materials and Oral Tissue Regeneration & Shandong Provincial Clinical Research Center for Oral Diseases, Jinan, China
| | - Ziyao Wang
- Department of Orthodontics, School and Hospital of Stomatology, Cheeloo College of Medicine, Shandong University, Jinan, China
- Shandong Key Laboratory of Oral Tissue Regeneration & Shandong Engineering Research Center of Dental Materials and Oral Tissue Regeneration & Shandong Provincial Clinical Research Center for Oral Diseases, Jinan, China
| | - Yajie Yu
- National Clinical Research Center for Mental Disorders, Peking University Sixth Hospital/Institute of Mental Health, The Key Laboratory of Mental Health, Ministry of Health, Peking University, Beijing, 100191, China
| | - Delu Zhao
- Department of Orthodontics, School and Hospital of Stomatology, Cheeloo College of Medicine, Shandong University, Jinan, China
- Shandong Key Laboratory of Oral Tissue Regeneration & Shandong Engineering Research Center of Dental Materials and Oral Tissue Regeneration & Shandong Provincial Clinical Research Center for Oral Diseases, Jinan, China
| | - Zhiyuan Shen
- Department of Oral and Maxillofacial Surgery, School and Hospital of Stomatology, Cheeloo College of Medicine, Shandong University, Jinan, China
- Shandong Key Laboratory of Oral Tissue Regeneration & Shandong Engineering Research Center of Dental Materials and Oral Tissue Regeneration & Shandong Provincial Clinical Research Center for Oral Diseases, Jinan, China
| | - Fulan Wei
- Department of Orthodontics, School and Hospital of Stomatology, Cheeloo College of Medicine, Shandong University, Jinan, China.
- Shandong Key Laboratory of Oral Tissue Regeneration & Shandong Engineering Research Center of Dental Materials and Oral Tissue Regeneration & Shandong Provincial Clinical Research Center for Oral Diseases, Jinan, China.
| |
Collapse
|
2
|
Nidiffer AR, Cao CZ, O'Sullivan A, Lalor EC. A representation of abstract linguistic categories in the visual system underlies successful lipreading. Neuroimage 2023; 282:120391. [PMID: 37757989 DOI: 10.1016/j.neuroimage.2023.120391] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2022] [Revised: 09/22/2023] [Accepted: 09/24/2023] [Indexed: 09/29/2023] Open
Abstract
There is considerable debate over how visual speech is processed in the absence of sound and whether neural activity supporting lipreading occurs in visual brain areas. Much of the ambiguity stems from a lack of behavioral grounding and neurophysiological analyses that cannot disentangle high-level linguistic and phonetic/energetic contributions from visual speech. To address this, we recorded EEG from human observers as they watched silent videos, half of which were novel and half of which were previously rehearsed with the accompanying audio. We modeled how the EEG responses to novel and rehearsed silent speech reflected the processing of low-level visual features (motion, lip movements) and a higher-level categorical representation of linguistic units, known as visemes. The ability of these visemes to account for the EEG - beyond the motion and lip movements - was significantly enhanced for rehearsed videos in a way that correlated with participants' trial-by-trial ability to lipread that speech. Source localization of viseme processing showed clear contributions from visual cortex, with no strong evidence for the involvement of auditory areas. We interpret this as support for the idea that the visual system produces its own specialized representation of speech that is (1) well-described by categorical linguistic features, (2) dissociable from lip movements, and (3) predictive of lipreading ability. We also suggest a reinterpretation of previous findings of auditory cortical activation during silent speech that is consistent with hierarchical accounts of visual and audiovisual speech perception.
Collapse
Affiliation(s)
- Aaron R Nidiffer
- Department of Biomedical Engineering, Department of Neuroscience, Del Monte Institute for Neuroscience, University of Rochester, Rochester, NY, USA
| | - Cody Zhewei Cao
- Department of Psychology, University of Michigan, Ann Arbor, MI, USA
| | - Aisling O'Sullivan
- School of Engineering, Trinity College Institute of Neuroscience, Trinity Centre for Biomedical Engineering, Trinity College, Dublin, Ireland
| | - Edmund C Lalor
- Department of Biomedical Engineering, Department of Neuroscience, Del Monte Institute for Neuroscience, University of Rochester, Rochester, NY, USA; School of Engineering, Trinity College Institute of Neuroscience, Trinity Centre for Biomedical Engineering, Trinity College, Dublin, Ireland.
| |
Collapse
|
3
|
Arioli M, Segatta C, Papagno C, Tettamanti M, Cattaneo Z. Social perception in deaf individuals: A meta-analysis of neuroimaging studies. Hum Brain Mapp 2023; 44:5402-5415. [PMID: 37609693 PMCID: PMC10543108 DOI: 10.1002/hbm.26444] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2023] [Revised: 06/22/2023] [Accepted: 07/22/2023] [Indexed: 08/24/2023] Open
Abstract
Deaf individuals may report difficulties in social interactions. However, whether these difficulties depend on deafness affecting social brain circuits is controversial. Here, we report the first meta-analysis comparing brain activations of hearing and (prelingually) deaf individuals during social perception. Our findings showed that deafness does not impact on the functional mechanisms supporting social perception. Indeed, both deaf and hearing control participants recruited regions of the action observation network during performance of different social tasks employing visual stimuli, and including biological motion perception, face identification, action observation, viewing, identification and memory for signs and lip reading. Moreover, we found increased recruitment of the superior-middle temporal cortex in deaf individuals compared with hearing participants, suggesting a preserved and augmented function during social communication based on signs and lip movements. Overall, our meta-analysis suggests that social difficulties experienced by deaf individuals are unlikely to be associated with brain alterations but may rather depend on non-supportive environments.
Collapse
Affiliation(s)
- Maria Arioli
- Department of Human and Social SciencesUniversity of BergamoBergamoItaly
| | - Cecilia Segatta
- Department of Human and Social SciencesUniversity of BergamoBergamoItaly
| | - Costanza Papagno
- Center for Mind/Brain Sciences (CIMeC)University of TrentoTrentoItaly
| | | | - Zaira Cattaneo
- Department of Human and Social SciencesUniversity of BergamoBergamoItaly
- IRCCS Mondino FoundationPaviaItaly
| |
Collapse
|
4
|
Krason A, Vigliocco G, Mailend ML, Stoll H, Varley R, Buxbaum LJ. Benefit of visual speech information for word comprehension in post-stroke aphasia. Cortex 2023; 165:86-100. [PMID: 37271014 PMCID: PMC10850036 DOI: 10.1016/j.cortex.2023.04.011] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2022] [Revised: 03/13/2023] [Accepted: 04/22/2023] [Indexed: 06/06/2023]
Abstract
Aphasia is a language disorder that often involves speech comprehension impairments affecting communication. In face-to-face settings, speech is accompanied by mouth and facial movements, but little is known about the extent to which they benefit aphasic comprehension. This study investigated the benefit of visual information accompanying speech for word comprehension in people with aphasia (PWA) and the neuroanatomic substrates of any benefit. Thirty-six PWA and 13 neurotypical matched control participants performed a picture-word verification task in which they indicated whether a picture of an animate/inanimate object matched a subsequent word produced by an actress in a video. Stimuli were either audiovisual (with visible mouth and facial movements) or auditory-only (still picture of a silhouette) with audio being clear (unedited) or degraded (6-band noise-vocoding). We found that visual speech information was more beneficial for neurotypical participants than PWA, and more beneficial for both groups when speech was degraded. A multivariate lesion-symptom mapping analysis for the degraded speech condition showed that lesions to superior temporal gyrus, underlying insula, primary and secondary somatosensory cortices, and inferior frontal gyrus were associated with reduced benefit of audiovisual compared to auditory-only speech, suggesting that the integrity of these fronto-temporo-parietal regions may facilitate cross-modal mapping. These findings provide initial insights into our understanding of the impact of audiovisual information on comprehension in aphasia and the brain regions mediating any benefit.
Collapse
Affiliation(s)
- Anna Krason
- Experimental Psychology, University College London, UK; Moss Rehabilitation Research Institute, Elkins Park, PA, USA.
| | - Gabriella Vigliocco
- Experimental Psychology, University College London, UK; Moss Rehabilitation Research Institute, Elkins Park, PA, USA
| | - Marja-Liisa Mailend
- Moss Rehabilitation Research Institute, Elkins Park, PA, USA; Department of Special Education, University of Tartu, Tartu Linn, Estonia
| | - Harrison Stoll
- Moss Rehabilitation Research Institute, Elkins Park, PA, USA; Applied Cognitive and Brain Science, Drexel University, Philadelphia, PA, USA
| | | | - Laurel J Buxbaum
- Moss Rehabilitation Research Institute, Elkins Park, PA, USA; Department of Rehabilitation Medicine, Thomas Jefferson University, Philadelphia, PA, USA
| |
Collapse
|
5
|
Banks MI, Krause BM, Berger DG, Campbell DI, Boes AD, Bruss JE, Kovach CK, Kawasaki H, Steinschneider M, Nourski KV. Functional geometry of auditory cortical resting state networks derived from intracranial electrophysiology. PLoS Biol 2023; 21:e3002239. [PMID: 37651504 PMCID: PMC10499207 DOI: 10.1371/journal.pbio.3002239] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2022] [Revised: 09/13/2023] [Accepted: 07/07/2023] [Indexed: 09/02/2023] Open
Abstract
Understanding central auditory processing critically depends on defining underlying auditory cortical networks and their relationship to the rest of the brain. We addressed these questions using resting state functional connectivity derived from human intracranial electroencephalography. Mapping recording sites into a low-dimensional space where proximity represents functional similarity revealed a hierarchical organization. At a fine scale, a group of auditory cortical regions excluded several higher-order auditory areas and segregated maximally from the prefrontal cortex. On mesoscale, the proximity of limbic structures to the auditory cortex suggested a limbic stream that parallels the classically described ventral and dorsal auditory processing streams. Identities of global hubs in anterior temporal and cingulate cortex depended on frequency band, consistent with diverse roles in semantic and cognitive processing. On a macroscale, observed hemispheric asymmetries were not specific for speech and language networks. This approach can be applied to multivariate brain data with respect to development, behavior, and disorders.
Collapse
Affiliation(s)
- Matthew I. Banks
- Department of Anesthesiology, University of Wisconsin, Madison, Wisconsin, United States of America
- Department of Neuroscience, University of Wisconsin, Madison, Wisconsin, United States of America
| | - Bryan M. Krause
- Department of Anesthesiology, University of Wisconsin, Madison, Wisconsin, United States of America
| | - D. Graham Berger
- Department of Anesthesiology, University of Wisconsin, Madison, Wisconsin, United States of America
| | - Declan I. Campbell
- Department of Anesthesiology, University of Wisconsin, Madison, Wisconsin, United States of America
| | - Aaron D. Boes
- Department of Neurology, The University of Iowa, Iowa City, Iowa, United States of America
| | - Joel E. Bruss
- Department of Neurology, The University of Iowa, Iowa City, Iowa, United States of America
| | - Christopher K. Kovach
- Department of Neurosurgery, The University of Iowa, Iowa City, Iowa, United States of America
| | - Hiroto Kawasaki
- Department of Neurosurgery, The University of Iowa, Iowa City, Iowa, United States of America
| | - Mitchell Steinschneider
- Department of Neurology, Albert Einstein College of Medicine, New York, New York, United States of America
- Department of Neuroscience, Albert Einstein College of Medicine, New York, New York, United States of America
| | - Kirill V. Nourski
- Department of Neurosurgery, The University of Iowa, Iowa City, Iowa, United States of America
- Iowa Neuroscience Institute, The University of Iowa, Iowa City, Iowa, United States of America
| |
Collapse
|
6
|
Bernstein LE, Auer ET, Eberhardt SP. Modality-Specific Perceptual Learning of Vocoded Auditory versus Lipread Speech: Different Effects of Prior Information. Brain Sci 2023; 13:1008. [PMID: 37508940 PMCID: PMC10377548 DOI: 10.3390/brainsci13071008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2023] [Revised: 05/26/2023] [Accepted: 06/08/2023] [Indexed: 07/30/2023] Open
Abstract
Traditionally, speech perception training paradigms have not adequately taken into account the possibility that there may be modality-specific requirements for perceptual learning with auditory-only (AO) versus visual-only (VO) speech stimuli. The study reported here investigated the hypothesis that there are modality-specific differences in how prior information is used by normal-hearing participants during vocoded versus VO speech training. Two different experiments, one with vocoded AO speech (Experiment 1) and one with VO, lipread, speech (Experiment 2), investigated the effects of giving different types of prior information to trainees on each trial during training. The training was for four ~20 min sessions, during which participants learned to label novel visual images using novel spoken words. Participants were assigned to different types of prior information during training: Word Group trainees saw a printed version of each training word (e.g., "tethon"), and Consonant Group trainees saw only its consonants (e.g., "t_th_n"). Additional groups received no prior information (i.e., Experiment 1, AO Group; Experiment 2, VO Group) or a spoken version of the stimulus in a different modality from the training stimuli (Experiment 1, Lipread Group; Experiment 2, Vocoder Group). That is, in each experiment, there was a group that received prior information in the modality of the training stimuli from the other experiment. In both experiments, the Word Groups had difficulty retaining the novel words they attempted to learn during training. However, when the training stimuli were vocoded, the Word Group improved their phoneme identification. When the training stimuli were visual speech, the Consonant Group improved their phoneme identification and their open-set sentence lipreading. The results are considered in light of theoretical accounts of perceptual learning in relationship to perceptual modality.
Collapse
Affiliation(s)
- Lynne E Bernstein
- Speech, Language, and Hearing Sciences Department, George Washington University, Washington, DC 20052, USA
| | - Edward T Auer
- Speech, Language, and Hearing Sciences Department, George Washington University, Washington, DC 20052, USA
| | - Silvio P Eberhardt
- Speech, Language, and Hearing Sciences Department, George Washington University, Washington, DC 20052, USA
| |
Collapse
|
7
|
Dopierała AAW, Pérez DL, Mercure E, Pluta A, Malinowska-Korczak A, Evans S, Wolak T, Tomalski P. The Development of Cortical Responses to the Integration of Audiovisual Speech in Infancy. Brain Topogr 2023:10.1007/s10548-023-00959-8. [PMID: 37171657 PMCID: PMC10176292 DOI: 10.1007/s10548-023-00959-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2022] [Accepted: 04/11/2023] [Indexed: 05/13/2023]
Abstract
In adults, the integration of audiovisual speech elicits specific higher (super-additive) or lower (sub-additive) cortical responses when compared to the responses to unisensory stimuli. Although there is evidence that the fronto-temporal network is active during perception of audiovisual speech in infancy, the development of fronto-temporal responses to audiovisual integration remains unknown. In the current study, 5-month-olds and 10-month-olds watched bimodal (audiovisual) and alternating unimodal (auditory + visual) syllables. In this context we use alternating unimodal to denote alternating auditory and visual syllables that are perceived as separate syllables by adults. Using fNIRS we measured responses over large cortical areas including the inferior frontal and superior temporal regions. We identified channels showing different responses to bimodal than alternating unimodal condition and used multivariate pattern analysis (MVPA) to decode patterns of cortical responses to bimodal (audiovisual) and alternating unimodal (auditory + visual) speech. Results showed that in both age groups integration elicits cortical responses consistent with both super- and sub-additive responses in the fronto-temporal cortex. The univariate analyses revealed that between 5 and 10 months spatial distribution of these responses becomes increasingly focal. MVPA correctly classified responses at 5 months, with key input from channels located in the inferior frontal and superior temporal channels of the right hemisphere. However, MVPA classification was not successful at 10 months, suggesting a potential cortical re-organisation of audiovisual speech perception at this age. These results show the complex and non-gradual development of the cortical responses to integration of congruent audiovisual speech in infancy.
Collapse
Affiliation(s)
- Aleksandra A W Dopierała
- Faculty of Psychology, University of Warsaw, Warsaw, Poland.
- Department of Psychology, University of British Columbia, Vancouver, Canada.
| | | | | | - Agnieszka Pluta
- Faculty of Psychology, University of Warsaw, Warsaw, Poland
- University of Westminister, London, UK
| | | | - Samuel Evans
- Kings College London, London, UK
- University of Westminister, London, UK
| | - Tomasz Wolak
- Institute of Physiology and Pathology of Hearing, Bioimaging Research Center, World Hearing Centre, Warsaw, Poland
| | - Przemysław Tomalski
- Faculty of Psychology, University of Warsaw, Warsaw, Poland.
- Institute of Psychology, Polish Academy of Sciences, Warsaw, Poland.
| |
Collapse
|
8
|
Teghipco A, Okada K, Murphy E, Hickok G. Predictive Coding and Internal Error Correction in Speech Production. NEUROBIOLOGY OF LANGUAGE (CAMBRIDGE, MASS.) 2023; 4:81-119. [PMID: 37229143 PMCID: PMC10205072 DOI: 10.1162/nol_a_00088] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/16/2022] [Accepted: 11/02/2022] [Indexed: 05/27/2023]
Abstract
Speech production involves the careful orchestration of sophisticated systems, yet overt speech errors rarely occur under naturalistic conditions. The present functional magnetic resonance imaging study sought neural evidence for internal error detection and correction by leveraging a tongue twister paradigm that induces the potential for speech errors while excluding any overt errors from analysis. Previous work using the same paradigm in the context of silently articulated and imagined speech production tasks has demonstrated forward predictive signals in auditory cortex during speech and presented suggestive evidence of internal error correction in left posterior middle temporal gyrus (pMTG) on the basis that this area tended toward showing a stronger response when potential speech errors are biased toward nonwords compared to words (Okada et al., 2018). The present study built on this prior work by attempting to replicate the forward prediction and lexicality effects in nearly twice as many participants but introduced novel stimuli designed to further tax internal error correction and detection mechanisms by biasing speech errors toward taboo words. The forward prediction effect was replicated. While no evidence was found for a significant difference in brain response as a function of lexical status of the potential speech error, biasing potential errors toward taboo words elicited significantly greater response in left pMTG than biasing errors toward (neutral) words. Other brain areas showed preferential response for taboo words as well but responded below baseline and were less likely to reflect language processing as indicated by a decoding analysis, implicating left pMTG in internal error correction.
Collapse
Affiliation(s)
- Alex Teghipco
- Department of Cognitive Sciences, University of California, Irvine, CA, USA
| | - Kayoko Okada
- Department of Psychology, Loyola Marymount University, Los Angeles, CA, USA
| | - Emma Murphy
- Department of Psychology, Loyola Marymount University, Los Angeles, CA, USA
| | - Gregory Hickok
- Department of Cognitive Sciences, University of California, Irvine, CA, USA
| |
Collapse
|
9
|
Benetti S, Ferrari A, Pavani F. Multimodal processing in face-to-face interactions: A bridging link between psycholinguistics and sensory neuroscience. Front Hum Neurosci 2023; 17:1108354. [PMID: 36816496 PMCID: PMC9932987 DOI: 10.3389/fnhum.2023.1108354] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2022] [Accepted: 01/11/2023] [Indexed: 02/05/2023] Open
Abstract
In face-to-face communication, humans are faced with multiple layers of discontinuous multimodal signals, such as head, face, hand gestures, speech and non-speech sounds, which need to be interpreted as coherent and unified communicative actions. This implies a fundamental computational challenge: optimally binding only signals belonging to the same communicative action while segregating signals that are not connected by the communicative content. How do we achieve such an extraordinary feat, reliably, and efficiently? To address this question, we need to further move the study of human communication beyond speech-centred perspectives and promote a multimodal approach combined with interdisciplinary cooperation. Accordingly, we seek to reconcile two explanatory frameworks recently proposed in psycholinguistics and sensory neuroscience into a neurocognitive model of multimodal face-to-face communication. First, we introduce a psycholinguistic framework that characterises face-to-face communication at three parallel processing levels: multiplex signals, multimodal gestalts and multilevel predictions. Second, we consider the recent proposal of a lateral neural visual pathway specifically dedicated to the dynamic aspects of social perception and reconceive it from a multimodal perspective ("lateral processing pathway"). Third, we reconcile the two frameworks into a neurocognitive model that proposes how multiplex signals, multimodal gestalts, and multilevel predictions may be implemented along the lateral processing pathway. Finally, we advocate a multimodal and multidisciplinary research approach, combining state-of-the-art imaging techniques, computational modelling and artificial intelligence for future empirical testing of our model.
Collapse
Affiliation(s)
- Stefania Benetti
- Centre for Mind/Brain Sciences, University of Trento, Trento, Italy,Interuniversity Research Centre “Cognition, Language, and Deafness”, CIRCLeS, Catania, Italy,*Correspondence: Stefania Benetti,
| | - Ambra Ferrari
- Max Planck Institute for Psycholinguistics, Donders Institute for Brain, Cognition, and Behaviour, Radboud University, Nijmegen, Netherlands
| | - Francesco Pavani
- Centre for Mind/Brain Sciences, University of Trento, Trento, Italy,Interuniversity Research Centre “Cognition, Language, and Deafness”, CIRCLeS, Catania, Italy
| |
Collapse
|
10
|
Van Engen KJ, Dey A, Sommers MS, Peelle JE. Audiovisual speech perception: Moving beyond McGurk. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 152:3216. [PMID: 36586857 PMCID: PMC9894660 DOI: 10.1121/10.0015262] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/28/2022] [Revised: 10/26/2022] [Accepted: 11/05/2022] [Indexed: 05/29/2023]
Abstract
Although it is clear that sighted listeners use both auditory and visual cues during speech perception, the manner in which multisensory information is combined is a matter of debate. One approach to measuring multisensory integration is to use variants of the McGurk illusion, in which discrepant auditory and visual cues produce auditory percepts that differ from those based on unimodal input. Not all listeners show the same degree of susceptibility to the McGurk illusion, and these individual differences are frequently used as a measure of audiovisual integration ability. However, despite their popularity, we join the voices of others in the field to argue that McGurk tasks are ill-suited for studying real-life multisensory speech perception: McGurk stimuli are often based on isolated syllables (which are rare in conversations) and necessarily rely on audiovisual incongruence that does not occur naturally. Furthermore, recent data show that susceptibility to McGurk tasks does not correlate with performance during natural audiovisual speech perception. Although the McGurk effect is a fascinating illusion, truly understanding the combined use of auditory and visual information during speech perception requires tasks that more closely resemble everyday communication: namely, words, sentences, and narratives with congruent auditory and visual speech cues.
Collapse
Affiliation(s)
- Kristin J Van Engen
- Department of Psychological and Brain Sciences, Washington University, St. Louis, Missouri 63130, USA
| | - Avanti Dey
- PLOS ONE, 1265 Battery Street, San Francisco, California 94111, USA
| | - Mitchell S Sommers
- Department of Psychological and Brain Sciences, Washington University, St. Louis, Missouri 63130, USA
| | - Jonathan E Peelle
- Department of Otolaryngology, Washington University, St. Louis, Missouri 63130, USA
| |
Collapse
|
11
|
Eising E, Mirza-Schreiber N, de Zeeuw EL, Wang CA, Truong DT, Allegrini AG, Shapland CY, Zhu G, Wigg KG, Gerritse ML, Molz B, Alagöz G, Gialluisi A, Abbondanza F, Rimfeld K, van Donkelaar M, Liao Z, Jansen PR, Andlauer TFM, Bates TC, Bernard M, Blokland K, Bonte M, Børglum AD, Bourgeron T, Brandeis D, Ceroni F, Csépe V, Dale PS, de Jong PF, DeFries JC, Démonet JF, Demontis D, Feng Y, Gordon SD, Guger SL, Hayiou-Thomas ME, Hernández-Cabrera JA, Hottenga JJ, Hulme C, Kere J, Kerr EN, Koomar T, Landerl K, Leonard GT, Lovett MW, Lyytinen H, Martin NG, Martinelli A, Maurer U, Michaelson JJ, Moll K, Monaco AP, Morgan AT, Nöthen MM, Pausova Z, Pennell CE, Pennington BF, Price KM, Rajagopal VM, Ramus F, Richer L, Simpson NH, Smith SD, Snowling MJ, Stein J, Strug LJ, Talcott JB, Tiemeier H, van der Schroeff MP, Verhoef E, Watkins KE, Wilkinson M, Wright MJ, Barr CL, Boomsma DI, Carreiras M, Franken MCJ, Gruen JR, Luciano M, Müller-Myhsok B, Newbury DF, Olson RK, Paracchini S, Paus T, Plomin R, Reilly S, Schulte-Körne G, Tomblin JB, van Bergen E, Whitehouse AJO, Willcutt EG, St Pourcain B, Francks C, Fisher SE. Genome-wide analyses of individual differences in quantitatively assessed reading- and language-related skills in up to 34,000 people. Proc Natl Acad Sci U S A 2022; 119:e2202764119. [PMID: 35998220 PMCID: PMC9436320 DOI: 10.1073/pnas.2202764119] [Citation(s) in RCA: 32] [Impact Index Per Article: 16.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/18/2022] [Accepted: 05/31/2022] [Indexed: 12/14/2022] Open
Abstract
The use of spoken and written language is a fundamental human capacity. Individual differences in reading- and language-related skills are influenced by genetic variation, with twin-based heritability estimates of 30 to 80% depending on the trait. The genetic architecture is complex, heterogeneous, and multifactorial, but investigations of contributions of single-nucleotide polymorphisms (SNPs) were thus far underpowered. We present a multicohort genome-wide association study (GWAS) of five traits assessed individually using psychometric measures (word reading, nonword reading, spelling, phoneme awareness, and nonword repetition) in samples of 13,633 to 33,959 participants aged 5 to 26 y. We identified genome-wide significant association with word reading (rs11208009, P = 1.098 × 10-8) at a locus that has not been associated with intelligence or educational attainment. All five reading-/language-related traits showed robust SNP heritability, accounting for 13 to 26% of trait variability. Genomic structural equation modeling revealed a shared genetic factor explaining most of the variation in word/nonword reading, spelling, and phoneme awareness, which only partially overlapped with genetic variation contributing to nonword repetition, intelligence, and educational attainment. A multivariate GWAS of word/nonword reading, spelling, and phoneme awareness maximized power for follow-up investigation. Genetic correlation analysis with neuroimaging traits identified an association with the surface area of the banks of the left superior temporal sulcus, a brain region linked to the processing of spoken and written language. Heritability was enriched for genomic elements regulating gene expression in the fetal brain and in chromosomal regions that are depleted of Neanderthal variants. Together, these results provide avenues for deciphering the biological underpinnings of uniquely human traits.
Collapse
Affiliation(s)
- Else Eising
- Language and Genetics Department, Max Planck Institute for Psycholinguistics, 6525 XD Nijmegen, the Netherlands
| | | | - Eveline L. de Zeeuw
- Department of Biological Psychology, Vrije Universiteit Amsterdam, 1081 BT Amsterdam, the Netherlands
| | - Carol A. Wang
- School of Medicine and Public Health, The University of Newcastle, Newcastle, NSW 2308, Australia
- Mothers and Babies Research Program, Hunter Medical Research Institute, Newcastle, NSW 2305, Australia
| | - Dongnhu T. Truong
- Department of Pediatrics and Genetics, Yale Medical School, New Haven, CT 06510
| | - Andrea G. Allegrini
- Social, Genetic and Developmental Psychiatry Centre, Institute of Psychiatry, Psychology and Neuroscience, King’s College London, London SE5 8AF, United Kingdom
| | - Chin Yang Shapland
- MRC Integrative Epidemiology Unit, University of Bristol, Bristol BS8 2BN, United Kingdom
- Population Health Sciences, University of Bristol, Bristol BS8 2PS, United Kingdom
| | - Gu Zhu
- Genetic Epidemiology Laboratory, QIMR Berghofer Medical Research Institute, Brisbane, QLD 4006, Australia
| | - Karen G. Wigg
- Division of Experimental and Translational Neuroscience, Krembil Research Institute, University Health Network, Toronto, ON M5T 0S8, Canada
| | - Margot L. Gerritse
- Language and Genetics Department, Max Planck Institute for Psycholinguistics, 6525 XD Nijmegen, the Netherlands
| | - Barbara Molz
- Language and Genetics Department, Max Planck Institute for Psycholinguistics, 6525 XD Nijmegen, the Netherlands
| | - Gökberk Alagöz
- Language and Genetics Department, Max Planck Institute for Psycholinguistics, 6525 XD Nijmegen, the Netherlands
| | - Alessandro Gialluisi
- Translational Research in Psychiatry, Max Planck Institute of Psychiatry, 80804 Munich, Germany
- Department of Epidemiology and Prevention, IRCCS Istituto Neurologico Mediterraneo Neuromed, 86077 Pozzilli, Italy
- Department of Medicine and Surgery, University of Insubria, 21100 Varese, Italy
| | - Filippo Abbondanza
- School of Medicine, University of St Andrews, KY16 9TF, St. Andrews, Scotland
| | - Kaili Rimfeld
- Social, Genetic and Developmental Psychiatry Centre, Institute of Psychiatry, Psychology and Neuroscience, King’s College London, London SE5 8AF, United Kingdom
- Department of Psychology, Royal Holloway, University of London, Egham TW20 0EY, United Kingdom
| | - Marjolein van Donkelaar
- Language and Genetics Department, Max Planck Institute for Psycholinguistics, 6525 XD Nijmegen, the Netherlands
| | - Zhijie Liao
- Department of Psychology, University of Toronto, Toronto, ON M5S 3G3,Canada
| | - Philip R. Jansen
- Department of Child and Adolescent Psychiatry/Psychology, Erasmus University Medical Center, 3000 CB Rotterdam, the Netherlands
- Department of Complex Trait Genetics, Center for Neurogenomics and Cognitive Research, Amsterdam Neuroscience, Vrije Universiteit Amsterdam, Amsterdam, 1081 HV the Netherlands
- Department of Human Genetics, VU Medical Center, Amsterdam University Medical Center, 1081 BT Amsterdam, the Netherlands
| | - Till F. M. Andlauer
- Translational Research in Psychiatry, Max Planck Institute of Psychiatry, 80804 Munich, Germany
- Department of Neurology, Klinikum rechts der Isar, School of Medicine, Technical University of Munich, 81675 Munich, Germany
| | - Timothy C. Bates
- Department of Psychology, University of Edinburgh, Edinburgh EH8 9JZ, United Kingdom
| | - Manon Bernard
- Department of Physiology and Nutritional Sciences, University of Toronto, Toronto, ON M5S 1A1, Canada
| | - Kirsten Blokland
- Program in Neuroscience and Mental Health, Hospital for Sick Children, Toronto, M5G 1X8 ON, Canada
| | - Milene Bonte
- Department of Cognitive Neuroscience and Maastricht Brain Imaging Center, Faculty of Psychology and Neuroscience, Maastricht University, 6229 ER Maastricht, the Netherlands
| | - Anders D. Børglum
- Department of Biomedicine, Aarhus University, 8000 Aarhus, Denmark
- The Lundbeck Foundation Initiative for Integrative Psychiatric Research, iPSYCH, 8210 Aarhus, Denmark
- Center for Genomics and Personalized Medicine (CGPM), 8000 Aarhus, Denmark
| | - Thomas Bourgeron
- Human Genetics and Cognitive Functions, Institut Pasteur, UMR3571 Centre national de la recherche scientifique (CNRS), Université Paris Cité, Paris, 75015, France
| | - Daniel Brandeis
- Department of Child and Adolescent Psychiatry and Psychotherapy, Psychiatric Hospital, University of Zurich, 8032 Zurich, Switzerland
- Zurich Center for Integrative Human Physiology, University of Zurich and ETH Zurich, 8057 Zurich, Switzerland
- Neuroscience Center Zurich, University of Zurich and ETH Zurich, 8057 Zurich, Switzerland
- Department of Child and Adolescent Psychiatry and Psychotherapy, Central Institute of Mental Health, Medical Faculty Mannheim, Heidelberg University, 68159 Mannheim, Germany
| | - Fabiola Ceroni
- Department of Pharmacy and Biotechnology, University of Bologna, 40126 Bologna, Italy
- Faculty of Health and Life Sciences, Oxford Brookes University, Oxford OX3 0BP, United Kingdom
| | - Valéria Csépe
- Brain Imaging Centre, Research Centre for Natural Sciences, Budapest, 1117 Hungary
- Multilingualism Doctoral School, Faculty of Modern Philology and Social Sciences, University of Pannonia, Veszprém, 8200 Hungary
| | - Philip S. Dale
- Department of Speech & Hearing Sciences, University of New Mexico, Albuquerque, NM 87131
| | - Peter F. de Jong
- Department of Child Development and Education, University of Amsterdam, 1012 WX Amsterdam, the Netherlands
| | - John C. DeFries
- Institute for Behavioral Genetics, University of Colorado, Boulder, CO 80309-0447
- Department of Psychology and Neuroscience, University of Colorado, Boulder, CO 80309-0447
| | - Jean-François Démonet
- Leenaards Memory Centre, Department of Clinical Neurosciences, Lausanne University Hospital (CHUV), University of Lausanne, CH-1011 Lausanne, Switzerland
| | - Ditte Demontis
- Department of Biomedicine, Aarhus University, 8000 Aarhus, Denmark
- The Lundbeck Foundation Initiative for Integrative Psychiatric Research, iPSYCH, 8210 Aarhus, Denmark
| | - Yu Feng
- Division of Experimental and Translational Neuroscience, Krembil Research Institute, University Health Network, Toronto, ON M5T 0S8, Canada
| | - Scott D. Gordon
- Genetic Epidemiology Laboratory, QIMR Berghofer Medical Research Institute, Brisbane, QLD 4006, Australia
| | - Sharon L. Guger
- Department of Psychology, Hospital for Sick Children, Toronto, ON M5G 1X8, Canada
| | | | - Juan A. Hernández-Cabrera
- Departamento de Psicología, Clínica Psicobiología y Metodología, 38200, La Laguna, Santa Cruz de Tenerife, Spain
| | - Jouke-Jan Hottenga
- Department of Biological Psychology, Vrije Universiteit Amsterdam, 1081 BT Amsterdam, the Netherlands
| | - Charles Hulme
- Department of Education, University of Oxford, Oxford, Oxfordshire OX2 6PY, United Kingdom
| | - Juha Kere
- Department of Biosciences and Nutrition, Karolinska Institutet, 171 77 Stockholm, Sweden
- Stem Cells and Metabolism Research Program, University of Helsinki and Folkhälsan Research Center, 00014 Helsinki, Finland
| | - Elizabeth N. Kerr
- Department of Psychology, Hospital for Sick Children, Toronto, ON M5G 1X8, Canada
- Department of Neurology, Hospital for Sick Children, Toronto, ON M5G 1X8, Canada
- Department of Paediatrics, University of Toronto, Toronto, ON M5G 1X8, Canada
| | - Tanner Koomar
- Department of Psychiatry, University of Iowa, Iowa City, IA 52242
| | - Karin Landerl
- Institute of Psychology, University of Graz, 8010 Graz, Austria
- BioTechMed-Graz, 8010 Graz, Austria
| | - Gabriel T. Leonard
- Cognitive Neuroscience Neurology and Neurosurgery, McGill University, Montreal, QC H3A 1G1, Canada
| | - Maureen W. Lovett
- Program in Neuroscience and Mental Health, Hospital for Sick Children, Toronto, M5G 1X8 ON, Canada
- Department of Paediatrics, University of Toronto, Toronto, ON M5G 1X8, Canada
| | - Heikki Lyytinen
- Department of Psychology, University of Jyväskylä, 40014 Jyväskylä, Finland
| | - Nicholas G. Martin
- Genetic Epidemiology Laboratory, QIMR Berghofer Medical Research Institute, Brisbane, QLD 4006, Australia
| | - Angela Martinelli
- School of Medicine, University of St Andrews, KY16 9TF, St. Andrews, Scotland
| | - Urs Maurer
- Department of Psychology, The Chinese University of Hong Kong, Hong Kong, China
| | | | - Kristina Moll
- Department of Child and Adolescent Psychiatry, Psychosomatics, and Psychotherapy, Ludwig-Maximilians-University Hospital Munich, Munich, 80336 Germany
| | | | - Angela T. Morgan
- Speech and Language, Murdoch Children's Research Institute, Melbourne, VIC 3052, Australia
- Department of Audiology and Speech Pathology, University of Melbourne, Melbourne, VIC 3052, Australia
- Speech Pathology Department, Royal Children's Hospital, Melbourne, VIC 3052, Australia
| | - Markus M. Nöthen
- Institute of Human Genetics, University Hospital of Bonn, 53127 Bonn, Germany
| | - Zdenka Pausova
- Department of Physiology and Nutritional Sciences, University of Toronto, Toronto, ON M5S 1A1, Canada
- Hospital for Sick Children, Toronto, ON M5G 1X8, Canada
| | - Craig E. Pennell
- School of Medicine and Public Health, The University of Newcastle, Newcastle, NSW 2308, Australia
- Mothers and Babies Research Program, Hunter Medical Research Institute, Newcastle, NSW 2305, Australia
- Maternity and Gynaecology, John Hunter Hospital, Newcastle, NSW 2305, Australia
| | | | - Kaitlyn M. Price
- Division of Experimental and Translational Neuroscience, Krembil Research Institute, University Health Network, Toronto, ON M5T 0S8, Canada
- Program in Neuroscience and Mental Health, Hospital for Sick Children, Toronto, M5G 1X8 ON, Canada
- Department of Physiology, University of Toronto, Toronto, ON M5S 1A8, Canada
| | - Veera M. Rajagopal
- Department of Biomedicine, Aarhus University, 8000 Aarhus, Denmark
- The Lundbeck Foundation Initiative for Integrative Psychiatric Research, iPSYCH, 8210 Aarhus, Denmark
| | - Franck Ramus
- Laboratoire de Sciences Cognitives et Psycholinguistique, Ecole Normale Supérieure, Paris Sciences & Lettres University, École des Hautes Études en Sciences Sociales (EHESS), Centre National de la Recherche Scientifique (CNRS), Paris, 75005 France
| | - Louis Richer
- Department of Health Sciences, Université du Québec à Chicoutimi, Chicoutimi, QC G7H 2B1, Canada
| | - Nuala H. Simpson
- Department of Experimental Psychology, University of Oxford, Oxford OX2 6GG, United Kingdom
| | - Shelley D. Smith
- Department of Neurological Sciences, College of Medicine, University of Nebraska Medical Center, Omaha, NE 68198
| | - Margaret J. Snowling
- Department of Experimental Psychology, University of Oxford, Oxford OX2 6GG, United Kingdom
- St. John’s College, University of Oxford, Oxford OX1 3JP, United Kingdom
| | - John Stein
- Department of Physiology, Anatomy and Genetics, Oxford University, Oxford OX1 3PT, United Kingdom
| | - Lisa J. Strug
- Department of Statistical Sciences and Computer Science and Division of Biostatistics, University of Toronto, Toronto, ON M5S 3G3, Canada
- Program in Genetics and Genome Biology and the Centre for Applied Genomics, Hospital for Sick Children, Toronto, ON M5G 1X8, Canada
| | - Joel B. Talcott
- Institute for Health and Neurodevelopment, Aston University, Birmingham B4 7ET, United Kingdom
| | - Henning Tiemeier
- Department of Child and Adolescent Psychiatry/Psychology, Erasmus University Medical Center, 3000 CB Rotterdam, the Netherlands
- T. H. Chan School of Public Health, Harvard, Boston, MA 02115
| | - Marc P. van der Schroeff
- Department of Otolaryngology, Head and Neck Surgery, Erasmus University Medical Center, 3015 GD Rotterdam, the Netherlands
- Generation R Study Group, Erasmus University Medical Center, 3015 GD Rotterdam, the Netherlands
| | - Ellen Verhoef
- Language and Genetics Department, Max Planck Institute for Psycholinguistics, 6525 XD Nijmegen, the Netherlands
| | - Kate E. Watkins
- Department of Experimental Psychology, University of Oxford, Oxford OX2 6GG, United Kingdom
| | - Margaret Wilkinson
- Program in Neuroscience and Mental Health, Hospital for Sick Children, Toronto, M5G 1X8 ON, Canada
| | - Margaret J. Wright
- Queensland Brain Institute, University of Queensland, Brisbane, QLD 4072, Australia
| | - Cathy L. Barr
- Division of Experimental and Translational Neuroscience, Krembil Research Institute, University Health Network, Toronto, ON M5T 0S8, Canada
- Program in Neuroscience and Mental Health, Hospital for Sick Children, Toronto, M5G 1X8 ON, Canada
- Department of Physiology, University of Toronto, Toronto, ON M5S 1A8, Canada
| | - Dorret I. Boomsma
- Department of Biological Psychology, Vrije Universiteit Amsterdam, 1081 BT Amsterdam, the Netherlands
- Netherlands Twin Register, 1081 BT Amsterdam, the Netherlands
- Amsterdam Reproduction and Development Research Institute, Amsterdam University Medical Center, 1105 AZ Amsterdam, the Netherlands
| | - Manuel Carreiras
- Basque Center on Cognition, Brain and Language, Donostia-San Sebastian, 20009 Gipuzkoa, Spain
- Ikerbasque, Basque Foundation for Science, 48009 Bilbao, Vizcaya, Spain
- Lengua Vasca y Comunicación, University of the Basque Country, 48940 Bilbao, Vizcaya, Spain
| | - Marie-Christine J. Franken
- Department of Otolaryngology, Head and Neck Surgery, Erasmus University Medical Center, 3015 GD Rotterdam, the Netherlands
| | - Jeffrey R. Gruen
- Department of Pediatrics and Genetics, Yale Medical School, New Haven, CT 06510
| | - Michelle Luciano
- Department of Psychology, University of Edinburgh, Edinburgh EH8 9JZ, United Kingdom
| | - Bertram Müller-Myhsok
- Translational Research in Psychiatry, Max Planck Institute of Psychiatry, 80804 Munich, Germany
- Department of Health Science, University of Liverpool, Liverpool L69 7ZX, United Kingdom
| | - Dianne F. Newbury
- Faculty of Health and Life Sciences, Oxford Brookes University, Oxford OX3 0BP, United Kingdom
| | - Richard K. Olson
- Institute for Behavioral Genetics, University of Colorado, Boulder, CO 80309-0447
| | - Silvia Paracchini
- School of Medicine, University of St Andrews, KY16 9TF, St. Andrews, Scotland
| | - Tomáš Paus
- Department of Psychiatry and Neuroscience and Centre Hospitalier Universitaire Sainte Justine, University of Montreal, Montreal, QC H3T 1J4, Canada
| | - Robert Plomin
- Social, Genetic and Developmental Psychiatry Centre, Institute of Psychiatry, Psychology and Neuroscience, King’s College London, London SE5 8AF, United Kingdom
| | - Sheena Reilly
- Speech and Language, Murdoch Children's Research Institute, Melbourne, VIC 3052, Australia
- Menzies Health Institute Queensland, Griffith University, Gold Coast, QLD 4222, Australia
| | - Gerd Schulte-Körne
- Department of Child and Adolescent Psychiatry, Psychosomatics, and Psychotherapy, Ludwig-Maximilians-University Hospital Munich, Munich, 80336 Germany
| | - J. Bruce Tomblin
- Communication Sciences and Disorders, University of Iowa, Iowa City, IA 52242
| | - Elsje van Bergen
- Department of Biological Psychology, Vrije Universiteit Amsterdam, 1081 BT Amsterdam, the Netherlands
- Netherlands Twin Register, 1081 BT Amsterdam, the Netherlands
- Research Institute LEARN!, Vrije Universiteit Amsterdam, 1081 BT Amsterdam, the Netherlands
| | | | - Erik G. Willcutt
- Department of Psychology and Neuroscience, University of Colorado, Boulder, CO 80309-0447
| | - Beate St Pourcain
- Language and Genetics Department, Max Planck Institute for Psycholinguistics, 6525 XD Nijmegen, the Netherlands
- MRC Integrative Epidemiology Unit, University of Bristol, Bristol BS8 2BN, United Kingdom
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, 6525 EN Nijmegen, the Netherlands
| | - Clyde Francks
- Language and Genetics Department, Max Planck Institute for Psycholinguistics, 6525 XD Nijmegen, the Netherlands
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, 6525 EN Nijmegen, the Netherlands
- Department of Human Genetics, Radboud University Medical Center, 6525 GA Nijmegen, the Netherlands
| | - Simon E. Fisher
- Language and Genetics Department, Max Planck Institute for Psycholinguistics, 6525 XD Nijmegen, the Netherlands
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, 6525 EN Nijmegen, the Netherlands
| |
Collapse
|
12
|
Bernstein LE, Jordan N, Auer ET, Eberhardt SP. Lipreading: A Review of Its Continuing Importance for Speech Recognition With an Acquired Hearing Loss and Possibilities for Effective Training. Am J Audiol 2022; 31:453-469. [PMID: 35316072 DOI: 10.1044/2021_aja-21-00112] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022] Open
Abstract
PURPOSE The goal of this review article is to reinvigorate interest in lipreading and lipreading training for adults with acquired hearing loss. Most adults benefit from being able to see the talker when speech is degraded; however, the effect size is related to their lipreading ability, which is typically poor in adults who have experienced normal hearing through most of their lives. Lipreading training has been viewed as a possible avenue for rehabilitation of adults with an acquired hearing loss, but most training approaches have not been particularly successful. Here, we describe lipreading and theoretically motivated approaches to its training, as well as examples of successful training paradigms. We discuss some extensions to auditory-only (AO) and audiovisual (AV) speech recognition. METHOD Visual speech perception and word recognition are described. Traditional and contemporary views of training and perceptual learning are outlined. We focus on the roles of external and internal feedback and the training task in perceptual learning, and we describe results of lipreading training experiments. RESULTS Lipreading is commonly characterized as limited to viseme perception. However, evidence demonstrates subvisemic perception of visual phonetic information. Lipreading words also relies on lexical constraints, not unlike auditory spoken word recognition. Lipreading has been shown to be difficult to improve through training, but under specific feedback and task conditions, training can be successful, and learning can generalize to untrained materials, including AV sentence stimuli in noise. The results on lipreading have implications for AO and AV training and for use of acoustically processed speech in face-to-face communication. CONCLUSION Given its importance for speech recognition with a hearing loss, we suggest that the research and clinical communities integrate lipreading in their efforts to improve speech recognition in adults with acquired hearing loss.
Collapse
Affiliation(s)
- Lynne E. Bernstein
- Department of Speech, Language & Hearing Sciences, George Washington University, Washington, DC
| | - Nicole Jordan
- Department of Speech, Language & Hearing Sciences, George Washington University, Washington, DC
| | - Edward T. Auer
- Department of Speech, Language & Hearing Sciences, George Washington University, Washington, DC
| | - Silvio P. Eberhardt
- Department of Speech, Language & Hearing Sciences, George Washington University, Washington, DC
| |
Collapse
|
13
|
Holmes S, Mar'i J, Simons LE, Zurakowski D, LeBel AA, O'Brien M, Borsook D. Integrated Features for Optimizing Machine Learning Classifiers of Pediatric and Young Adults With a Post-Traumatic Headache From Healthy Controls. FRONTIERS IN PAIN RESEARCH 2022; 3:859881. [PMID: 35655747 PMCID: PMC9152124 DOI: 10.3389/fpain.2022.859881] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/21/2022] [Accepted: 03/09/2022] [Indexed: 11/14/2022] Open
Abstract
Post-traumatic headache (PTH) is a challenging clinical condition to identify and treat as it integrates multiple subjectively defined symptoms with underlying physiological processes. The precise mechanisms underlying PTH are unclear, and it remains to be understood how to integrate the patient experience with underlying biology when attempting to classify persons with PTH, particularly in the pediatric setting where patient self-report may be highly variable. The objective of this investigation was to evaluate the use of different machine learning (ML) classifiers to differentiate pediatric and young adult subjects with PTH from healthy controls using behavioral data from self-report questionnaires that reflect concussion symptoms, mental health, pain experience of the participants, and structural brain imaging from cortical and sub-cortical locations. Behavioral data, alongside brain imaging, survived data reduction methods and both contributed toward final models. Behavioral data that contributed towards the final model included both the child and parent perspective of the pain-experience. Brain imaging features produced two unique clusters that reflect regions that were previously found in mild traumatic brain injury (mTBI) and PTH. Affinity-based propagation analysis demonstrated that behavioral data remained independent relative to neuroimaging data that suggest there is a role for both behavioral and brain imaging data when attempting to classify children with PTH.
Collapse
Affiliation(s)
- Scott Holmes
- Pediatric Pain Pathway Lab, Department of Anesthesia, Critical Care, and Pain Medicine, Boston Children's Hospital – Harvard Medical School, Boston, MA, United States
- Pain and Affective Neuroscience Center, Boston Children's Hospital, Boston, MA, United States
- *Correspondence: Scott Holmes
| | - Joud Mar'i
- Pediatric Pain Pathway Lab, Department of Anesthesia, Critical Care, and Pain Medicine, Boston Children's Hospital – Harvard Medical School, Boston, MA, United States
| | - Laura E. Simons
- Department of Anesthesiology, Perioperative, and Pain Medicine, Stanford University School of Medicine, Palo Alto, CA, United States
| | - David Zurakowski
- Department of Anesthesia, Critical Care, and Pain Medicine, Boston Children's Hospital, Boston, MA, United States
| | - Alyssa Ann LeBel
- Department of Anesthesia, Critical Care, and Pain Medicine, Boston Children's Hospital, Boston, MA, United States
| | - Michael O'Brien
- Sports Medicine Division, Sports Concussion Clinic, Orthopedic Surgery, Harvard Medical School, Boston, MA, United States
| | - David Borsook
- Departments of Psychiatry ad Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, MA, United States
| |
Collapse
|
14
|
Matsuzaki J, Kagitani-Shimono K, Aoki S, Hanaie R, Kato Y, Nakanishi M, Tatsumi A, Tominaga K, Yamamoto T, Nagai Y, Mohri I, Taniike M. Abnormal cortical responses elicited by audiovisual movies in patients with autism spectrum disorder with atypical sensory behavior: A magnetoencephalographic study. Brain Dev 2022; 44:81-94. [PMID: 34563417 DOI: 10.1016/j.braindev.2021.08.007] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/03/2021] [Revised: 08/09/2021] [Accepted: 08/30/2021] [Indexed: 12/23/2022]
Abstract
BACKGROUND Atypical sensory behavior disrupts behavioral adaptation in children with autism spectrum disorder (ASD); however, neural correlates of sensory dysfunction using magnetoencephalography (MEG) remain unclear. METHOD We used MEG to measure the cortical activation elicited by visual (uni)/audiovisual (multisensory) movies in 46 children (7-14 years) were included in final analysis: 13 boys with atypical audiovisual behavior in ASD (AAV+), 10 without this condition, and 23 age-matched typically developing boys. RESULTS The AAV+ group demonstrated an increase in the cortical activation in the bilateral insula in response to unisensory movies and in the left occipital, right superior temporal sulcus (rSTS), and temporal regions to multisensory movies. These increased responses were correlated with severity of the sensory impairment. Increased theta-low gamma oscillations were observed in the rSTS in AAV+. CONCLUSION The findings suggest that AAV is attributed to atypical neural networks centered on the rSTS.
Collapse
Affiliation(s)
- Junko Matsuzaki
- Division of Developmental Neuroscience, Department of Child Development, United Graduate School of Child Development, Osaka University, Osaka, Japan; Molecular Research Center for Children's Mental Development, Osaka University Graduate School of Medicine, Osaka, Japan
| | - Kuriko Kagitani-Shimono
- Division of Developmental Neuroscience, Department of Child Development, United Graduate School of Child Development, Osaka University, Osaka, Japan; Molecular Research Center for Children's Mental Development, Osaka University Graduate School of Medicine, Osaka, Japan; Department of Pediatrics, Osaka University Graduate School of Medicine, Osaka, Japan.
| | - Sho Aoki
- Division of Developmental Neuroscience, Department of Child Development, United Graduate School of Child Development, Osaka University, Osaka, Japan
| | - Ryuzo Hanaie
- Molecular Research Center for Children's Mental Development, Osaka University Graduate School of Medicine, Osaka, Japan
| | - Yoko Kato
- Division of Developmental Neuroscience, Department of Child Development, United Graduate School of Child Development, Osaka University, Osaka, Japan
| | - Mariko Nakanishi
- Molecular Research Center for Children's Mental Development, Osaka University Graduate School of Medicine, Osaka, Japan
| | - Aika Tatsumi
- Molecular Research Center for Children's Mental Development, Osaka University Graduate School of Medicine, Osaka, Japan
| | - Koji Tominaga
- Division of Developmental Neuroscience, Department of Child Development, United Graduate School of Child Development, Osaka University, Osaka, Japan; Department of Pediatrics, Osaka University Graduate School of Medicine, Osaka, Japan
| | - Tomoka Yamamoto
- Molecular Research Center for Children's Mental Development, Osaka University Graduate School of Medicine, Osaka, Japan
| | - Yukie Nagai
- International Research Center for Neurointelligence, The University of Tokyo, Tokyo, Japan
| | - Ikuko Mohri
- Division of Developmental Neuroscience, Department of Child Development, United Graduate School of Child Development, Osaka University, Osaka, Japan; Molecular Research Center for Children's Mental Development, Osaka University Graduate School of Medicine, Osaka, Japan; Department of Pediatrics, Osaka University Graduate School of Medicine, Osaka, Japan
| | - Masako Taniike
- Division of Developmental Neuroscience, Department of Child Development, United Graduate School of Child Development, Osaka University, Osaka, Japan; Molecular Research Center for Children's Mental Development, Osaka University Graduate School of Medicine, Osaka, Japan; Department of Pediatrics, Osaka University Graduate School of Medicine, Osaka, Japan
| |
Collapse
|
15
|
Peelle JE, Spehar B, Jones MS, McConkey S, Myerson J, Hale S, Sommers MS, Tye-Murray N. Increased Connectivity among Sensory and Motor Regions during Visual and Audiovisual Speech Perception. J Neurosci 2022; 42:435-442. [PMID: 34815317 PMCID: PMC8802926 DOI: 10.1523/jneurosci.0114-21.2021] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2021] [Revised: 10/29/2021] [Accepted: 11/08/2021] [Indexed: 11/21/2022] Open
Abstract
In everyday conversation, we usually process the talker's face as well as the sound of the talker's voice. Access to visual speech information is particularly useful when the auditory signal is degraded. Here, we used fMRI to monitor brain activity while adult humans (n = 60) were presented with visual-only, auditory-only, and audiovisual words. The audiovisual words were presented in quiet and in several signal-to-noise ratios. As expected, audiovisual speech perception recruited both auditory and visual cortex, with some evidence for increased recruitment of premotor cortex in some conditions (including in substantial background noise). We then investigated neural connectivity using psychophysiological interaction analysis with seed regions in both primary auditory cortex and primary visual cortex. Connectivity between auditory and visual cortices was stronger in audiovisual conditions than in unimodal conditions, including a wide network of regions in posterior temporal cortex and prefrontal cortex. In addition to whole-brain analyses, we also conducted a region-of-interest analysis on the left posterior superior temporal sulcus (pSTS), implicated in many previous studies of audiovisual speech perception. We found evidence for both activity and effective connectivity in pSTS for visual-only and audiovisual speech, although these were not significant in whole-brain analyses. Together, our results suggest a prominent role for cross-region synchronization in understanding both visual-only and audiovisual speech that complements activity in integrative brain regions like pSTS.SIGNIFICANCE STATEMENT In everyday conversation, we usually process the talker's face as well as the sound of the talker's voice. Access to visual speech information is particularly useful when the auditory signal is hard to understand (e.g., background noise). Prior work has suggested that specialized regions of the brain may play a critical role in integrating information from visual and auditory speech. Here, we show a complementary mechanism relying on synchronized brain activity among sensory and motor regions may also play a critical role. These findings encourage reconceptualizing audiovisual integration in the context of coordinated network activity.
Collapse
Affiliation(s)
- Jonathan E Peelle
- Department of Otolaryngology, Washington University in St. Louis, St. Louis, Missouri 63110
| | - Brent Spehar
- Department of Otolaryngology, Washington University in St. Louis, St. Louis, Missouri 63110
| | - Michael S Jones
- Department of Otolaryngology, Washington University in St. Louis, St. Louis, Missouri 63110
| | - Sarah McConkey
- Department of Otolaryngology, Washington University in St. Louis, St. Louis, Missouri 63110
| | - Joel Myerson
- Department of Psychological and Brain Sciences, Washington University in St. Louis, St. Louis, Missouri 63130
| | - Sandra Hale
- Department of Psychological and Brain Sciences, Washington University in St. Louis, St. Louis, Missouri 63130
| | - Mitchell S Sommers
- Department of Psychological and Brain Sciences, Washington University in St. Louis, St. Louis, Missouri 63130
| | - Nancy Tye-Murray
- Department of Otolaryngology, Washington University in St. Louis, St. Louis, Missouri 63110
| |
Collapse
|
16
|
Ylinen A, Wikman P, Leminen M, Alho K. Task-dependent cortical activations during selective attention to audiovisual speech. Brain Res 2022; 1775:147739. [PMID: 34843702 DOI: 10.1016/j.brainres.2021.147739] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2021] [Revised: 10/21/2021] [Accepted: 11/21/2021] [Indexed: 11/28/2022]
Abstract
Selective listening to speech depends on widespread networks of the brain, but how the involvement of different neural systems in speech processing is affected by factors such as the task performed by a listener and speech intelligibility remains poorly understood. We used functional magnetic resonance imaging to systematically examine the effects that performing different tasks has on neural activations during selective attention to continuous audiovisual speech in the presence of task-irrelevant speech. Participants viewed audiovisual dialogues and attended either to the semantic or the phonological content of speech, or ignored speech altogether and performed a visual control task. The tasks were factorially combined with good and poor auditory and visual speech qualities. Selective attention to speech engaged superior temporal regions and the left inferior frontal gyrus regardless of the task. Frontoparietal regions implicated in selective auditory attention to simple sounds (e.g., tones, syllables) were not engaged by the semantic task, suggesting that this network may not be not as crucial when attending to continuous speech. The medial orbitofrontal cortex, implicated in social cognition, was most activated by the semantic task. Activity levels during the phonological task in the left prefrontal, premotor, and secondary somatosensory regions had a distinct temporal profile as well as the highest overall activity, possibly relating to the role of the dorsal speech processing stream in sub-lexical processing. Our results demonstrate that the task type influences neural activations during selective attention to speech, and emphasize the importance of ecologically valid experimental designs.
Collapse
Affiliation(s)
- Artturi Ylinen
- Department of Psychology and Logopedics, University of Helsinki, Helsinki, Finland.
| | - Patrik Wikman
- Department of Psychology and Logopedics, University of Helsinki, Helsinki, Finland; Department of Neuroscience, Georgetown University, Washington D.C., USA
| | - Miika Leminen
- Analytics and Data Services, HUS Helsinki University Hospital, Helsinki, Finland
| | - Kimmo Alho
- Department of Psychology and Logopedics, University of Helsinki, Helsinki, Finland; Advanced Magnetic Imaging Centre, Aalto NeuroImaging, Aalto University, Espoo, Finland
| |
Collapse
|
17
|
Kaganovich N, Christ S. Event-related potentials evidence for long-term audiovisual representations of phonemes in adults. Eur J Neurosci 2021; 54:7860-7875. [PMID: 34750895 DOI: 10.1111/ejn.15519] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2021] [Revised: 11/03/2021] [Accepted: 11/05/2021] [Indexed: 10/19/2022]
Abstract
The presence of long-term auditory representations for phonemes has been well-established. However, since speech perception is typically audiovisual, we hypothesized that long-term phoneme representations may also contain information on speakers' mouth shape during articulation. We used an audiovisual oddball paradigm in which, on each trial, participants saw a face and heard one of two vowels. One vowel occurred frequently (standard), while another occurred rarely (deviant). In one condition (neutral), the face had a closed, non-articulating mouth. In the other condition (audiovisual violation), the mouth shape matched the frequent vowel. Although in both conditions stimuli were audiovisual, we hypothesized that identical auditory changes would be perceived differently by participants. Namely, in the neutral condition, deviants violated only the audiovisual pattern specific to each block. By contrast, in the audiovisual violation condition, deviants additionally violated long-term representations for how a speaker's mouth looks during articulation. We compared the amplitude of mismatch negativity (MMN) and P3 components elicited by deviants in the two conditions. The MMN extended posteriorly over temporal and occipital sites even though deviants contained no visual changes, suggesting that deviants were perceived as interruptions in audiovisual, rather than auditory only, sequences. As predicted, deviants elicited larger MMN and P3 in the audiovisual violation compared to the neutral condition. The results suggest that long-term representations of phonemes are indeed audiovisual.
Collapse
Affiliation(s)
- Natalya Kaganovich
- Department of Speech, Language, and Hearing Sciences, Purdue University, West Lafayette, Indiana, USA.,Department of Psychological Sciences, Purdue University, West Lafayette, Indiana, USA
| | - Sharon Christ
- Department of Human Development and Family Studies, Purdue University, West Lafayette, Indiana, USA.,Department of Statistics, Purdue University, West Lafayette, Indiana, USA
| |
Collapse
|
18
|
Auditory cortical micro-networks show differential connectivity during voice and speech processing in humans. Commun Biol 2021; 4:801. [PMID: 34172824 PMCID: PMC8233416 DOI: 10.1038/s42003-021-02328-2] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2021] [Accepted: 06/09/2021] [Indexed: 02/05/2023] Open
Abstract
The temporal voice areas (TVAs) in bilateral auditory cortex (AC) appear specialized for voice processing. Previous research assumed a uniform functional profile for the TVAs which are broadly spread along the bilateral AC. Alternatively, the TVAs might comprise separate AC nodes controlling differential neural functions for voice and speech decoding, organized as local micro-circuits. To investigate micro-circuits, we modeled the directional connectivity between TVA nodes during voice processing in humans while acquiring brain activity using neuroimaging. Results show several bilateral AC nodes for general voice decoding (speech and non-speech voices) and for speech decoding in particular. Furthermore, non-hierarchical and differential bilateral AC networks manifest distinct excitatory and inhibitory pathways for voice and speech processing. Finally, while voice and speech processing seem to have distinctive but integrated neural circuits in the left AC, the right AC reveals disintegrated neural circuits for both sounds. Altogether, we demonstrate a functional heterogeneity in the TVAs for voice decoding based on local micro-circuits.
Collapse
|
19
|
Neural correlates of the production effect: An fMRI study. Brain Cogn 2021; 152:105757. [PMID: 34130081 DOI: 10.1016/j.bandc.2021.105757] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2020] [Revised: 05/16/2021] [Accepted: 05/18/2021] [Indexed: 11/20/2022]
Abstract
Recognition memory is improved for items produced at study (e.g., by reading them aloud) relative to a non-produced control condition (e.g., silent reading). This production effect is typically attributed to the extra elements in the production task (e.g., motor activation, auditory perception) enhancing item distinctiveness. To evaluate this claim, the present study examined the neural mechanisms underlying the production effect. Prior to a recognition memory test, different words within a study list were read either aloud, silently, or while saying "check" (as a sensorimotor control condition). Production improved recognition, and aloud words yielded higher rates of both recollection and familiarity judgments than either silent or control words. During encoding, fMRI revealed stronger activation in regions associated with motor, somatosensory, and auditory processing for aloud items than for either silent or control items. These activations were predictive of recollective success for aloud items at test. Together, our findings are compatible with a distinctiveness-based account of the production effect, while also pointing to the possible role of other processing differences during the aloud trials as compared to silent and control.
Collapse
|
20
|
Wiesman AI, Christopher-Hayes NJ, Wilson TW. Stairway to memory: Left-hemispheric alpha dynamics index the progressive loading of items into a short-term store. Neuroimage 2021; 235:118024. [PMID: 33836267 PMCID: PMC8354033 DOI: 10.1016/j.neuroimage.2021.118024] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2020] [Revised: 03/25/2021] [Accepted: 03/27/2021] [Indexed: 01/29/2023] Open
Abstract
The encoding, maintenance, and subsequent retrieval of memories over short time intervals is an essential cognitive function. Load effects on the neural dynamics supporting the maintenance of short-term memories have been well studied, but experimental design limitations have hindered the study of similar effects during the encoding of information into online memory stores. Theoretically, the active encoding of complex visual stimuli into memory must also recruit neural resources in a manner that scales with memory load. Understanding the neural systems supporting this encoding load effect is of particular importance, as some patient populations exhibit difficulties specifically with the encoding, and not the maintenance, of short-term memories. Using magnetoencephalography, a visual sequence memory paradigm, and a novel encoding slope analysis, we provide evidence for a left-lateralized network of regions, oscillating in the alpha frequency range, that exhibit a progressive loading effect of complex visual stimulus information during memory encoding. This progressive encoding load effect significantly tracked the eventual retrieval of item-order memories at the single trial level, and neural activity in these regions was functionally dissociated from that of earlier visual networks. These findings suggest that the active encoding of stimulus information into short-term stores recruits a left-lateralized network of frontal, parietal, and temporal regions, and might be susceptible to modulation (e.g., using non-invasive stimulation) in the alpha band.
Collapse
Affiliation(s)
- Alex I Wiesman
- College of Medicine, University of Nebraska Medical Center, Omaha 68198-8422, NE, United States.
| | - Nicholas J Christopher-Hayes
- College of Medicine, University of Nebraska Medical Center, Omaha 68198-8422, NE, United States; Institute for Human Neuroscience, Boys Town National Research Hospital, Omaha, NE, United States
| | - Tony W Wilson
- College of Medicine, University of Nebraska Medical Center, Omaha 68198-8422, NE, United States; Institute for Human Neuroscience, Boys Town National Research Hospital, Omaha, NE, United States
| |
Collapse
|
21
|
Walker GM, Rollo PS, Tandon N, Hickok G. Effect of Bilateral Opercular Syndrome on Speech Perception. NEUROBIOLOGY OF LANGUAGE (CAMBRIDGE, MASS.) 2021; 2:335-353. [PMID: 37213256 PMCID: PMC10158595 DOI: 10.1162/nol_a_00037] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/28/2020] [Accepted: 03/23/2021] [Indexed: 05/23/2023]
Abstract
Speech perception ability and structural neuroimaging were investigated in two cases of bilateral opercular syndrome. Due to bilateral ablation of the motor control center for the lower face and surrounds, these rare cases provide an opportunity to evaluate the necessity of cortical motor representations for speech perception, a cornerstone of some neurocomputational theories of language processing. Speech perception, including audiovisual integration (i.e., the McGurk effect), was mostly unaffected in these cases, although verbal short-term memory impairment hindered performance on several tasks that are traditionally used to evaluate speech perception. The results suggest that the role of the cortical motor system in speech perception is context-dependent and supplementary, not inherent or necessary.
Collapse
Affiliation(s)
- Grant M. Walker
- Department of Cognitive Sciences, University of California, Irvine
- * Corresponding Author:
| | | | - Nitin Tandon
- Department of Neurosurgery, University of Texas Medical School at Houston
| | - Gregory Hickok
- Department of Cognitive Sciences, University of California, Irvine
- Department of Language Science, University of California, Irvine
| |
Collapse
|
22
|
Audio-visual combination of syllables involves time-sensitive dynamics following from fusion failure. Sci Rep 2020; 10:18009. [PMID: 33093570 PMCID: PMC7583249 DOI: 10.1038/s41598-020-75201-7] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2020] [Accepted: 10/05/2020] [Indexed: 11/08/2022] Open
Abstract
In face-to-face communication, audio-visual (AV) stimuli can be fused, combined or perceived as mismatching. While the left superior temporal sulcus (STS) is presumably the locus of AV integration, the process leading to combination is unknown. Based on previous modelling work, we hypothesize that combination results from a complex dynamic originating in a failure to integrate AV inputs, followed by a reconstruction of the most plausible AV sequence. In two different behavioural tasks and one MEG experiment, we observed that combination is more time demanding than fusion. Using time-/source-resolved human MEG analyses with linear and dynamic causal models, we show that both fusion and combination involve early detection of AV incongruence in the STS, whereas combination is further associated with enhanced activity of AV asynchrony-sensitive regions (auditory and inferior frontal cortices). Based on neural signal decoding, we finally show that only combination can be decoded from the IFG activity and that combination is decoded later than fusion in the STS. These results indicate that the AV speech integration outcome primarily depends on whether the STS converges or not onto an existing multimodal syllable representation, and that combination results from subsequent temporal processing, presumably the off-line re-ordering of incongruent AV stimuli.
Collapse
|
23
|
Nourski KV, Steinschneider M, Rhone AE, Kovach CK, Banks MI, Krause BM, Kawasaki H, Howard MA. Electrophysiology of the Human Superior Temporal Sulcus during Speech Processing. Cereb Cortex 2020; 31:1131-1148. [PMID: 33063098 DOI: 10.1093/cercor/bhaa281] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/18/2020] [Revised: 08/06/2020] [Accepted: 09/01/2020] [Indexed: 12/20/2022] Open
Abstract
The superior temporal sulcus (STS) is a crucial hub for speech perception and can be studied with high spatiotemporal resolution using electrodes targeting mesial temporal structures in epilepsy patients. Goals of the current study were to clarify functional distinctions between the upper (STSU) and the lower (STSL) bank, hemispheric asymmetries, and activity during self-initiated speech. Electrophysiologic properties were characterized using semantic categorization and dialog-based tasks. Gamma-band activity and alpha-band suppression were used as complementary measures of STS activation. Gamma responses to auditory stimuli were weaker in STSL compared with STSU and had longer onset latencies. Activity in anterior STS was larger during speaking than listening; the opposite pattern was observed more posteriorly. Opposite hemispheric asymmetries were found for alpha suppression in STSU and STSL. Alpha suppression in the STS emerged earlier than in core auditory cortex, suggesting feedback signaling within the auditory cortical hierarchy. STSL was the only region where gamma responses to words presented in the semantic categorization tasks were larger in subjects with superior task performance. More pronounced alpha suppression was associated with better task performance in Heschl's gyrus, superior temporal gyrus, and STS. Functional differences between STSU and STSL warrant their separate assessment in future studies.
Collapse
Affiliation(s)
- Kirill V Nourski
- Department of Neurosurgery, The University of Iowa, Iowa City, IA 52242, USA.,Iowa Neuroscience Institute, The University of Iowa, Iowa City, IA 52242, USA
| | - Mitchell Steinschneider
- Departments of Neurology and Neuroscience, Albert Einstein College of Medicine, Bronx, NY 10461, USA
| | - Ariane E Rhone
- Department of Neurosurgery, The University of Iowa, Iowa City, IA 52242, USA
| | | | - Matthew I Banks
- Department of Anesthesiology, University of Wisconsin-Madison, Madison, WI 53705, USA.,Department of Neuroscience, University of Wisconsin-Madison, Madison, WI 53705, USA
| | - Bryan M Krause
- Department of Anesthesiology, University of Wisconsin-Madison, Madison, WI 53705, USA
| | - Hiroto Kawasaki
- Department of Neurosurgery, The University of Iowa, Iowa City, IA 52242, USA
| | - Matthew A Howard
- Department of Neurosurgery, The University of Iowa, Iowa City, IA 52242, USA.,Iowa Neuroscience Institute, The University of Iowa, Iowa City, IA 52242, USA.,Pappajohn Biomedical Institute, The University of Iowa, Iowa City, IA 52242, USA
| |
Collapse
|
24
|
Scurry AN, Huber E, Matera C, Jiang F. Increased Right Posterior STS Recruitment Without Enhanced Directional-Tuning During Tactile Motion Processing in Early Deaf Individuals. Front Neurosci 2020; 14:864. [PMID: 32982667 PMCID: PMC7477335 DOI: 10.3389/fnins.2020.00864] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/07/2020] [Accepted: 07/24/2020] [Indexed: 01/19/2023] Open
Abstract
Upon early sensory deprivation, the remaining modalities often exhibit cross-modal reorganization, such as primary auditory cortex (PAC) recruitment for visual motion processing in early deafness (ED). Previous studies of compensatory plasticity in ED individuals have given less attention to tactile motion processing. In the current study, we aimed to examine the effects of early auditory deprivation on tactile motion processing. We simulated four directions of tactile motion on each participant's right index finger and characterized their tactile motion responses and directional-tuning profiles using population receptive field analysis. Similar tactile motion responses were found within primary (SI) and secondary (SII) somatosensory cortices between ED and hearing control groups, whereas ED individuals showed a reduced proportion of voxels with directionally tuned responses in SI contralateral to stimulation. There were also significant but minimal responses to tactile motion within PAC for both groups. While early deaf individuals show significantly larger recruitment of right posterior superior temporal sulcus (pSTS) region upon tactile motion stimulation, there was no evidence of enhanced directional tuning. Greater recruitment of right pSTS region is consistent with prior studies reporting reorganization of multimodal areas due to sensory deprivation. The absence of increased directional tuning within the right pSTS region may suggest a more distributed population of neurons dedicated to processing tactile spatial information as a consequence of early auditory deprivation.
Collapse
Affiliation(s)
- Alexandra N Scurry
- Department of Psychology, University of Nevada, Reno, Reno, NV, United States
| | - Elizabeth Huber
- Department of Speech and Hearing Sciences, Institute for Learning & Brain Sciences, University of Washington, Seattle, WA, United States
| | - Courtney Matera
- Department of Psychology, University of Nevada, Reno, Reno, NV, United States
| | - Fang Jiang
- Department of Psychology, University of Nevada, Reno, Reno, NV, United States
| |
Collapse
|
25
|
Pujol J, Blanco-Hinojo L, Martínez-Vilavella G, Canu-Martín L, Pujol A, Pérez-Sola V, Deus J. Brain activity during traditional textbook and audiovisual-3D learning. Brain Behav 2019; 9:e01427. [PMID: 31571423 PMCID: PMC6790317 DOI: 10.1002/brb3.1427] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/15/2019] [Revised: 09/09/2019] [Accepted: 09/10/2019] [Indexed: 12/21/2022] Open
Abstract
INTRODUCTION Audiovisual educational tools have increasingly been used during the past years to complement and compete with traditional textbooks. However, little is known as to how the brain processes didactic information presented in different formats. We directly assessed brain activity during learning using both traditional textbook and audiovisual-3D material. METHODS A homogeneous sample of 30 young adults with active study habits was assessed. Educational material on the subject of Cardiology was adapted to be presented during the acquisition of functional MRI. RESULTS When tested after image acquisition, participants obtained similar examination scores for both formats. Evoked brain activity was robust during both traditional textbook and audiovisual-3D lessons, but a greater number of brain systems were implicated in the processing of audiovisual-3D information, consistent with its multisource sensory nature. However, learning was not associated with group mean brain activations, but was instead predicted by distinct functional MRI signal changes in the frontal lobes and showed distinct cognitive correlates. In the audiovisual-3D version, examination scores were positively correlated with late-evoked prefrontal cortex activity and working memory, and negatively correlated with language-related frontal areas and verbal memory. As for the traditional textbook version, the fewer results obtained suggested the opposite pattern, with examination scores negatively correlating with prefrontal cortex activity evoked during the lesson. CONCLUSIONS Overall, the results indicate that a similar level of knowledge may be achieved via different cognitive strategies. In our experiment, audiovisual learning appeared to benefit from prefrontal executive resources (as opposed to memorizing verbal information) more than traditional textbook learning.
Collapse
Affiliation(s)
- Jesus Pujol
- MRI Research Unit, Department of Radiology, Hospital del Mar, Barcelona, Spain.,Centro Investigación Biomédica en Red de Salud Mental, CIBERSAM G21, Barcelona, Spain
| | - Laura Blanco-Hinojo
- MRI Research Unit, Department of Radiology, Hospital del Mar, Barcelona, Spain.,Centro Investigación Biomédica en Red de Salud Mental, CIBERSAM G21, Barcelona, Spain
| | | | - Lucila Canu-Martín
- MRI Research Unit, Department of Radiology, Hospital del Mar, Barcelona, Spain
| | - Anna Pujol
- MRI Research Unit, Department of Radiology, Hospital del Mar, Barcelona, Spain
| | - Víctor Pérez-Sola
- Centro Investigación Biomédica en Red de Salud Mental, CIBERSAM G21, Barcelona, Spain.,Institute of Neuropsychiatry and Addictions, Hospital del Mar, IMIM, Barcelona, Spain
| | - Joan Deus
- MRI Research Unit, Department of Radiology, Hospital del Mar, Barcelona, Spain.,Department of Psychobiology and Methodology in Health Sciences, Autonomous University of Barcelona, Barcelona, Spain
| |
Collapse
|
26
|
Retter TL, Webster MA, Jiang F. Directional Visual Motion Is Represented in the Auditory and Association Cortices of Early Deaf Individuals. J Cogn Neurosci 2019; 31:1126-1140. [PMID: 30726181 PMCID: PMC6599583 DOI: 10.1162/jocn_a_01378] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Individuals who are deaf since early life may show enhanced performance at some visual tasks, including discrimination of directional motion. The neural substrates of such behavioral enhancements remain difficult to identify in humans, although neural plasticity has been shown for early deaf people in the auditory and association cortices, including the primary auditory cortex (PAC) and STS region, respectively. Here, we investigated whether neural responses in auditory and association cortices of early deaf individuals are reorganized to be sensitive to directional visual motion. To capture direction-selective responses, we recorded fMRI responses frequency-tagged to the 0.1-Hz presentation of central directional (100% coherent random dot) motion persisting for 2 sec contrasted with nondirectional (0% coherent) motion for 8 sec. We found direction-selective responses in the STS region in both deaf and hearing participants, but the extent of activation in the right STS region was 5.5 times larger for deaf participants. Minimal but significant direction-selective responses were also found in the PAC of deaf participants, both at the group level and in five of six individuals. In response to stimuli presented separately in the right and left visual fields, the relative activation across the right and left hemispheres was similar in both the PAC and STS region of deaf participants. Notably, the enhanced right-hemisphere activation could support the right visual field advantage reported previously in behavioral studies. Taken together, these results show that the reorganized auditory cortices of early deaf individuals are sensitive to directional motion. Speculatively, these results suggest that auditory and association regions can be remapped to support enhanced visual performance.
Collapse
|
27
|
Ching ASM, Kim J, Davis C. Auditory-visual integration during nonconscious perception. Cortex 2019; 117:1-15. [PMID: 30925308 DOI: 10.1016/j.cortex.2019.02.014] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2018] [Revised: 02/11/2019] [Accepted: 02/12/2019] [Indexed: 10/27/2022]
Abstract
Our study proposes a test of a key assumption of the most prominent model of consciousness - the global workspace (GWS) model (e.g., Baars, 2002, 2005, 2007; Dehaene & Naccache, 2001; Mudrik, Faivre, & Koch, 2014). This assumption is that multimodal integration requires consciousness; however, few studies have explicitly tested if integration can occur between nonconscious information from different modalities. The proposed study examined whether a classic indicator of multimodal integration - the McGurk effect - can be elicited with subliminal auditory-visual speech stimuli. We used a masked speech priming paradigm developed by Kouider and Dupoux (2005) in conjunction with continuous flash suppression (CFS; Tsuchiya & Koch, 2005), a binocular rivalry technique for presenting video stimuli subliminally. Applying these techniques together, we carried out two experiments in which participants categorised auditory syllable targets which were preceded by subliminal auditory-visual (AV) speech primes. Subliminal AV primes were either illusion-inducing (McGurk) or illusion-neutral (Incongruent) combinations of speech stimuli. In Experiment 1, the categorisation of the syllable target ("pa") was facilitated by the same syllable prime when it was part of a McGurk combination (auditory "pa" and visual "ka") but not when part of an Incongruent combination (auditory "pa" and visual "wa"). This dependency on specific AV combinations indicated a nonconscious AV interaction. Experiment 2 presented a different syllable target ("ta") which matched the predicted illusory outcome of the McGurk combination - here, both the McGurk combination (auditory "pa" and visual "ka") and the Incongruent combination (auditory "ta" and visual "ka") failed to facilitate target categorisation. The combined results of both Experiments demonstrate a type of nonconscious multimodal interaction that is distinct from integration - it allows unimodal information that is compatible for integration (i.e., McGurk combinations) to persist and influence later processes, but does not actually combine and alter that information. As the GWS model does not account for non-integrative multimodal interactions, this places some pressure on such models of consciousness.
Collapse
Affiliation(s)
| | - Jeesun Kim
- The MARCS Institute, Western Sydney University, Australia
| | - Chris Davis
- The MARCS Institute, Western Sydney University, Australia
| |
Collapse
|
28
|
Improving Human–Computer Interface Design through Application of Basic Research on Audiovisual Integration and Amplitude Envelope. MULTIMODAL TECHNOLOGIES AND INTERACTION 2019. [DOI: 10.3390/mti3010004] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
Quality care for patients requires effective communication amongst medical teams. Increasingly, communication is required not only between team members themselves, but between members and the medical devices monitoring and managing patient well-being. Most human–computer interfaces use either auditory or visual displays, and despite significant experimentation, they still elicit well-documented concerns. Curiously, few interfaces explore the benefits of multimodal communication, despite extensive documentation of the brain’s sensitivity to multimodal signals. New approaches built on insights from basic audiovisual integration research hold the potential to improve future human–computer interfaces. In particular, recent discoveries regarding the acoustic property of amplitude envelope illustrate that it can enhance audiovisual integration while also lowering annoyance. Here, we share key insights from recent research with the potential to inform applications related to human–computer interface design. Ultimately, this could lead to a cost-effective way to improve communication in medical contexts—with signification implications for both human health and the burgeoning medical device industry.
Collapse
|
29
|
A functional MRI investigation of crossmodal interference in an audiovisual Stroop task. PLoS One 2019; 14:e0210736. [PMID: 30645634 PMCID: PMC6333399 DOI: 10.1371/journal.pone.0210736] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2018] [Accepted: 01/01/2019] [Indexed: 01/08/2023] Open
Abstract
The visual color-word Stroop task is widely used in clinical and research settings as a measure of cognitive control. Numerous neuroimaging studies have used color-word Stroop tasks to investigate the neural resources supporting cognitive control, but to our knowledge all have used unimodal (typically visual) Stroop paradigms. Thus, it is possible that this classic measure of cognitive control is not capturing the resources involved in multisensory cognitive control. The audiovisual integration and crossmodal correspondence literatures identify regions sensitive to congruency of auditory and visual stimuli, but it is unclear how these regions relate to the unimodal cognitive control literature. In this study we aimed to identify brain regions engaged by crossmodal cognitive control during an audiovisual color-word Stroop task, and how they relate to previous unimodal Stroop and audiovisual integration findings. First, we replicated previous behavioral audiovisual Stroop findings in an fMRI-adapted audiovisual Stroop paradigm: incongruent visual information increased reaction time towards an auditory stimulus and congruent visual information decreased reaction time. Second, we investigated the brain regions supporting cognitive control during an audiovisual color-word Stroop task using fMRI. Similar to unimodal cognitive control tasks, a left superior parietal region exhibited an interference effect of visual information on the auditory stimulus. This superior parietal region was also identified using a standard audiovisual integration localizing procedure, indicating that audiovisual integration resources are sensitive to cognitive control demands. Facilitation of the auditory stimulus by congruent visual information was found in posterior superior temporal cortex, including in the posterior STS which has been found to support audiovisual integration. The dorsal anterior cingulate cortex, often implicated in unimodal Stroop tasks, was not modulated by the audiovisual Stroop task. Overall the findings indicate that an audiovisual color-word Stroop task engages overlapping resources with audiovisual integration and overlapping but distinct resources compared to unimodal Stroop tasks.
Collapse
|
30
|
Wang X, Gu J, Xu J, Li X, Geng J, Wang B, Liu B. Decoding natural scenes based on sounds of objects within scenes using multivariate pattern analysis. Neurosci Res 2018; 148:9-18. [PMID: 30513353 DOI: 10.1016/j.neures.2018.11.009] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2018] [Revised: 11/21/2018] [Accepted: 11/30/2018] [Indexed: 10/27/2022]
Abstract
Scene recognition plays an important role in spatial navigation and scene classification. It remains unknown whether the occipitotemporal cortex could represent the semantic association between the scenes and sounds of objects within the scenes. In this study, we used the functional magnetic resonance imaging (fMRI) technique and multivariate pattern analysis to assess whether diff ; ;erent scenes could be discriminated based on the patterns evoked by sounds of objects within the scenes. We found that patterns evoked by scenes could be predicted with patterns evoked by sounds of objects within the scenes in the posterior fusiform area (pF), lateral occipital area (LO) and superior temporal sulcus (STS). The further functional connectivity analysis suggested significant correlations between pF, LO and parahippocampal place area (PPA) except that between STS and other three regions under the scene and sound conditions. A distinct network in processing scenes and sounds was discovered using a seed-to-voxel analysis with STS as the seed. This study may provide a cross-modal channel of scene decoding through the sounds of objects within the scenes in the occipitotemporal cortex, which could complement the single-modal channel of scene decoding based on the global scene properties or objects within the scenes.
Collapse
Affiliation(s)
- Xiaojing Wang
- College of Intelligence and Computing, Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin, 300350, China
| | - Jin Gu
- College of Intelligence and Computing, Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin, 300350, China
| | - Junhai Xu
- College of Intelligence and Computing, Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin, 300350, China
| | - Xianglin Li
- Medical Imaging Research Institute, Binzhou Medical University, Yantai, Shandong 264003, China
| | - Junzu Geng
- Department of Radiology, Yantai Affiliated Hospital of Binzhou Medical University, Yantai, Shandong, 264003, China
| | - Bin Wang
- Medical Imaging Research Institute, Binzhou Medical University, Yantai, Shandong 264003, China
| | - Baolin Liu
- College of Intelligence and Computing, Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin University, Tianjin, 300350, China; State Key Laboratory of Intelligent Technology and Systems, National Laboratory for Information Science and Technology, Tsinghua University, Beijing, 100084, China.
| |
Collapse
|
31
|
Audiovisual Lexical Retrieval Deficits Following Left Hemisphere Stroke. Brain Sci 2018; 8:brainsci8120206. [PMID: 30486517 PMCID: PMC6316523 DOI: 10.3390/brainsci8120206] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2018] [Revised: 11/18/2018] [Accepted: 11/27/2018] [Indexed: 11/27/2022] Open
Abstract
Binding sensory features of multiple modalities of what we hear and see allows formation of a coherent percept to access semantics. Previous work on object naming has focused on visual confrontation naming with limited research in nonverbal auditory or multisensory processing. To investigate neural substrates and sensory effects of lexical retrieval, we evaluated healthy adults (n = 118) and left hemisphere stroke patients (LHD, n = 42) in naming manipulable objects across auditory (sound), visual (picture), and multisensory (audiovisual) conditions. LHD patients were divided into cortical, cortical–subcortical, or subcortical lesions (CO, CO–SC, SC), and specific lesion location investigated in a predictive model. Subjects produced lower accuracy in auditory naming relative to other conditions. Controls demonstrated greater naming accuracy and faster reaction times across all conditions compared to LHD patients. Naming across conditions was most severely impaired in CO patients. Both auditory and visual naming accuracy were impacted by temporal lobe involvement, although auditory naming was sensitive to lesions extending subcortically. Only controls demonstrated significant improvement over visual naming with the addition of auditory cues (i.e., multisensory condition). Results support overlapping neural networks for visual and auditory modalities related to semantic integration in lexical retrieval and temporal lobe involvement, while multisensory integration was impacted by both occipital and temporal lobe lesion involvement. The findings support modality specificity in naming and suggest that auditory naming is mediated by a distributed cortical–subcortical network overlapping with networks mediating spatiotemporal aspects of skilled movements producing sound.
Collapse
|
32
|
Abbott NT, Shahin AJ. Cross-modal phonetic encoding facilitates the McGurk illusion and phonemic restoration. J Neurophysiol 2018; 120:2988-3000. [PMID: 30303762 DOI: 10.1152/jn.00262.2018] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
In spoken language, audiovisual (AV) perception occurs when the visual modality influences encoding of acoustic features (e.g., phonetic representations) at the auditory cortex. We examined how visual speech (mouth movements) transforms phonetic representations, indexed by changes to the N1 auditory evoked potential (AEP). EEG was acquired while human subjects watched and listened to videos of a speaker uttering consonant vowel (CV) syllables, /ba/ and /wa/, presented in auditory-only or AV congruent or incongruent contexts or in a context in which the consonants were replaced by white noise (noise replaced). Subjects reported whether they heard "ba" or "wa." We hypothesized that the auditory N1 amplitude during illusory perception (caused by incongruent AV input, as in the McGurk illusion, or white noise-replaced consonants in CV utterances) should shift to reflect the auditory N1 characteristics of the phonemes conveyed visually (by mouth movements) as opposed to acoustically. Indeed, the N1 AEP became larger and occurred earlier when listeners experienced illusory "ba" (video /ba/, audio /wa/, heard as "ba") and vice versa when they experienced illusory "wa" (video /wa/, audio /ba/, heard as "wa"), mirroring the N1 AEP characteristics for /ba/ and /wa/ observed in natural acoustic situations (e.g., auditory-only setting). This visually mediated N1 behavior was also observed for noise-replaced CVs. Taken together, the findings suggest that information relayed by the visual modality modifies phonetic representations at the auditory cortex and that similar neural mechanisms support the McGurk illusion and visually mediated phonemic restoration. NEW & NOTEWORTHY Using a variant of the McGurk illusion experimental design (using the syllables /ba/ and /wa/), we demonstrate that lipreading influences phonetic encoding at the auditory cortex. We show that the N1 auditory evoked potential morphology shifts to resemble the N1 morphology of the syllable conveyed visually. We also show similar N1 shifts when the consonants are replaced by white noise, suggesting that the McGurk illusion and the visually mediated phonemic restoration rely on common mechanisms.
Collapse
Affiliation(s)
- Noelle T Abbott
- Center for Mind and Brain, University of California, Davis, California.,San Diego State University-University of California, San Diego Joint Doctoral Program in Language and Communicative Disorders, San Diego, California
| | - Antoine J Shahin
- Center for Mind and Brain, University of California, Davis, California.,Department of Cognitive and Information Sciences, University of California, Merced, California
| |
Collapse
|
33
|
Follmann R, Goldsmith CJ, Stein W. Multimodal sensory information is represented by a combinatorial code in a sensorimotor system. PLoS Biol 2018; 16:e2004527. [PMID: 30321170 PMCID: PMC6201955 DOI: 10.1371/journal.pbio.2004527] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2017] [Revised: 10/25/2018] [Accepted: 10/02/2018] [Indexed: 11/22/2022] Open
Abstract
A ubiquitous feature of the nervous system is the processing of simultaneously arriving sensory inputs from different modalities. Yet, because of the difficulties of monitoring large populations of neurons with the single resolution required to determine their sensory responses, the cellular mechanisms of how populations of neurons encode different sensory modalities often remain enigmatic. We studied multimodal information encoding in a small sensorimotor system of the crustacean stomatogastric nervous system that drives rhythmic motor activity for the processing of food. This system is experimentally advantageous, as it produces a fictive behavioral output in vitro, and distinct sensory modalities can be selectively activated. It has the additional advantage that all sensory information is routed through a hub ganglion, the commissural ganglion, a structure with fewer than 220 neurons. Using optical imaging of a population of commissural neurons to track each individual neuron's response across sensory modalities, we provide evidence that multimodal information is encoded via a combinatorial code of recruited neurons. By selectively stimulating chemosensory and mechanosensory inputs that are functionally important for processing of food, we find that these two modalities were processed in a distributed network comprising the majority of commissural neurons imaged. In a total of 12 commissural ganglia, we show that 98% of all imaged neurons were involved in sensory processing, with the two modalities being processed by a highly overlapping set of neurons. Of these, 80% were multimodal, 18% were unimodal, and only 2% of the neurons did not respond to either modality. Differences between modalities were represented by the identities of the neurons participating in each sensory condition and by differences in response sign (excitation versus inhibition), with 46% changing their responses in the other modality. Consistent with the hypothesis that the commissural network encodes different sensory conditions in the combination of activated neurons, a new combination of excitation and inhibition was found when both pathways were activated simultaneously. The responses to this bimodal condition were distinct from either unimodal condition, and for 30% of the neurons, they were not predictive from the individual unimodal responses. Thus, in a sensorimotor network, different sensory modalities are encoded using a combinatorial code of neurons that are activated or inhibited. This provides motor networks with the ability to differentially respond to categorically different sensory conditions and may serve as a model to understand higher-level processing of multimodal information.
Collapse
Affiliation(s)
- Rosangela Follmann
- School of Biological Sciences, Illinois State University, Normal, Illinois, United States of America
| | | | - Wolfgang Stein
- School of Biological Sciences, Illinois State University, Normal, Illinois, United States of America
| |
Collapse
|
34
|
Casado-Aranda LA, Van der Laan LN, Sánchez-Fernández J. Neural correlates of gender congruence in audiovisual commercials for gender-targeted products: An fMRI study. Hum Brain Mapp 2018; 39:4360-4372. [PMID: 29964348 DOI: 10.1002/hbm.24276] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2018] [Revised: 05/08/2018] [Accepted: 05/31/2018] [Indexed: 11/07/2022] Open
Abstract
This article explores neural and self-report responses to gender congruence in product-voice combinations in commercials. An fMRI study was carried out in which participants (n = 30) were presented with gender-targeted pictures of characteristic male or female products accompanied by either gender congruent or incongruent voices. The findings show that attitudes are more positive toward commercials with gender congruent than with gender incongruent product-voice combinations. fMRI analyses revealed that primary visual brain areas, namely calcarine and cuneus, responded stronger to congruent than incongruent combinations suggesting that participants enhanced their endogenous attention toward congruent commercials. Incongruent combinations, by contrast, elicited stronger activation in areas related to the perception of conflicts in information processing and error monitoring, such as the supramarginal, inferior parietal gyri and superior, and middle temporal gyri. Interestingly, increased activation in the posterior cingulate cortex (an area related to value encoding) predicted more positive attitudes toward congruent commercials. Together, these results advance our understanding of the neural correlates of processing congruent and incongruent audiovisual stimuli. These findings may advice advertising professionals in designing successful campaigns of everyday products, namely by making use of congruent instead of incongruent product-voice combinations.
Collapse
Affiliation(s)
- Luis-Alberto Casado-Aranda
- Department of Marketing and Market Research, University of Granada, Campus Universitario la Cartuja, Granada, Spain
| | - Laura Nynke Van der Laan
- University of Amsterdam, Amsterdam School of Communication Research (ASCoR), NG Amsterdam, The Netherlands
| | - Juan Sánchez-Fernández
- Department of Marketing and Market Research, University of Granada, Campus Universitario la Cartuja, Granada, Spain
| |
Collapse
|
35
|
Neural networks supporting audiovisual integration for speech: A large-scale lesion study. Cortex 2018; 103:360-371. [PMID: 29705718 DOI: 10.1016/j.cortex.2018.03.030] [Citation(s) in RCA: 25] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2017] [Revised: 12/05/2017] [Accepted: 03/23/2018] [Indexed: 10/17/2022]
Abstract
Auditory and visual speech information are often strongly integrated resulting in perceptual enhancements for audiovisual (AV) speech over audio alone and sometimes yielding compelling illusory fusion percepts when AV cues are mismatched, the McGurk-MacDonald effect. Previous research has identified three candidate regions thought to be critical for AV speech integration: the posterior superior temporal sulcus (STS), early auditory cortex, and the posterior inferior frontal gyrus. We assess the causal involvement of these regions (and others) in the first large-scale (N = 100) lesion-based study of AV speech integration. Two primary findings emerged. First, behavioral performance and lesion maps for AV enhancement and illusory fusion measures indicate that classic metrics of AV speech integration are not necessarily measuring the same process. Second, lesions involving superior temporal auditory, lateral occipital visual, and multisensory zones in the STS are the most disruptive to AV speech integration. Further, when AV speech integration fails, the nature of the failure-auditory vs visual capture-can be predicted from the location of the lesions. These findings show that AV speech processing is supported by unimodal auditory and visual cortices as well as multimodal regions such as the STS at their boundary. Motor related frontal regions do not appear to play a role in AV speech integration.
Collapse
|
36
|
Neural Mechanisms Underlying Cross-Modal Phonetic Encoding. J Neurosci 2017; 38:1835-1849. [PMID: 29263241 DOI: 10.1523/jneurosci.1566-17.2017] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2017] [Revised: 11/17/2017] [Accepted: 12/08/2017] [Indexed: 11/21/2022] Open
Abstract
Audiovisual (AV) integration is essential for speech comprehension, especially in adverse listening situations. Divergent, but not mutually exclusive, theories have been proposed to explain the neural mechanisms underlying AV integration. One theory advocates that this process occurs via interactions between the auditory and visual cortices, as opposed to fusion of AV percepts in a multisensory integrator. Building upon this idea, we proposed that AV integration in spoken language reflects visually induced weighting of phonetic representations at the auditory cortex. EEG was recorded while male and female human subjects watched and listened to videos of a speaker uttering consonant vowel (CV) syllables /ba/ and /fa/, presented in Auditory-only, AV congruent or incongruent contexts. Subjects reported whether they heard /ba/ or /fa/. We hypothesized that vision alters phonetic encoding by dynamically weighting which phonetic representation in the auditory cortex is strengthened or weakened. That is, when subjects are presented with visual /fa/ and acoustic /ba/ and hear /fa/ (illusion-fa), the visual input strengthens the weighting of the phone /f/ representation. When subjects are presented with visual /ba/ and acoustic /fa/ and hear /ba/ (illusion-ba), the visual input weakens the weighting of the phone /f/ representation. Indeed, we found an enlarged N1 auditory evoked potential when subjects perceived illusion-ba, and a reduced N1 when they perceived illusion-fa, mirroring the N1 behavior for /ba/ and /fa/ in Auditory-only settings. These effects were especially pronounced in individuals with more robust illusory perception. These findings provide evidence that visual speech modifies phonetic encoding at the auditory cortex.SIGNIFICANCE STATEMENT The current study presents evidence that audiovisual integration in spoken language occurs when one modality (vision) acts on representations of a second modality (audition). Using the McGurk illusion, we show that visual context primes phonetic representations at the auditory cortex, altering the auditory percept, evidenced by changes in the N1 auditory evoked potential. This finding reinforces the theory that audiovisual integration occurs via visual networks influencing phonetic representations in the auditory cortex. We believe that this will lead to the generation of new hypotheses regarding cross-modal mapping, particularly whether it occurs via direct or indirect routes (e.g., via a multisensory mediator).
Collapse
|
37
|
Starke J, Ball F, Heinze HJ, Noesselt T. The spatio-temporal profile of multisensory integration. Eur J Neurosci 2017; 51:1210-1223. [PMID: 29057531 DOI: 10.1111/ejn.13753] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2017] [Revised: 10/13/2017] [Accepted: 10/16/2017] [Indexed: 12/29/2022]
Abstract
Task-irrelevant visual stimuli can enhance auditory perception. However, while there is some neurophysiological evidence for mechanisms that underlie the phenomenon, the neural basis of visually induced effects on auditory perception remains unknown. Combining fMRI and EEG with psychophysical measurements in two independent studies, we identified the neural underpinnings and temporal dynamics of visually induced auditory enhancement. Lower- and higher-intensity sounds were paired with a non-informative visual stimulus, while participants performed an auditory detection task. Behaviourally, visual co-stimulation enhanced auditory sensitivity. Using fMRI, enhanced BOLD signals were observed in primary auditory cortex for low-intensity audiovisual stimuli which scaled with subject-specific enhancement in perceptual sensitivity. Concordantly, a modulation of event-related potentials could already be observed over frontal electrodes at an early latency (30-80 ms), which again scaled with subject-specific behavioural benefits. Later modulations starting around 280 ms, that is in the time range of the P3, did not fit this pattern of brain-behaviour correspondence. Hence, the latency of the corresponding fMRI-EEG brain-behaviour modulation points at an early interplay of visual and auditory signals in low-level auditory cortex, potentially mediated by crosstalk at the level of the thalamus. However, fMRI signals in primary auditory cortex, auditory thalamus and the P50 for higher-intensity auditory stimuli were also elevated by visual co-stimulation (in the absence of any behavioural effect) suggesting a general, intensity-independent integration mechanism. We propose that this automatic interaction occurs at the level of the thalamus and might signify a first step of audiovisual interplay necessary for visually induced perceptual enhancement of auditory perception.
Collapse
Affiliation(s)
- Johanna Starke
- Department of Biological Psychology, Faculty of Natural Science, Otto-von-Guericke-University Magdeburg, Magdeburg, Germany.,Department of Neurology, Faculty of Medicine, Otto-von-Guericke-University Magdeburg, Magdeburg, Germany
| | - Felix Ball
- Department of Biological Psychology, Faculty of Natural Science, Otto-von-Guericke-University Magdeburg, Magdeburg, Germany.,Department of Neurology, Faculty of Medicine, Otto-von-Guericke-University Magdeburg, Magdeburg, Germany.,Center for Behavioural Brain Sciences, Otto-von-Guericke-University Magdeburg, Magdeburg, Germany
| | - Hans-Jochen Heinze
- Department of Neurology, Faculty of Medicine, Otto-von-Guericke-University Magdeburg, Magdeburg, Germany.,Center for Behavioural Brain Sciences, Otto-von-Guericke-University Magdeburg, Magdeburg, Germany
| | - Toemme Noesselt
- Department of Biological Psychology, Faculty of Natural Science, Otto-von-Guericke-University Magdeburg, Magdeburg, Germany.,Center for Behavioural Brain Sciences, Otto-von-Guericke-University Magdeburg, Magdeburg, Germany
| |
Collapse
|