1
|
Rizzi R, Bidelman GM. Functional benefits of continuous vs. categorical listening strategies on the neural encoding and perception of noise-degraded speech. Brain Res 2024; 1844:149166. [PMID: 39151718 PMCID: PMC11399885 DOI: 10.1016/j.brainres.2024.149166] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/19/2024] [Revised: 07/26/2024] [Accepted: 08/13/2024] [Indexed: 08/19/2024]
Abstract
Acoustic information in speech changes continuously, yet listeners form discrete perceptual categories to ease the demands of perception. Being a more continuous/gradient as opposed to a more discrete/categorical listener may be further advantageous for understanding speech in noise by increasing perceptual flexibility and resolving ambiguity. The degree to which a listener's responses to a continuum of speech sounds are categorical versus continuous can be quantified using visual analog scaling (VAS) during speech labeling tasks. Here, we recorded event-related brain potentials (ERPs) to vowels along an acoustic-phonetic continuum (/u/ to /a/) while listeners categorized phonemes in both clean and noise conditions. Behavior was assessed using standard two alternative forced choice (2AFC) and VAS paradigms to evaluate categorization under task structures that promote discrete vs. continuous hearing, respectively. Behaviorally, identification curves were steeper under 2AFC vs. VAS categorization but were relatively immune to noise, suggesting robust access to abstract, phonetic categories even under signal degradation. Behavioral slopes were correlated with listeners' QuickSIN scores; shallower slopes corresponded with better speech in noise performance, suggesting a perceptual advantage to noise degraded speech comprehension conferred by a more gradient listening strategy. At the neural level, P2 amplitudes and latencies of the ERPs were modulated by task and noise; VAS responses were larger and showed greater noise-related latency delays than 2AFC responses. More gradient responders had smaller shifts in ERP latency with noise, suggesting their neural encoding of speech was more resilient to noise degradation. Interestingly, source-resolved ERPs showed that more gradient listening was also correlated with stronger neural responses in left superior temporal gyrus. Our results demonstrate that listening strategy modulates the categorical organization of speech and behavioral success, with more continuous/gradient listening being advantageous to sentential speech in noise perception.
Collapse
Affiliation(s)
- Rose Rizzi
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA; Program in Neuroscience, Indiana University, Bloomington, IN, USA
| | - Gavin M Bidelman
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA; Program in Neuroscience, Indiana University, Bloomington, IN, USA; Cognitive Science Program, Indiana University, Bloomington, IN, USA.
| |
Collapse
|
2
|
He D, Buder EH, Bidelman GM. Cross-linguistic and acoustic-driven effects on multiscale neural synchrony to stress rhythms. BRAIN AND LANGUAGE 2024; 256:105463. [PMID: 39243486 DOI: 10.1016/j.bandl.2024.105463] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/05/2023] [Revised: 09/01/2024] [Accepted: 09/03/2024] [Indexed: 09/09/2024]
Abstract
We investigated how neural oscillations code the hierarchical nature of stress rhythms in speech and how stress processing varies with language experience. By measuring phase synchrony of multilevel EEG-acoustic tracking and intra-brain cross-frequency coupling, we show the encoding of stress involves different neural signatures (delta rhythms = stress foot rate; theta rhythms = syllable rate), is stronger for amplitude vs. duration stress cues, and induces nested delta-theta coherence mirroring the stress-syllable hierarchy in speech. Only native English, but not Mandarin, speakers exhibited enhanced neural entrainment at central stress (2 Hz) and syllable (4 Hz) rates intrinsic to natural English. English individuals with superior cortical-stress tracking capabilities also displayed stronger neural hierarchical coherence, highlighting a nuanced interplay between internal nesting of brain rhythms and external entrainment rooted in language-specific speech rhythms. Our cross-language findings reveal brain-speech synchronization is not purely a "bottom-up" but benefits from "top-down" processing from listeners' language-specific experience.
Collapse
Affiliation(s)
- Deling He
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA
| | - Eugene H Buder
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA
| | - Gavin M Bidelman
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA; Program in Neuroscience, Indiana University, Bloomington, IN, USA; Cognitive Science Program, Indiana University, Bloomington, IN, USA.
| |
Collapse
|
3
|
Jo HS, Hsieh TH, Chien WC, Shaw FZ, Liang SF, Kung CC. Probing the neural dynamics of musicians' and non-musicians' consonant/dissonant perception: Joint analyses of electrical encephalogram (EEG) and functional magnetic resonance imaging (fMRI). Neuroimage 2024; 298:120784. [PMID: 39147290 DOI: 10.1016/j.neuroimage.2024.120784] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/25/2023] [Revised: 08/06/2024] [Accepted: 08/12/2024] [Indexed: 08/17/2024] Open
Abstract
The perception of two (or more) simultaneous musical notes, depending on their pitch interval(s), could be broadly categorized as consonant or dissonant. Previous literature has suggested that musicians and non-musicians adopt different strategies when discerning music intervals: while musicians rely on the frequency ratios between the two fundamental frequencies, such as "perfect fifth" (3:2) as consonant and "tritone" (45:32) as dissonant intervals; non-musicians may rely on the presence of 'roughness' or 'beats', generated by the difference of fundamental frequencies, as the key elements of 'dissonance'. The separate Event-Related Potential (ERP) differences in N1 and P2 along the midline electrodes provided evidence congruent with such 'separate reliances'. To replicate and to extend, in this study we reran the previous experiment, and separately collected fMRI data of the same protocol (with sparse sampling modifications). The behavioral and EEG results largely corresponded to our previous finding. The fMRI results, with the joint analyses by univariate, psycho-physiological interaction, and representational similarity analysis (RSA) approaches, further reinforce the involvement of central midline-related brain regions, such as ventromedial prefrontal and dorsal anterior cingulate cortex, in consonant/dissonance judgments. The final spatiotemporal searchlight RSA provided convincing evidence that the medial prefrontal cortex, along with the bilateral superior temporal cortex, is the joint locus of midline N1 and dorsal anterior cingulate cortex for the P2 effect (for musicians). Together, these analyses reaffirm that musicians rely more on experience-driven knowledge for consonance/dissonance perception; but also demonstrate the advantages of multiple analyses in constraining the findings from both EEG and fMRI.
Collapse
Affiliation(s)
- Han Shin Jo
- Institute of Medical Informatics, National Cheng Kung University (NCKU), Tainan, 70101, Taiwan
| | - Tsung-Hao Hsieh
- Department of Computer Science and Information Engineering, NCKU, Tainan, 70101, Taiwan; Department of Computer Science, Tunghai University, Taichung, 407224, Taiwan
| | - Wei-Che Chien
- Department of Computer Science and Information Engineering, NCKU, Tainan, 70101, Taiwan
| | - Fu-Zen Shaw
- Department of Psychology, NCKU, Tainan, 70101, Taiwan; Mind Research and Imaging Center, NCKU, Tainan, 70101, Taiwan
| | - Sheng-Fu Liang
- Institute of Medical Informatics, National Cheng Kung University (NCKU), Tainan, 70101, Taiwan; Department of Computer Science and Information Engineering, NCKU, Tainan, 70101, Taiwan
| | - Chun-Chia Kung
- Department of Psychology, NCKU, Tainan, 70101, Taiwan; Mind Research and Imaging Center, NCKU, Tainan, 70101, Taiwan.
| |
Collapse
|
4
|
Bidelman GM, York A, Pearson C. Neural correlates of phonetic categorization under auditory (phoneme) and visual (grapheme) modalities. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.07.24.604940. [PMID: 39211275 PMCID: PMC11361091 DOI: 10.1101/2024.07.24.604940] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/04/2024]
Abstract
We tested whether the neural mechanisms of phonetic categorization are specific to speech sounds or generalize to graphemes (i.e., visual letters) of the same phonetic label. Given that linguistic experience shapes categorical processing, and letter-speech sound matching plays a crucial role during early reading acquisition, we hypothesized sound phoneme and visual grapheme tokens representing the same linguistic identity might recruit common neural substrates, despite originating from different sensory modalities. Behavioral and neuroelectric brain responses (ERPs) were acquired as participants categorized stimuli from sound (phoneme) and homologous letter (grapheme) continua each spanning a /da/ - /ga/ gradient. Behaviorally, listeners were faster and showed stronger categorization of phoneme compared to graphemes. At the neural level, multidimensional scaling of the EEG revealed responses self-organized in a categorial fashion such that tokens clustered within their respective modality beginning ∼150-250 ms after stimulus onset. Source-resolved ERPs further revealed modality-specific and overlapping brain regions supporting phonetic categorization. Left inferior frontal gyrus and auditory cortex showed stronger responses for sound category members compared to phonetically ambiguous tokens, whereas early visual cortices paralleled this categorical organization for graphemes. Auditory and visual categorization also recruited common visual association areas in extrastriate cortex but in opposite hemispheres (auditory = left; visual=right). Our findings reveal both auditory and visual sensory cortex supports categorical organization for phonetic labels within their respective modalities. However, a partial overlap in phoneme and grapheme processing among occipital brain areas implies the presence of an isomorphic, domain-general mapping for phonetic categories in dorsal visual system.
Collapse
|
5
|
Rizzi R, Bidelman GM. Functional benefits of continuous vs. categorical listening strategies on the neural encoding and perception of noise-degraded speech. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.05.15.594387. [PMID: 38798410 PMCID: PMC11118460 DOI: 10.1101/2024.05.15.594387] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/29/2024]
Abstract
Acoustic information in speech changes continuously, yet listeners form discrete perceptual categories to ease the demands of perception. Being a more continuous/gradient as opposed to a discrete/categorical listener may be further advantageous for understanding speech in noise by increasing perceptual flexibility and resolving ambiguity. The degree to which a listener's responses to a continuum of speech sounds are categorical versus continuous can be quantified using visual analog scaling (VAS) during speech labeling tasks. Here, we recorded event-related brain potentials (ERPs) to vowels along an acoustic-phonetic continuum (/u/ to /a/) while listeners categorized phonemes in both clean and noise conditions. Behavior was assessed using standard two alternative forced choice (2AFC) and VAS paradigms to evaluate categorization under task structures that promote discrete (2AFC) vs. continuous (VAS) hearing, respectively. Behaviorally, identification curves were steeper under 2AFC vs. VAS categorization but were relatively immune to noise, suggesting robust access to abstract, phonetic categories even under signal degradation. Behavioral slopes were positively correlated with listeners' QuickSIN scores, suggesting a behavioral advantage for speech in noise comprehension conferred by gradient listening strategy. At the neural level, electrode level data revealed P2 peak amplitudes of the ERPs were modulated by task and noise; responses were larger under VAS vs. 2AFC categorization and showed larger noise-related delay in latency in the VAS vs. 2AFC condition. More gradient responders also had smaller shifts in ERP latency with noise, suggesting their neural encoding of speech was more resilient to noise degradation. Interestingly, source-resolved ERPs showed that more gradient listening was also correlated with stronger neural responses in left superior temporal gyrus. Our results demonstrate that listening strategy (i.e., being a discrete vs. continuous listener) modulates the categorical organization of speech and behavioral success, with continuous/gradient listening being more advantageous to speech in noise perception.
Collapse
|
6
|
Momtaz S, Bidelman GM. Effects of Stimulus Rate and Periodicity on Auditory Cortical Entrainment to Continuous Sounds. eNeuro 2024; 11:ENEURO.0027-23.2024. [PMID: 38253583 PMCID: PMC10913036 DOI: 10.1523/eneuro.0027-23.2024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2023] [Revised: 01/14/2024] [Accepted: 01/16/2024] [Indexed: 01/24/2024] Open
Abstract
The neural mechanisms underlying the exogenous coding and neural entrainment to repetitive auditory stimuli have seen a recent surge of interest. However, few studies have characterized how parametric changes in stimulus presentation alter entrained responses. We examined the degree to which the brain entrains to repeated speech (i.e., /ba/) and nonspeech (i.e., click) sounds using phase-locking value (PLV) analysis applied to multichannel human electroencephalogram (EEG) data. Passive cortico-acoustic tracking was investigated in N = 24 normal young adults utilizing EEG source analyses that isolated neural activity stemming from both auditory temporal cortices. We parametrically manipulated the rate and periodicity of repetitive, continuous speech and click stimuli to investigate how speed and jitter in ongoing sound streams affect oscillatory entrainment. Neuronal synchronization to speech was enhanced at 4.5 Hz (the putative universal rate of speech) and showed a differential pattern to that of clicks, particularly at higher rates. PLV to speech decreased with increasing jitter but remained superior to clicks. Surprisingly, PLV entrainment to clicks was invariant to periodicity manipulations. Our findings provide evidence that the brain's neural entrainment to complex sounds is enhanced and more sensitized when processing speech-like stimuli, even at the syllable level, relative to nonspeech sounds. The fact that this specialization is apparent even under passive listening suggests a priority of the auditory system for synchronizing to behaviorally relevant signals.
Collapse
Affiliation(s)
- Sara Momtaz
- School of Communication Sciences & Disorders, University of Memphis, Memphis, Tennessee 38152
- Boys Town National Research Hospital, Boys Town, Nebraska 68131
| | - Gavin M Bidelman
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, Indiana 47408
- Program in Neuroscience, Indiana University, Bloomington, Indiana 47405
| |
Collapse
|
7
|
MacLean J, Stirn J, Sisson A, Bidelman GM. Short- and long-term neuroplasticity interact during the perceptual learning of concurrent speech. Cereb Cortex 2024; 34:bhad543. [PMID: 38212291 PMCID: PMC10839853 DOI: 10.1093/cercor/bhad543] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2023] [Revised: 12/20/2023] [Accepted: 12/21/2023] [Indexed: 01/13/2024] Open
Abstract
Plasticity from auditory experience shapes the brain's encoding and perception of sound. However, whether such long-term plasticity alters the trajectory of short-term plasticity during speech processing has yet to be investigated. Here, we explored the neural mechanisms and interplay between short- and long-term neuroplasticity for rapid auditory perceptual learning of concurrent speech sounds in young, normal-hearing musicians and nonmusicians. Participants learned to identify double-vowel mixtures during ~ 45 min training sessions recorded simultaneously with high-density electroencephalography (EEG). We analyzed frequency-following responses (FFRs) and event-related potentials (ERPs) to investigate neural correlates of learning at subcortical and cortical levels, respectively. Although both groups showed rapid perceptual learning, musicians showed faster behavioral decisions than nonmusicians overall. Learning-related changes were not apparent in brainstem FFRs. However, plasticity was highly evident in cortex, where ERPs revealed unique hemispheric asymmetries between groups suggestive of different neural strategies (musicians: right hemisphere bias; nonmusicians: left hemisphere). Source reconstruction and the early (150-200 ms) time course of these effects localized learning-induced cortical plasticity to auditory-sensory brain areas. Our findings reinforce the domain-general benefits of musicianship but reveal that successful speech sound learning is driven by a critical interplay between long- and short-term mechanisms of auditory plasticity, which first emerge at a cortical level.
Collapse
Affiliation(s)
- Jessica MacLean
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA
- Program in Neuroscience, Indiana University, Bloomington, IN, USA
| | - Jack Stirn
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA
| | - Alexandria Sisson
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA
| | - Gavin M Bidelman
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA
- Program in Neuroscience, Indiana University, Bloomington, IN, USA
- Cognitive Science Program, Indiana University, Bloomington, IN, USA
| |
Collapse
|
8
|
Stahl AN, Racca JM, Kerley CI, Anderson A, Landman B, Hood LJ, Gifford RH, Rex TS. Comprehensive behavioral and physiologic assessment of peripheral and central auditory function in individuals with mild traumatic brain injury. Hear Res 2024; 441:108928. [PMID: 38086151 DOI: 10.1016/j.heares.2023.108928] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/28/2023] [Revised: 12/01/2023] [Accepted: 12/04/2023] [Indexed: 12/27/2023]
Abstract
Auditory complaints are frequently reported by individuals with mild traumatic brain injury (mTBI) yet remain difficult to detect in the absence of clinically significant hearing loss. This highlights a growing need to identify sensitive indices of auditory-related mTBI pathophysiology beyond pure-tone thresholds for improved hearing healthcare diagnosis and treatment. Given the heterogeneity of mTBI etiology and the diverse peripheral and central processes required for normal auditory function, the present study sought to determine the audiologic assessments sensitive to mTBI pathophysiology at the group level using a well-rounded test battery of both peripheral and central auditory system function. This test battery included pure-tone detection thresholds, word understanding in quiet, sentence understanding in noise, distortion product otoacoustic emissions (DPOAEs), middle-ear muscle reflexes (MEMRs), and auditory evoked potentials (AEPs), including auditory brainstem responses (ABRs), middle latency responses (MLRs), and late latency responses (LLRs). Each participant also received magnetic resonance imaging (MRI). Compared to the control group, we found that individuals with mTBI had reduced DPOAE amplitudes that revealed a compound effect of age, elevated MEMR thresholds for an ipsilateral broadband noise elicitor, longer ABR Wave I latencies for click and 4 kHz tone burst elicitors, longer ABR Wave III latencies for 4 kHz tone bursts, larger MLR Na and Nb amplitudes, smaller MLR Pb amplitudes, longer MLR Pa latencies, and smaller LLR N1 amplitudes for older individuals with mTBI. Further, mTBI individuals with combined hearing difficulty and noise sensitivity had a greater number of deficits on thalamic and cortical AEP measures compared to those with only one/no self-reported auditory symptoms. This finding was corroborated with MRI, which revealed significant structural differences in the auditory cortical areas of mTBI participants who reported combined hearing difficulty and noise sensitivity, including an enlargement of left transverse temporal gyrus (TTG) and bilateral planum polare (PP). These findings highlight the need for continued investigations toward identifying individualized audiologic assessments and treatments that are sensitive to mTBI pathophysiology.
Collapse
Affiliation(s)
- Amy N Stahl
- Neuroscience Graduate Program, Vanderbilt University, Nashville, TN USA; Department of Ophthalmology & Visual Sciences, Vanderbilt Eye Institute, Vanderbilt University Medical Center, Nashville, TN USA
| | - Jordan M Racca
- Department of Hearing & Speech Sciences, Vanderbilt University Medical Center, Nashville, TN USA; Collaborative for STEM Education and Outreach, Vanderbilt Peabody College of Education, Vanderbilt University, Nashville, TN USA
| | - Cailey I Kerley
- Department of Electrical and Computer Engineering, Vanderbilt University, Nashville, TN, USA
| | - Adam Anderson
- Department of Biomedical Engineering, Vanderbilt University, Nashville, TN, USA
| | - Bennett Landman
- Department of Electrical and Computer Engineering, Vanderbilt University, Nashville, TN, USA
| | - Linda J Hood
- Department of Hearing & Speech Sciences, Vanderbilt University Medical Center, Nashville, TN USA
| | - René H Gifford
- Department of Hearing & Speech Sciences, Vanderbilt University Medical Center, Nashville, TN USA
| | - Tonia S Rex
- Department of Biomedical Engineering, Vanderbilt University, Nashville, TN, USA.
| |
Collapse
|
9
|
He D, Buder EH, Bidelman GM. Cross-linguistic and acoustic-driven effects on multiscale neural synchrony to stress rhythms. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.12.04.570012. [PMID: 38106017 PMCID: PMC10723321 DOI: 10.1101/2023.12.04.570012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/19/2023]
Abstract
We investigated how neural oscillations code the hierarchical nature of stress rhythms in speech and how stress processing varies with language experience. By measuring phase synchrony of multilevel EEG-acoustic tracking and intra-brain cross-frequency coupling, we show the encoding of stress involves different neural signatures (delta rhythms = stress foot rate; theta rhythms = syllable rate), is stronger for amplitude vs. duration stress cues, and induces nested delta-theta coherence mirroring the stress-syllable hierarchy in speech. Only native English, but not Mandarin, speakers exhibited enhanced neural entrainment at central stress (2 Hz) and syllable (4 Hz) rates intrinsic to natural English. English individuals with superior cortical-stress tracking capabilities also displayed stronger neural hierarchical coherence, highlighting a nuanced interplay between internal nesting of brain rhythms and external entrainment rooted in language-specific speech rhythms. Our cross-language findings reveal brain-speech synchronization is not purely a "bottom-up" but benefits from "top-down" processing from listeners' language-specific experience.
Collapse
Affiliation(s)
- Deling He
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA
| | - Eugene H. Buder
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA
| | - Gavin M. Bidelman
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA
- Program in Neuroscience, Indiana University, Bloomington, IN, USA
| |
Collapse
|
10
|
Shen D, Ross B, Alain C. Temporal deployment of attention in musicians: Evidence from an attentional blink paradigm. Ann N Y Acad Sci 2023; 1530:110-123. [PMID: 37823710 DOI: 10.1111/nyas.15069] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/13/2023]
Abstract
The generalization of music training to unrelated nonmusical domains is well established and may reflect musicians' superior ability to regulate attention. We investigated the temporal deployment of attention in musicians and nonmusicians using scalp-recording of event-related potentials in an attentional blink (AB) paradigm. Participants listened to rapid sequences of stimuli and identified target and probe sounds. The AB was defined as a probe identification deficit when the probe closely follows the target. The sequence of stimuli was preceded by a neutral or informative cue about the probe position within the sequence. Musicians outperformed nonmusicians in identifying the target and probe. In both groups, cueing improved target and probe identification and reduced the AB. The informative cue elicited a sustained potential, which was more prominent in musicians than nonmusicians over left temporal areas and yielded a larger N1 amplitude elicited by the target. The N1 was larger in musicians than nonmusicians, and its amplitude over the left frontocentral cortex of musicians correlated with accuracy. Together, these results reveal musicians' superior ability to regulate attention, allowing them to prepare for incoming stimuli, thereby improving sound object identification. This capacity to manage attentional resources to optimize task performance may generalize to nonmusical activities.
Collapse
Affiliation(s)
- Dawei Shen
- Rotman Research Institute, Baycrest Centre for Geriatric Care, Toronto, Ontario, Canada
| | - Bernhard Ross
- Rotman Research Institute, Baycrest Centre for Geriatric Care, Toronto, Ontario, Canada
- Department of Medical Biophysics, University of Toronto, Toronto, Ontario, Canada
- Institute of Medical Sciences, University of Toronto, Toronto, Ontario, Canada
- Music and Health Science Research Collaboratory, University of Toronto, Toronto, Ontario, Canada
| | - Claude Alain
- Rotman Research Institute, Baycrest Centre for Geriatric Care, Toronto, Ontario, Canada
- Institute of Medical Sciences, University of Toronto, Toronto, Ontario, Canada
- Music and Health Science Research Collaboratory, University of Toronto, Toronto, Ontario, Canada
- Department of Psychology, University of Toronto, Toronto, Ontario, Canada
| |
Collapse
|
11
|
Arao H, Suwazono S, Kimura A, Asano H, Suzuki H. Measuring auditory event-related potentials at the external ear canal: A demonstrative study using a new electrode and error-feedback paradigm. Eur J Neurosci 2023; 58:4310-4327. [PMID: 37875165 DOI: 10.1111/ejn.16175] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2021] [Revised: 09/21/2023] [Accepted: 10/04/2023] [Indexed: 10/26/2023]
Abstract
Although ear canal electroencephalogram (EEG) recording has received interest from basic and applied research communities, evidence on how it can be implemented in practice is limited. The present study involving eight male participants including the authors presents the utility of our ear canal electrode and method by demonstrating both comparability of ear canal EEG to those at nearby sites and distinctiveness that ear canal event-related potentials (ERPs) could have. For this purpose, we used the balanced noncephalic electrode reference and an experimental paradigm with an error-feedback sound. Clear auditory ERPs were detected at the ear canal sites with a sufficiently low noise level comparable with those at conventional sites. The N1c, a temporal maximum subcomponent, spread over the bilateral temporal sites, including the ear canals and earlobes. While consecutive signals are generally highly similar between the ear canal and the earlobe, the N1c was larger at the ear canal than the earlobe, as demonstrated by the conventional frequentist and the hierarchical Bayesian modelling approaches. Although an evident caveat is that our sample was limited in terms of size and sex, the general capability indicates that the structure of our ear canal electrode provides EEG measurement that can be used in basic and applied settings. Our experimental method can also be an ERP-based test that conveniently assesses the capability of existing and future ear canal electrodes. The distinctive nature of the ERPs to the error-feedback sound may be utilized to examine the basic aspects of auditory ERPs and to test the processes involved in feedback-guided behaviour of participants.
Collapse
Affiliation(s)
- Hiroshi Arao
- Department of Human Sciences, Taisho University, Tokyo, Japan
| | - Shugo Suwazono
- Department of Neurology and Center for Clinical Neuroscience, National Hospital Organization Okinawa National Hospital, Ginowan, Japan
| | | | - Hirotoshi Asano
- Department of Computer Science, Kogakuin University, Tokyo, Japan
| | - Hiroaki Suzuki
- Department of Human Sciences, Taisho University, Tokyo, Japan
| |
Collapse
|
12
|
Brown JA, Bidelman GM. Attention, Musicality, and Familiarity Shape Cortical Speech Tracking at the Musical Cocktail Party. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.10.28.562773. [PMID: 37961204 PMCID: PMC10634879 DOI: 10.1101/2023.10.28.562773] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/15/2023]
Abstract
The "cocktail party problem" challenges our ability to understand speech in noisy environments, which often include background music. Here, we explored the role of background music in speech-in-noise listening. Participants listened to an audiobook in familiar and unfamiliar music while tracking keywords in either speech or song lyrics. We used EEG to measure neural tracking of the audiobook. When speech was masked by music, the modeled peak latency at 50 ms (P1TRF) was prolonged compared to unmasked. Additionally, P1TRF amplitude was larger in unfamiliar background music, suggesting improved speech tracking. We observed prolonged latencies at 100 ms (N1TRF) when speech was not the attended stimulus, though only in less musical listeners. Our results suggest early neural representations of speech are enhanced with both attention and concurrent unfamiliar music, indicating familiar music is more distracting. One's ability to perceptually filter "musical noise" at the cocktail party depends on objective musical abilities.
Collapse
Affiliation(s)
- Jane A. Brown
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA
- Institute for Intelligent Systems, University of Memphis, Memphis, TN 38152, USA
| | - Gavin M. Bidelman
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA
- Program in Neuroscience, Indiana University, Bloomington, IN, USA
- Cognitive Science Program, Indiana University, Bloomington, IN, USA
| |
Collapse
|
13
|
Momtaz S, Moncrieff D, Ray MA, Bidelman GM. Children with amblyaudia show less flexibility in auditory cortical entrainment to periodic non-speech sounds. Int J Audiol 2023; 62:920-926. [PMID: 35822427 PMCID: PMC10026530 DOI: 10.1080/14992027.2022.2094289] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/19/2021] [Revised: 06/16/2022] [Accepted: 06/21/2022] [Indexed: 11/05/2022]
Abstract
OBJECTIVE We investigated auditory temporal processing in children with amblyaudia (AMB), a subtype of auditory processing disorder (APD), via cortical neural entrainment. DESIGN AND STUDY SAMPLES Evoked responses were recorded to click-trains at slow vs. fast (8.5 vs. 14.9/s) rates in n = 14 children with AMB and n = 11 age-matched controls. Source and time-frequency analyses (TFA) decomposed EEGs into oscillations (reflecting neural entrainment) stemming from bilateral auditory cortex. RESULTS Phase-locking strength in AMB depended critically on the speed of auditory stimuli. In contrast to age-matched peers, AMB responses were largely insensitive to rate manipulations. This rate resistance occurred regardless of the ear of presentation and in both cortical hemispheres. CONCLUSIONS Children with AMB show less rate-related changes in auditory cortical entrainment. In addition to reduced capacity to integrate information between the ears, we identify more rigid tagging of external auditory stimuli. Our neurophysiological findings may account for domain-general temporal processing deficits commonly observed in AMB and related APDs behaviourally. More broadly, our findings may inform communication strategies and future rehabilitation programmes; increasing the rate of stimuli above a normal (slow) speech rate is likely to make stimulus processing more challenging for individuals with AMB/APD.
Collapse
Affiliation(s)
- Sara Momtaz
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, USA
| | - Deborah Moncrieff
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, USA
| | - Meredith A. Ray
- Division of Epidemiology, Biostatistics, and Environmental Health, School of Public Health, University of Memphis, Memphis, TN, USA
| | - Gavin M. Bidelman
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, USA
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA
| |
Collapse
|
14
|
MacLean J, Stirn J, Sisson A, Bidelman GM. Short- and long-term experience-dependent neuroplasticity interact during the perceptual learning of concurrent speech. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.09.26.559640. [PMID: 37808665 PMCID: PMC10557636 DOI: 10.1101/2023.09.26.559640] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/10/2023]
Abstract
Plasticity from auditory experiences shapes brain encoding and perception of sound. However, whether such long-term plasticity alters the trajectory of short-term plasticity during speech processing has yet to be investigated. Here, we explored the neural mechanisms and interplay between short- and long-term neuroplasticity for rapid auditory perceptual learning of concurrent speech sounds in young, normal-hearing musicians and nonmusicians. Participants learned to identify double-vowel mixtures during ∼45 minute training sessions recorded simultaneously with high-density EEG. We analyzed frequency-following responses (FFRs) and event-related potentials (ERPs) to investigate neural correlates of learning at subcortical and cortical levels, respectively. While both groups showed rapid perceptual learning, musicians showed faster behavioral decisions than nonmusicians overall. Learning-related changes were not apparent in brainstem FFRs. However, plasticity was highly evident in cortex, where ERPs revealed unique hemispheric asymmetries between groups suggestive of different neural strategies (musicians: right hemisphere bias; nonmusicians: left hemisphere). Source reconstruction and the early (150-200 ms) time course of these effects localized learning-induced cortical plasticity to auditory-sensory brain areas. Our findings confirm domain-general benefits for musicianship but reveal successful speech sound learning is driven by a critical interplay between long- and short-term mechanisms of auditory plasticity that first emerge at a cortical level.
Collapse
|
15
|
Elmer S, Kurthen I, Meyer M, Giroud N. A multidimensional characterization of the neurocognitive architecture underlying age-related temporal speech processing. Neuroimage 2023; 278:120285. [PMID: 37481009 DOI: 10.1016/j.neuroimage.2023.120285] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2022] [Revised: 07/11/2023] [Accepted: 07/19/2023] [Indexed: 07/24/2023] Open
Abstract
Healthy aging is often associated with speech comprehension difficulties in everyday life situations despite a pure-tone hearing threshold in the normative range. Drawing on this background, we used a multidimensional approach to assess the functional and structural neural correlates underlying age-related temporal speech processing while controlling for pure-tone hearing acuity. Accordingly, we combined structural magnetic resonance imaging and electroencephalography, and collected behavioral data while younger and older adults completed a phonetic categorization and discrimination task with consonant-vowel syllables varying along a voice-onset time continuum. The behavioral results confirmed age-related temporal speech processing singularities which were reflected in a shift of the boundary of the psychometric categorization function, with older adults perceiving more syllable characterized by a short voice-onset time as /ta/ compared to younger adults. Furthermore, despite the absence of any between-group differences in phonetic discrimination abilities, older adults demonstrated longer N100/P200 latencies as well as increased P200 amplitudes while processing the consonant-vowel syllables varying in voice-onset time. Finally, older adults also exhibited a divergent anatomical gray matter infrastructure in bilateral auditory-related and frontal brain regions, as manifested in reduced cortical thickness and surface area. Notably, in the younger adults but not in the older adult cohort, cortical surface area in these two gross anatomical clusters correlated with the categorization of consonant-vowel syllables characterized by a short voice-onset time, suggesting the existence of a critical gray matter threshold that is crucial for consistent mapping of phonetic categories varying along the temporal dimension. Taken together, our results highlight the multifaceted dimensions of age-related temporal speech processing characteristics, and pave the way toward a better understanding of the relationships between hearing, speech and the brain in older age.
Collapse
Affiliation(s)
- Stefan Elmer
- Department of Computational Linguistics, Computational Neuroscience of Speech & Hearing, University of Zurich, Zurich, Switzerland; Competence center Language & Medicine, University of Zurich, Switzerland.
| | - Ira Kurthen
- Department of Computational Linguistics, Computational Neuroscience of Speech & Hearing, University of Zurich, Zurich, Switzerland
| | - Martin Meyer
- Department of Comparative Language Science, University of Zurich, Zurich, Switzerland; Center for Neuroscience Zurich, University and ETH of Zurich, Zurich, Switzerland; Center for the Interdisciplinary Study of Language Evolution (ISLE), University of Zurich, Zurich, Switzerland; Cognitive Psychology Unit, Alpen-Adria University, Klagenfurt, Austria
| | - Nathalie Giroud
- Department of Computational Linguistics, Computational Neuroscience of Speech & Hearing, University of Zurich, Zurich, Switzerland; Center for Neuroscience Zurich, University and ETH of Zurich, Zurich, Switzerland; Competence center Language & Medicine, University of Zurich, Switzerland
| |
Collapse
|
16
|
Amaral MSAD, Zamberlan-Amorin NE, Mendes KDS, Bernal SC, Massuda ET, Hyppolito MA, Reis ACMB. The P300 Auditory Evoked Potential in Cochlear Implant Users: A Scoping Review. Int Arch Otorhinolaryngol 2023; 27:e518-e527. [PMID: 37564465 PMCID: PMC10411132 DOI: 10.1055/s-0042-1744172] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2021] [Accepted: 01/23/2022] [Indexed: 10/17/2022] Open
Abstract
Introduction The P300 auditory evoked potential is a long-latency cortical potential evoked with auditory stimulation, which provides information on neural mechanisms underlying the central auditory processing. Objectives To identify and gather scientific evidence regarding the P300 in adult cochlear implant (CI) users. Data Synthesis A total of 87 articles, 20 of which were selected for this study, were identified and exported to the Rayyan search software. Those 20 articles did not propose a homogeneous methodology, which made comparison more difficult. Most articles (60%) in this review compare CI users with typical hearing people, showing prolonged P300 latency in CI users. Among the studies, 35% show that CI users present a smaller P300 amplitude. Another variable is the influence of the kind of stimulus used to elicit P300, which was prolonged in 30% of the studies that used pure tone stimuli, 10% of the studies that used pure tone and speech stimuli, and 60% of the studies that used speech stimuli. Conclusion This review has contributed with evidence that shows the importance of applying a controlled P300 protocol to diagnose and monitor CI users. Regardless of the stimuli used to elicit P300, we noticed a pattern in the increase in latency and decrease in amplitude in CI users. The user's experience with the CI speech processor over time and the speech test results seem to be related to the P300 latency and amplitude measurements.
Collapse
Affiliation(s)
- Maria Stella Arantes do Amaral
- Department of Ophthalmology, Otorhinolaryngology, and Head and Neck Surgery, Hospital das Clínicas, Faculdade de Medicina de Ribeirão Preto, Universidade de São Paulo, São Paulo, SP, Brazil
| | - Nelma Ellen Zamberlan-Amorin
- Centro Especializado de Otorrinolaringologia e Fonoaudiologia (CEOF), Hospital das Clínicas, Faculdade de Medicina de Ribeirão Preto, Universidade de São Paulo, São Paulo, Brazil
| | - Karina Dal Sasso Mendes
- Department of General and Specialized Nursing, Faculdade de Enfermagem de Ribeirão Preto, Universidade de São Paulo, São Paulo, Brazil
| | - Sarah Carolina Bernal
- Health Sciences Department, Faculdade de Medicina de Ribeirão Preto, Universidade de São Paulo, São Paulo, Brazil
| | - Eduardo Tanaka Massuda
- Department of Ophthalmology, Otorhinolaryngology, and Head and Neck Surgery, Faculdade de Medicina de Ribeirão Preto, Universidade de São Paulo, São Paulo, Brazil
| | - Miguel Angelo Hyppolito
- Department of Ophthalmology, Otorhinolaryngology, and Head and Neck Surgery, Faculdade de Medicina de Ribeirão Preto, Universidade de São Paulo, São Paulo, Brazil
| | | |
Collapse
|
17
|
He D, Buder EH, Bidelman GM. Effects of Syllable Rate on Neuro-Behavioral Synchronization Across Modalities: Brain Oscillations and Speech Productions. NEUROBIOLOGY OF LANGUAGE (CAMBRIDGE, MASS.) 2023; 4:344-360. [PMID: 37229510 PMCID: PMC10205147 DOI: 10.1162/nol_a_00102] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/08/2022] [Accepted: 01/25/2023] [Indexed: 05/27/2023]
Abstract
Considerable work suggests the dominant syllable rhythm of the acoustic envelope is remarkably similar across languages (∼4-5 Hz) and that oscillatory brain activity tracks these quasiperiodic rhythms to facilitate speech processing. However, whether this fundamental periodicity represents a common organizing principle in both auditory and motor systems involved in speech has not been explicitly tested. To evaluate relations between entrainment in the perceptual and production domains, we measured individuals' (i) neuroacoustic tracking of the EEG to speech trains and their (ii) simultaneous and non-simultaneous productions synchronized to syllable rates between 2.5 and 8.5 Hz. Productions made without concurrent auditory presentation isolated motor speech functions more purely. We show that neural synchronization flexibly adapts to the heard stimuli in a rate-dependent manner, but that phase locking is boosted near ∼4.5 Hz, the purported dominant rate of speech. Cued speech productions (recruit sensorimotor interaction) were optimal between 2.5 and 4.5 Hz, suggesting a low-frequency constraint on motor output and/or sensorimotor integration. In contrast, "pure" motor productions (without concurrent sound cues) were most precisely generated at rates of 4.5 and 5.5 Hz, paralleling the neuroacoustic data. Correlations further revealed strong links between receptive (EEG) and production synchronization abilities; individuals with stronger auditory-perceptual entrainment better matched speech rhythms motorically. Together, our findings support an intimate link between exogenous and endogenous rhythmic processing that is optimized at 4-5 Hz in both auditory and motor systems. Parallels across modalities could result from dynamics of the speech motor system coupled with experience-dependent tuning of the perceptual system via the sensorimotor interface.
Collapse
Affiliation(s)
- Deling He
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA
| | - Eugene H. Buder
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA
| | - Gavin M. Bidelman
- Department of Speech, Language and Hearing Sciences, Indiana University, Bloomington, IN, USA
- Program in Neuroscience, Indiana University, Bloomington, IN, USA
| |
Collapse
|
18
|
Elmer S, Besson M, Rodriguez-Fornells A, Giroud N. Foreign speech sound discrimination and associative word learning lead to a fast reconfiguration of resting-state networks. Neuroimage 2023; 271:120026. [PMID: 36921678 DOI: 10.1016/j.neuroimage.2023.120026] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2022] [Revised: 03/09/2023] [Accepted: 03/12/2023] [Indexed: 03/18/2023] Open
Abstract
Learning new words in an unfamiliar language is a complex endeavor that requires the orchestration of multiple perceptual and cognitive functions. Although the neural mechanisms governing word learning are becoming better understood, little is known about the predictive value of resting-state (RS) metrics for foreign word discrimination and word learning attainment. In addition, it is still unknown which of the multistep processes involved in word learning have the potential to rapidly reconfigure RS networks. To address these research questions, we used electroencephalography (EEG), measured forty participants, and examined scalp-based power spectra, source-based spectral density maps and functional connectivity metrics before (RS1), in between (RS2) and after (RS3) a series of tasks which are known to facilitate the acquisition of new words in a foreign language, namely word discrimination, word-referent mapping and semantic generalization. Power spectra at the scalp level consistently revealed a reconfiguration of RS networks as a function of foreign word discrimination (RS1 vs. RS2) and word learning (RS1 vs. RS3) tasks in the delta, lower and upper alpha, and upper beta frequency ranges. Otherwise, functional reconfigurations at the source level were restricted to the theta (spectral density maps) and to the lower and upper alpha frequency bands (spectral density maps and functional connectivity). Notably, scalp RS changes related to the word discrimination tasks (difference between RS2 and RS1) correlated with word discrimination abilities (upper alpha band) and semantic generalization performance (theta and upper alpha bands), whereas functional changes related to the word learning tasks (difference between RS3 and RS1) correlated with word discrimination scores (lower alpha band). Taken together, these results highlight that foreign speech sound discrimination and word learning have the potential to rapidly reconfigure RS networks at multiple functional scales.
Collapse
Affiliation(s)
- Stefan Elmer
- Department of Computational Linguistics, Computational Neuroscience of Speech & Hearing, University of Zurich, Zurich, Switzerland; Bellvitge Biomedical Research Institute, Barcelona, Spain; Competence center Language & Medicine, University of Zurich, Switzerland.
| | - Mireille Besson
- Laboratoire de Neurosciences Cognitives, Université Publique de France, CNRS & Aix-Marseille University, Marseille, France
| | - Antoni Rodriguez-Fornells
- Bellvitge Biomedical Research Institute, Barcelona, Spain; University of Barcelona, Spain; Institució Catalana de Recerca i Estudis Avançats, Barcelona, Spain
| | - Nathalie Giroud
- Department of Computational Linguistics, Computational Neuroscience of Speech & Hearing, University of Zurich, Zurich, Switzerland; Center for Neuroscience Zurich, University and ETH of Zurich, Zurich, Switzerland; Competence center Language & Medicine, University of Zurich, Switzerland
| |
Collapse
|
19
|
Ignatiadis K, Barumerli R, Tóth B, Baumgartner R. Effects of individualized brain anatomies and EEG electrode positions on inferred activity of the primary auditory cortex. Front Neuroinform 2022; 16:970372. [PMID: 36313125 PMCID: PMC9606706 DOI: 10.3389/fninf.2022.970372] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2022] [Accepted: 09/05/2022] [Indexed: 09/07/2024] Open
Abstract
Due to its high temporal resolution and non-invasive nature, electroencephalography (EEG) is considered a method of great value for the field of auditory cognitive neuroscience. In performing source space analyses, localization accuracy poses a bottleneck, which precise forward models based on individualized attributes such as subject anatomy or electrode locations aim to overcome. Yet acquiring anatomical images or localizing EEG electrodes requires significant additional funds and processing time, making it an oftentimes inaccessible asset. Neuroscientific software offers template solutions, on which analyses can be based. For localizing the source of auditory evoked responses, we here compared the results of employing such template anatomies and electrode positions versus the subject-specific ones, as well as combinations of the two. All considered cases represented approaches commonly used in electrophysiological studies. We considered differences between two commonly used inverse solutions (dSPM, sLORETA) and targeted the primary auditory cortex; a notoriously small cortical region that is located within the lateral sulcus, thus particularly prone to errors in localization. Through systematical comparison of early evoked component metrics and spatial leakage, we assessed how the individualization steps impacted the analyses outcomes. Both electrode locations as well as subject anatomies were found to have an effect, which though varied based on the configuration considered. When comparing the inverse solutions, we moreover found that dSPM more consistently benefited from individualization of subject morphologies compared to sLORETA, suggesting it to be the better choice for auditory cortex localization.
Collapse
Affiliation(s)
| | - Roberto Barumerli
- Acoustics Research Institute, Austrian Academy of Sciences, Vienna, Austria
| | - Brigitta Tóth
- Institute of Cognitive Neuroscience and Psychology, Research Centre for Natural Sciences, Budapest, Hungary
| | - Robert Baumgartner
- Acoustics Research Institute, Austrian Academy of Sciences, Vienna, Austria
| |
Collapse
|
20
|
Corina DP, Coffey-Corina S, Pierotti E, Bormann B, LaMarr T, Lawyer L, Backer KC, Miller LM. Electrophysiological Examination of Ambient Speech Processing in Children With Cochlear Implants. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:3502-3517. [PMID: 36037517 PMCID: PMC9913291 DOI: 10.1044/2022_jslhr-22-00004] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/17/2022] [Revised: 05/05/2022] [Accepted: 06/11/2022] [Indexed: 06/15/2023]
Abstract
PURPOSE This research examined the expression of cortical auditory evoked potentials in a cohort of children who received cochlear implants (CIs) for treatment of congenital deafness (n = 28) and typically hearing controls (n = 28). METHOD We make use of a novel electroencephalography paradigm that permits the assessment of auditory responses to ambiently presented speech and evaluates the contributions of concurrent visual stimulation on this activity. RESULTS Our findings show group differences in the expression of auditory sensory and perceptual event-related potential components occurring in 80- to 200-ms and 200- to 300-ms time windows, with reductions in amplitude and a greater latency difference for CI-using children. Relative to typically hearing children, current source density analysis showed muted responses to concurrent visual stimulation in CI-using children, suggesting less cortical specialization and/or reduced responsiveness to auditory information that limits the detection of the interaction between sensory systems. CONCLUSION These findings indicate that even in the face of early interventions, CI-using children may exhibit disruptions in the development of auditory and multisensory processing.
Collapse
Affiliation(s)
- David P. Corina
- Department of Linguistics, University of California, Davis
- Department of Psychology, University of California, Davis
- Center for Mind and Brain, University of California, Davis
| | | | - Elizabeth Pierotti
- Department of Psychology, University of California, Davis
- Center for Mind and Brain, University of California, Davis
| | - Brett Bormann
- Center for Mind and Brain, University of California, Davis
- Neurobiology, Physiology and Behavior, University of California, Davis
| | - Todd LaMarr
- Center for Mind and Brain, University of California, Davis
| | - Laurel Lawyer
- Center for Mind and Brain, University of California, Davis
| | | | - Lee M. Miller
- Center for Mind and Brain, University of California, Davis
- Neurobiology, Physiology and Behavior, University of California, Davis
- Department of Otolaryngology/Head and Neck Surgery, University of California, Davis
| |
Collapse
|
21
|
Bidelman GM, Chow R, Noly-Gandon A, Ryan JD, Bell KL, Rizzi R, Alain C. Transcranial Direct Current Stimulation Combined With Listening to Preferred Music Alters Cortical Speech Processing in Older Adults. Front Neurosci 2022; 16:884130. [PMID: 35873829 PMCID: PMC9298650 DOI: 10.3389/fnins.2022.884130] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2022] [Accepted: 06/17/2022] [Indexed: 11/13/2022] Open
Abstract
Emerging evidence suggests transcranial direct current stimulation (tDCS) can improve cognitive performance in older adults. Similarly, music listening may improve arousal and stimulate subsequent performance on memory-related tasks. We examined the synergistic effects of tDCS paired with music listening on auditory neurobehavioral measures to investigate causal evidence of short-term plasticity in speech processing among older adults. In a randomized sham-controlled crossover study, we measured how combined anodal tDCS over dorsolateral prefrontal cortex (DLPFC) paired with listening to autobiographically salient music alters neural speech processing in older adults compared to either music listening (sham stimulation) or tDCS alone. EEG assays included both frequency-following responses (FFRs) and auditory event-related potentials (ERPs) to trace neuromodulation-related changes at brainstem and cortical levels. Relative to music without tDCS (sham), we found tDCS alone (without music) modulates the early cortical neural encoding of speech in the time frame of ∼100-150 ms. Whereas tDCS by itself appeared to largely produce suppressive effects (i.e., reducing ERP amplitude), concurrent music with tDCS restored responses to those of the music+sham levels. However, the interpretation of this effect is somewhat ambiguous as this neural modulation could be attributable to a true effect of tDCS or presence/absence music. Still, the combined benefit of tDCS+music (above tDCS alone) was correlated with listeners' education level suggesting the benefit of neurostimulation paired with music might depend on listener demographics. tDCS changes in speech-FFRs were not observed with DLPFC stimulation. Improvements in working memory pre to post session were also associated with better speech-in-noise listening skills. Our findings provide new causal evidence that combined tDCS+music relative to tDCS-alone (i) modulates the early (100-150 ms) cortical encoding of speech and (ii) improves working memory, a cognitive skill which may indirectly bolster noise-degraded speech perception in older listeners.
Collapse
Affiliation(s)
- Gavin M. Bidelman
- Department of Speech, Language and Hearing Sciences, Indiana University Bloomington, Bloomington, IN, United States
- School of Communication Sciences and Disorders, The University of Memphis, Memphis, TN, United States
| | - Ricky Chow
- Rotman Research Institute, Baycrest Centre, Toronto, ON, Canada
| | | | - Jennifer D. Ryan
- Rotman Research Institute, Baycrest Centre, Toronto, ON, Canada
- Department of Psychology, University of Toronto, Toronto, ON, Canada
- Department of Psychiatry, University of Toronto, Toronto, ON, Canada
- Institute of Medical Science, University of Toronto, Toronto, ON, Canada
| | - Karen L. Bell
- Department of Audiology, San José State University, San Jose, CA, United States
| | - Rose Rizzi
- Department of Speech, Language and Hearing Sciences, Indiana University Bloomington, Bloomington, IN, United States
- School of Communication Sciences and Disorders, The University of Memphis, Memphis, TN, United States
| | - Claude Alain
- Rotman Research Institute, Baycrest Centre, Toronto, ON, Canada
- Department of Psychology, University of Toronto, Toronto, ON, Canada
- Institute of Medical Science, University of Toronto, Toronto, ON, Canada
- Music and Health Science Research Collaboratory, University of Toronto, Toronto, ON, Canada
| |
Collapse
|
22
|
Aksu S, Uslu A, İşçen P, Tülay EE, Barham H, Soyata AZ, Demirtas-Tatlidede A, Yıldız GB, Bilgiç B, Hanağası H, Woods AJ, Karamürsel S, Uyar FA. Does transcranial direct current stimulation enhance cognitive performance in Parkinson's disease mild cognitive impairment? An event-related potentials and neuropsychological assessment study. Neurol Sci 2022; 43:4029-4044. [PMID: 35322340 DOI: 10.1007/s10072-022-06020-z] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2021] [Accepted: 03/16/2022] [Indexed: 11/29/2022]
Abstract
BACKGROUND Parkinson's disease-mild cognitive impairment (PD-MCI) is garnering attention as a key interventional period for cognitive impairment. Currently, there are no approved treatments for PD-MCI and encouraging results of transcranial direct current stimulation (tDCS) combined with other interventions have been proposed, though the efficacy and neural mechanisms of tDCS alone have not been studied in PD-MCI yet. OBJECTIVES The present double-blind, randomized, sham-controlled study assessed the effects of tDCS over the dorsolateral prefrontal cortex on cognitive functions via neuropsychological and electrophysiological evaluations in individuals with PD-MCI for the first time. METHOD Twenty-six individuals with PD-MCI were administered 10 sessions of active (n = 13) or sham (n = 13) prefrontal tDCS twice a day, for 5 days. Changes were tested through a comprehensive neuropsychological battery and event-related potential recordings, which were performed before, immediately, and 1 month after the administrations. RESULTS Neuropsychological assessment showed an improvement in delayed recall and executive functions in the active group. N1 amplitudes in response to targets in the oddball test-likely indexing attention and discriminability and NoGo N2 amplitudes in the continuous performance test-likely indexing cognitive control and conflict monitoring increased in the active group. Active stimulation elicited higher benefits 1 month after the administrations. CONCLUSION The present findings substantiate the efficacy of tDCS on cognitive control and episodic memory, along with the neural underpinnings of cognitive control, highlighting its potential for therapeutic utility in PD-MCI. TRIAL REGISTRATION NCT 04,171,804. Date of registration: 21/11/2019.
Collapse
Affiliation(s)
- Serkan Aksu
- Department of Physiology, Graduate School of Health Sciences, Istanbul University, Istanbul, Turkey.
- Department of Physiology, Istanbul Faculty of Medicine, Istanbul University, Istanbul, Turkey.
- Department of Physiology, Faculty of Medicine, Muğla Sıtkı Koçman University, Muğla, Turkey.
| | - Atilla Uslu
- Department of Physiology, Istanbul Faculty of Medicine, Istanbul University, Istanbul, Turkey
| | - Pınar İşçen
- Department of Neurology, Istanbul Faculty of Medicine, Istanbul University, Istanbul, Turkey
| | - Emine Elif Tülay
- Department of Software Engineering, Faculty of Engineering, Muğla Sıtkı Koçman University, Muğla, Turkey
| | - Huzeyfe Barham
- Department of Psychiatry, Kırklareli Research and Training Hospital, Kırklareli, Turkey
| | | | | | | | - Başar Bilgiç
- Department of Neurology, Istanbul Faculty of Medicine, Istanbul University, Istanbul, Turkey
| | - Haşmet Hanağası
- Department of Neurology, Istanbul Faculty of Medicine, Istanbul University, Istanbul, Turkey
| | - Adam J Woods
- Department of Clinical and Health Psychology, Center for Cognitive Aging and Memory, McKnight Brain Institute, Cognitive Aging and Memory Clinical Translational Research Program, University of Florida, Gainesville, USA
| | - Sacit Karamürsel
- Department of Physiology, School of Medicine, Koç University, Istanbul, Turkey
| | - Fatma Aytül Uyar
- Department of Physiology, Graduate School of Health Sciences, Istanbul University, Istanbul, Turkey
- Department of Physiology, Istanbul Faculty of Medicine, Istanbul University, Istanbul, Turkey
| |
Collapse
|
23
|
Muncke J, Kuruvila I, Hoppe U. Prediction of Speech Intelligibility by Means of EEG Responses to Sentences in Noise. Front Neurosci 2022; 16:876421. [PMID: 35720724 PMCID: PMC9198593 DOI: 10.3389/fnins.2022.876421] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2022] [Accepted: 03/13/2022] [Indexed: 11/13/2022] Open
Abstract
Objective Understanding speech in noisy conditions is challenging even for people with mild hearing loss, and intelligibility for an individual person is usually evaluated by using several subjective test methods. In the last few years, a method has been developed to determine a temporal response function (TRF) between speech envelope and simultaneous electroencephalographic (EEG) measurements. By using this TRF it is possible to predict the EEG signal for any speech signal. Recent studies have suggested that the accuracy of this prediction varies with the level of noise added to the speech signal and can predict objectively the individual speech intelligibility. Here we assess the variations of the TRF itself when it is calculated for measurements with different signal-to-noise ratios and apply these variations to predict speech intelligibility. Methods For 18 normal hearing subjects the individual threshold of 50% speech intelligibility was determined by using a speech in noise test. Additionally, subjects listened passively to speech material of the speech in noise test at different signal-to-noise ratios close to individual threshold of 50% speech intelligibility while an EEG was recorded. Afterwards the shape of TRFs for each signal-to-noise ratio and subject were compared with the derived intelligibility. Results The strongest effect of variations in stimulus signal-to-noise ratio on the TRF shape occurred close to 100 ms after the stimulus presentation, and was located in the left central scalp region. The investigated variations in TRF morphology showed a strong correlation with speech intelligibility, and we were able to predict the individual threshold of 50% speech intelligibility with a mean deviation of less then 1.5 dB. Conclusion The intelligibility of speech in noise can be predicted by analyzing the shape of the TRF derived from different stimulus signal-to-noise ratios. Because TRFs are interpretable, in a manner similar to auditory evoked potentials, this method offers new options for clinical diagnostics.
Collapse
Affiliation(s)
- Jan Muncke
- Department of Audiology, ENT-Clinic, University Hospital Erlangen, Erlangen, Germany
| | - Ivine Kuruvila
- Department of Audiology, ENT-Clinic, University Hospital Erlangen, Erlangen, Germany
- WS Audiology, Erlangen, Germany
| | - Ulrich Hoppe
- Department of Audiology, ENT-Clinic, University Hospital Erlangen, Erlangen, Germany
| |
Collapse
|
24
|
ElShafei HA, Masson R, Fakche C, Fornoni L, Moulin A, Caclin A, Bidet-Caulet A. Age-related differences in bottom-up and top-down attention: Insights from EEG and MEG. Eur J Neurosci 2022; 55:1215-1231. [PMID: 35112420 PMCID: PMC9303169 DOI: 10.1111/ejn.15617] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2020] [Revised: 01/21/2022] [Accepted: 01/29/2022] [Indexed: 11/26/2022]
Abstract
Attention operates through top‐down and bottom‐up processes, and a balance between these processes is crucial for daily tasks. Imperilling such balance could explain ageing‐associated attentional problems such as exacerbated distractibility. In this study, we aimed to characterize this enhanced distractibility by investigating the impact of ageing upon event‐related components associated with top‐down and bottom‐up attentional processes. MEG and EEG data were acquired from 14 older and 14 younger healthy adults while performing a task that conjointly evaluates top‐down and bottom‐up attention. Event‐related components were analysed on sensor and source levels. In comparison with the younger group, the older mainly displayed (1) reduced target anticipation processes (reduced CMV), (2) increased early target processing (larger P50 but smaller N1) and (3) increased processing of early distracting sounds (larger N1 but reduced P3a), followed by a (4) prolonged reorientation towards the main task (larger RON). Taken together, our results suggest that the enhanced distractibility in ageing could stem from top‐down deficits, in particular from reduced inhibitory and reorientation processes.
Collapse
Affiliation(s)
- Hesham A ElShafei
- Lyon Neuroscience Research Center; CRNL, INSERM U1028, CNRS UMR5292, University of Lyon 1, Université de Lyon, Lyon, France.,Donders Institute for Brain, Cognition & Behavior, Radboud University, Nijmegen, The Netherlands.,Donders Centre for Cognitive Neuroimaging, EN, Nijmegen, Netherlands
| | - Rémy Masson
- Lyon Neuroscience Research Center; CRNL, INSERM U1028, CNRS UMR5292, University of Lyon 1, Université de Lyon, Lyon, France
| | - Camille Fakche
- Lyon Neuroscience Research Center; CRNL, INSERM U1028, CNRS UMR5292, University of Lyon 1, Université de Lyon, Lyon, France
| | - Lesly Fornoni
- Lyon Neuroscience Research Center; CRNL, INSERM U1028, CNRS UMR5292, University of Lyon 1, Université de Lyon, Lyon, France
| | - Annie Moulin
- Lyon Neuroscience Research Center; CRNL, INSERM U1028, CNRS UMR5292, University of Lyon 1, Université de Lyon, Lyon, France
| | - Anne Caclin
- Lyon Neuroscience Research Center; CRNL, INSERM U1028, CNRS UMR5292, University of Lyon 1, Université de Lyon, Lyon, France
| | - Aurélie Bidet-Caulet
- Lyon Neuroscience Research Center; CRNL, INSERM U1028, CNRS UMR5292, University of Lyon 1, Université de Lyon, Lyon, France
| |
Collapse
|
25
|
Dercksen TT, Widmann A, Scharf F, Wetzel N. Sound omission related brain responses in children. Dev Cogn Neurosci 2022; 53:101045. [PMID: 34923314 PMCID: PMC8688889 DOI: 10.1016/j.dcn.2021.101045] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/04/2021] [Revised: 10/21/2021] [Accepted: 12/09/2021] [Indexed: 11/03/2022] Open
Abstract
Action is an important way for children to learn about the world. Recent theories suggest that action is inherently accompanied by the sensory prediction of its effects. Such predictions can be revealed by rarely omitting the expected sensory consequence of the action, resulting in an omission response that is observable in the EEG. Although prediction errors play an important role in models of learning and development, little is known about omission-related brain responses in children. This study used a motor-auditory omission paradigm, testing a group of 6-8-year-old children and an adult group (N = 31 each). In an identity-specific condition, the sound coupled to the motor action was predictable, while in an identity unspecific condition the sound was unpredictable. Results of a temporal principal component analysis revealed that sound-related brain responses underlying the N1-complex differed considerably between age groups. Despite these developmental differences, omission responses (oN1) were similar between age groups. Two subcomponents of the oN1 were differently affected by specific and unspecific predictions. Results demonstrate that children, independent from the maturation of sound processing mechanisms, can implement specific and unspecific predictions as flexibly as adults. This supports theories that regard action and prediction error as important drivers of cognitive development.
Collapse
Affiliation(s)
- Tjerk T Dercksen
- Leibniz Institute for Neurobiology, Brenneckestraße 6, 39118 Magdeburg, Germany; Center for Behavioral Brain Sciences, Universitätsplatz 2, D-39106 Magdeburg, Germany.
| | - Andreas Widmann
- Leibniz Institute for Neurobiology, Brenneckestraße 6, 39118 Magdeburg, Germany; Leipzig University, Neumarkt 9-19, D-04109 Leipzig, Germany
| | - Florian Scharf
- University of Münster, Fliednerstraße 21, 48149 Münster, Germany
| | - Nicole Wetzel
- Leibniz Institute for Neurobiology, Brenneckestraße 6, 39118 Magdeburg, Germany; Center for Behavioral Brain Sciences, Universitätsplatz 2, D-39106 Magdeburg, Germany; University of Applied Sciences Magdeburg-Stendal, Osterburgerstraße 25, 39576 Stendal, Germany
| |
Collapse
|
26
|
Krishnan A, Suresh CH, Gandour JT. Cortical hemisphere preference and brainstem ear asymmetry reflect experience-dependent functional modulation of pitch. BRAIN AND LANGUAGE 2021; 221:104995. [PMID: 34303110 PMCID: PMC8559596 DOI: 10.1016/j.bandl.2021.104995] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/19/2021] [Revised: 05/07/2021] [Accepted: 07/07/2021] [Indexed: 06/13/2023]
Abstract
Temporal attributes of pitch processing at cortical and subcortical levels are differentially weighted and well-coordinated. The question is whether language experience induces functional modulation of hemispheric preference complemented by brainstem ear symmetry for pitch processing. Brainstem frequency-following and cortical pitch responses were recorded concurrently from Mandarin and English participants. A Mandarin syllable with a rising pitch contour was presented to both ears with monaural stimulation. At the cortical level, left ear stimulation in the Chinese group revealed an experience-dependent response for pitch processing in the right hemisphere, consistent with a functionalaccount. The English group revealed a contralateral hemisphere preference consistent with a structuralaccount. At the brainstem level, Chinese participants showed a functional leftward ear asymmetry, whereas English were consistent with a structural account. Overall, language experience modulates both cortical hemispheric preference and brainstem ear asymmetry in a complementary manner to optimize processing of temporal attributes of pitch.
Collapse
Affiliation(s)
- Ananthanarayan Krishnan
- Department of Speech Language Hearing Sciences, Purdue University, Lyles Porter Hall, 715 Clinic Drive, West Lafayette, IN 47907, USA.
| | - Chandan H Suresh
- Department of Speech Language Hearing Sciences, Purdue University, Lyles Porter Hall, 715 Clinic Drive, West Lafayette, IN 47907, USA; Department of Communication Disorders, California State, University, 5151 State University Drive, Los Angeles, CA 90032, USA.
| | - Jackson T Gandour
- Department of Speech Language Hearing Sciences, Purdue University, Lyles Porter Hall, 715 Clinic Drive, West Lafayette, IN 47907, USA.
| |
Collapse
|
27
|
Ghani U, Signal N, Niazi IK, Taylor D. Efficacy of a Single-Task ERP Measure to Evaluate Cognitive Workload During a Novel Exergame. Front Hum Neurosci 2021; 15:742384. [PMID: 34566610 PMCID: PMC8456040 DOI: 10.3389/fnhum.2021.742384] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2021] [Accepted: 08/18/2021] [Indexed: 11/24/2022] Open
Abstract
This study aimed to validate the efficacy of single-task event-related potential (ERP) measures of cognitive workload to be implemented in exergame-based rehabilitation. Twenty-four healthy participants took part in a novel gamified balance task where task-irrelevant auditory tones were presented in the background to generate ERPs in the participants’ electroencephalogram (EEG) as a measure of cognitive workload. For the balance task, a computer-based tilt-ball game was combined with a balance board. Participants played the game by shifting their weight to tilt the balance board, which moved a virtual ball to score goals. The game was manipulated by adjusting the size of the goalposts to set three predefined levels of game difficulty (easy, medium, and hard). The participant’s experience of game difficulty was evaluated based on the number of goals scored and their subjective reporting of perceived difficulty. Participants experienced a significant difference in the three levels of task difficulty based on the number of goals scored and perceived difficulty (p < 0.001). Post hoc analysis revealed the lowest performance for the hardest level. The mean amplitude of the N1 ERP component was used to measure the cognitive workload associated with the three difficulty levels. The N1 component’s amplitude decreased significantly (p < 0.001), with an increase in the task difficulty. Moreover, the amplitude of the N1 component for the hard level was significantly smaller compared to medium (p = 0.0003) and easy (p < 0.001) levels. These results support the efficacy of the N1 ERP component to measure cognitive workload in dynamic and real-life scenarios such as exergames and other rehabilitation exercises.
Collapse
Affiliation(s)
- Usman Ghani
- Rehabilitation Innovation Centre, Auckland University of Technology, Auckland, New Zealand
| | - Nada Signal
- Rehabilitation Innovation Centre, Auckland University of Technology, Auckland, New Zealand
| | - Imran Khan Niazi
- Rehabilitation Innovation Centre, Auckland University of Technology, Auckland, New Zealand.,Department of Health Science and Technology, Aalborg University, Aalborg, Denmark.,Centre for Chiropractic Research, New Zealand College of Chiropractic, Auckland, New Zealand
| | - Denise Taylor
- Rehabilitation Innovation Centre, Auckland University of Technology, Auckland, New Zealand
| |
Collapse
|
28
|
Momtaz S, Moncrieff D, Bidelman GM. Dichotic listening deficits in amblyaudia are characterized by aberrant neural oscillations in auditory cortex. Clin Neurophysiol 2021; 132:2152-2162. [PMID: 34284251 DOI: 10.1016/j.clinph.2021.04.022] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2021] [Revised: 04/16/2021] [Accepted: 04/29/2021] [Indexed: 12/25/2022]
Abstract
OBJECTIVE Children diagnosed with auditory processing disorder (APD) show deficits in processing complex sounds that are associated with difficulties in higher-order language, learning, cognitive, and communicative functions. Amblyaudia (AMB) is a subcategory of APD characterized by abnormally large ear asymmetries in dichotic listening tasks. METHODS Here, we examined frequency-specific neural oscillations and functional connectivity via high-density electroencephalography (EEG) in children with and without AMB during passive listening of nonspeech stimuli. RESULTS Time-frequency maps of these "brain rhythms" revealed stronger phase-locked beta-gamma (~35 Hz) oscillations in AMB participants within bilateral auditory cortex for sounds presented to the right ear, suggesting a hypersynchronization and imbalance of auditory neural activity. Brain-behavior correlations revealed neural asymmetries in cortical responses predicted the larger than normal right-ear advantage seen in participants with AMB. Additionally, we found weaker functional connectivity in the AMB group from right to left auditory cortex, despite their stronger neural responses overall. CONCLUSION Our results reveal abnormally large auditory sensory encoding and an imbalance in communication between cerebral hemispheres (ipsi- to -contralateral signaling) in AMB. SIGNIFICANCE These neurophysiological changes might lead to the functionally poorer behavioral capacity to integrate information between the two ears in children with AMB.
Collapse
Affiliation(s)
- Sara Momtaz
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA.
| | - Deborah Moncrieff
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA
| | - Gavin M Bidelman
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA; University of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, USA
| |
Collapse
|
29
|
Wang Y, Wan C, Zhang Y, Zhou Y, Wang H, Yan F, Song D, Du R, Wang Q, Huang L. Detecting Connected Consciousness During Propofol-Induced Anesthesia Using EEG Based Brain Decoding. Int J Neural Syst 2021; 31:2150021. [PMID: 33970056 DOI: 10.1142/s0129065721500210] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/29/2022]
Abstract
Connected consciousness refers to the state when external stimuli can enter into the stream of our consciousness experience. Emerging evidence suggests that although patients may not respond behaviorally to external stimuli during anesthesia, they may be aware of their surroundings. In this work, we investigated whether EEG based brain decoding could be used for detecting connected consciousness in the absence of behavioral responses during propofol infusion. A total of 14 subjects participated in our experiment. Subjects were asked to discriminate two types of auditory stimuli with a finger press during an ultraslow propofol infusion. We trained an EEG based brain decoding model using data collected in the awakened state using the same auditory stimuli and tested the model on data collected during the propofol infusion. The model provided a correct classification rate (CCR) of [Formula: see text]% when subjects were able to respond to the stimuli during the propofol infusion. The CCR dropped to [Formula: see text]% when subjects ceased responding and further decreased to [Formula: see text]% when we increased the propofol concentration by another 0.2 [Formula: see text]g/ml. After terminating the propofol infusion, we observed that the CCR rebounded to [Formula: see text]% before the subjects regained consciousness. With the classification results, we provided evidence that loss of consciousness is a gradual process and may progress from full consciousness to connected consciousness and then to disconnected consciousness.
Collapse
Affiliation(s)
- Yubo Wang
- School of Life Science and Technology, Xidian University, Xi'an, P. R. China
| | - Chenghao Wan
- School of Life Science and Technology, Xidian University, Xi'an, P. R. China
| | - Yun Zhang
- School of Life Science and Technology, Xidian University, Xi'an, P. R. China
| | - Yu Zhou
- School of Life Science and Technology, Xidian University, Xi'an, P. R. China
| | - Haidong Wang
- School of Life Science and Technology, Xidian University, Xi'an, P. R. China
| | - Fei Yan
- Department of Anesthesiology and Center for Brain Science, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, P. R. China
| | - Dawei Song
- Department of Anesthesiology and Center for Brain Science, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, P. R. China
| | - Ruini Du
- Department of Anesthesiology and Center for Brain Science, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, P. R. China
| | - Qiang Wang
- Department of Anesthesiology and Center for Brain Science, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, P. R. China
| | - Liyu Huang
- School of Life Science and Technology, Xidian University, Xi'an, P. R. China
| |
Collapse
|
30
|
Elmer S, Valizadeh SA, Cunillera T, Rodriguez-Fornells A. Statistical learning and prosodic bootstrapping differentially affect neural synchronization during speech segmentation. Neuroimage 2021; 235:118051. [PMID: 33848624 DOI: 10.1016/j.neuroimage.2021.118051] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2020] [Revised: 03/12/2021] [Accepted: 04/05/2021] [Indexed: 10/21/2022] Open
Abstract
Neural oscillations constitute an intrinsic property of functional brain organization that facilitates the tracking of linguistic units at multiple time scales through brain-to-stimulus alignment. This ubiquitous neural principle has been shown to facilitate speech segmentation and word learning based on statistical regularities. However, there is no common agreement yet on whether speech segmentation is mediated by a transition of neural synchronization from syllable to word rate, or whether the two time scales are concurrently tracked. Furthermore, it is currently unknown whether syllable transition probability contributes to speech segmentation when lexical stress cues can be directly used to extract word forms. Using Inter-Trial Coherence (ITC) analyses in combinations with Event-Related Potentials (ERPs), we showed that speech segmentation based on both statistical regularities and lexical stress cues was accompanied by concurrent neural synchronization to syllables and words. In particular, ITC at the word rate was generally higher in structured compared to random sequences, and this effect was particularly pronounced in the flat condition. Furthermore, ITC at the syllable rate dynamically increased across the blocks of the flat condition, whereas a similar modulation was not observed in the stressed condition. Notably, in the flat condition ITC at both time scales correlated with each other, and changes in neural synchronization were accompanied by a rapid reconfiguration of the P200 and N400 components with a close relationship between ITC and ERPs. These results highlight distinct computational principles governing neural synchronization to pertinent linguistic units while segmenting speech under different listening conditions.
Collapse
Affiliation(s)
- Stefan Elmer
- Auditory Research Group Zurich (ARGZ), Division Neuropsychology, Institute of Psychology, University of Zurich, Binzmühlestrasse 14/25, Zurich 8050, Switzerland; Cognition and Brain Plasticity Group, Bellvitge Biomedical Research Institute, L'Hospitalet de Llobregat, Barcelona 08097, Spain.
| | - Seyed Abolfazl Valizadeh
- Auditory Research Group Zurich (ARGZ), Division Neuropsychology, Institute of Psychology, University of Zurich, Binzmühlestrasse 14/25, Zurich 8050, Switzerland; Department of Internal Medicine, University Hospital, University of Zurich, Zurich 8091, Switzerland; University Research Priority Program, "Dynamics of Healthy Aging", University of Zurich, Zurich 8050, Switzerland.
| | - Toni Cunillera
- Department of Cognition, Development and Educational Psychology, Barcelona 08035, University of Barcelona, Spain.
| | - Antoni Rodriguez-Fornells
- Department of Cognition, Development and Educational Psychology, Campus Bellvitge, University of Barcelona, 5L'Hospitalet de Llobregat, Barcelona 08097, Spain; Cognition and Brain Plasticity Group, Bellvitge Biomedical Research Institute, L'Hospitalet de Llobregat, Barcelona 08097, Spain; Institució Catalana de Recerca i Estudis Avançats, ICREA, Barcelona 08010, Spain.
| |
Collapse
|
31
|
Development of central auditory processes in Polish children and adolescents at the age from 7 to 16 years. CURRENT PSYCHOLOGY 2021. [DOI: 10.1007/s12144-021-01540-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
Abstract
AbstractThere are discrepancies in the literature regarding the course of central auditory processes (CAP) maturation in typically developing children and adolescents. The purpose of the study was to provide an overview of age – related improvement in CAP in Polish primary and secondary school students aged 7–16 years. 180 children/adolescents, subdivided into 9 age categories, and 20 adults (aged 18–24 years) performed the Dichotic Digit Test (DDT), Duration Pattern Test (DPT), Frequency Pattern Test (FPT), Gap Detection Test (GDT) and adaptive Speech-in-Noise (aSpN). The 12-year-olds was retested after w week. We found the age effects only for the DDT, DPT and FPT. In the right ear DDT the 7-year-olds performed more poorly than all groups ≥12. In the left ear DDT both 7- and 8-year-olds achieved less correct responses compared with the 13-, 14-, 15-year-olds and with the adults. The right ear advantage was greater in the 7-year-olds than in the 15-year-olds and adult group. At the age of 7 there was lower DPT and FPT scores than in all participants ≥13 whereas the 8-year-olds obtained less correct responses in the FPT than all age categories ≥12. Almost all groups (except for the 7-year-olds) performed better in the DPT than FPT. The test-retest reliability for all tests was satisfactory. The study demonstrated that different CAP have their own patterns of improvement with age and some of them are specific for the Polish population. The psychoacoustic battery may be useful in screening for CAP disorders in Poland.
Collapse
|
32
|
Kamal F, Morrison C, Campbell K, Taler V. Event-related potential evidence that very slowly presented auditory stimuli are passively processed differently in younger and older adults. Neurobiol Aging 2021; 103:12-21. [PMID: 33774574 DOI: 10.1016/j.neurobiolaging.2021.02.014] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/09/2020] [Revised: 02/16/2021] [Accepted: 02/18/2021] [Indexed: 12/31/2022]
Abstract
The occurrence of a very infrequent and unattended auditory stimulus is highly salient and may result in an interruption of the frontoparietal network controlling processing priorities. Research has suggested that older adults may be unable to compute the level of salience of unattended stimulus inputs. A multi-channel EEG was recorded in 20 younger adults and 20 older adults. In different conditions, a single 80 dB SPL auditory stimulus was presented relatively rapidly, every 1.5 s or very slowly, every 12.0 s. Participants ignored the auditory stimuli while watching a silent video. When the stimuli were presented rapidly, group differences were not observed for the amplitudes of N1 and P2, which peaked at 100 and 180 ms respectively. When stimuli were presented very slowly, their amplitudes were much enhanced for younger adults, but did not increase for older adults. The failure to observe a large increase in the amplitude of N1 and P2 in older adults for very slowly presented auditory stimuli provides strong evidence of a dysfunction of the salience network in this group.
Collapse
Affiliation(s)
- Farooq Kamal
- School of Psychology, University of Ottawa, Ottawa, Ontario, Canada; Bruyère Research Institute, Ottawa, Ontario, Canada
| | - Cassandra Morrison
- School of Psychology, University of Ottawa, Ottawa, Ontario, Canada; Bruyère Research Institute, Ottawa, Ontario, Canada
| | - Kenneth Campbell
- School of Psychology, University of Ottawa, Ottawa, Ontario, Canada; Bruyère Research Institute, Ottawa, Ontario, Canada
| | - Vanessa Taler
- School of Psychology, University of Ottawa, Ottawa, Ontario, Canada; Bruyère Research Institute, Ottawa, Ontario, Canada.
| |
Collapse
|
33
|
Wang S, Lin M, Sun L, Chen X, Fu X, Yan L, Li C, Zhang X. Neural Mechanisms of Hearing Recovery for Cochlear-Implanted Patients: An Electroencephalogram Follow-Up Study. Front Neurosci 2021; 14:624484. [PMID: 33633529 PMCID: PMC7901906 DOI: 10.3389/fnins.2020.624484] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2020] [Accepted: 12/22/2020] [Indexed: 12/11/2022] Open
Abstract
Background Patients with severe profound hearing loss could benefit from cochlear implantation (CI). However, the neural mechanism of such benefit is still unclear. Therefore, we analyzed the electroencephalogram (EEG) and behavioral indicators of auditory function remodeling in patients with CI. Both indicators were sampled at multiple time points after implantation (1, 90, and 180 days). Methods First, the speech perception ability was evaluated with the recording of a list of Chinese words and sentences in 15 healthy controls (HC group) and 10 patients with CI (CI group). EEG data were collected using an oddball paradigm. Then, the characteristics of event-related potentials (ERPs) and mismatch negative (MMN) were compared between the CI group and the HC group. In addition, we analyzed the phase lag indices (PLI) in the CI group and the HC group and calculated the difference in functional connectivity between the two groups at different stages after implantation. Results The behavioral indicator, speech recognition ability, in CI patients improved as the implantation time increased. The MMN analysis showed that CI patients could recognize the difference between standard and deviation stimuli just like the HCs 90 days after cochlear implantation. Comparing the latencies of N1/P2/MMN between the CI group and the HC group, we found that the latency of N1/P2 in CI patients was longer, while the latency of MMN in CI users was shorter. In addition, PLI-based whole-brain functional connectivity (PLI-FC) showed that the difference between the CI group and the HC group mainly exists in electrode pairs between the bilateral auditory area and the frontal area. Furthermore, all those differences gradually decreased with the increase in implantation time. Conclusion The N1 amplitude, N1/P2/MMN latency, and PLI-FC in the alpha band may reflect the process of auditory function remodeling and could be an objective index for the assessment of speech perception ability and the effect of cochlear implantation.
Collapse
Affiliation(s)
- Songjian Wang
- School of Biomedical Engineering, Capital Medical University, Beijing, China.,Beijing Key Laboratory of Fundamental Research on Biomechanics in Clinical Application, Capital Medical University, Beijing, China
| | - Meng Lin
- School of Biomedical Engineering, Capital Medical University, Beijing, China.,Beijing Key Laboratory of Fundamental Research on Biomechanics in Clinical Application, Capital Medical University, Beijing, China
| | - Liwei Sun
- School of Biomedical Engineering, Capital Medical University, Beijing, China.,Beijing Key Laboratory of Fundamental Research on Biomechanics in Clinical Application, Capital Medical University, Beijing, China
| | - Xueqing Chen
- Key Laboratory of Otolaryngology Head and Neck Surgery, Beijing Tongren Hospital, Beijing Institute of Otolaryngology, Ministry of Education, Beijing, China
| | - Xinxing Fu
- Key Laboratory of Otolaryngology Head and Neck Surgery, Beijing Tongren Hospital, Beijing Institute of Otolaryngology, Ministry of Education, Beijing, China
| | - LiLi Yan
- Key Laboratory of Otolaryngology Head and Neck Surgery, Beijing Tongren Hospital, Beijing Institute of Otolaryngology, Ministry of Education, Beijing, China
| | - Chunlin Li
- School of Biomedical Engineering, Capital Medical University, Beijing, China.,Beijing Key Laboratory of Fundamental Research on Biomechanics in Clinical Application, Capital Medical University, Beijing, China
| | - Xu Zhang
- School of Biomedical Engineering, Capital Medical University, Beijing, China.,Beijing Key Laboratory of Fundamental Research on Biomechanics in Clinical Application, Capital Medical University, Beijing, China
| |
Collapse
|
34
|
Kaiser M, Senkowski D, Keil J. Mediofrontal theta-band oscillations reflect top-down influence in the ventriloquist illusion. Hum Brain Mapp 2021; 42:452-466. [PMID: 33617132 PMCID: PMC7775991 DOI: 10.1002/hbm.25236] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2020] [Revised: 09/01/2020] [Accepted: 09/29/2020] [Indexed: 11/12/2022] Open
Abstract
In the ventriloquist illusion, spatially disparate visual signals can influence the perceived location of simultaneous sounds. Previous studies have shown asymmetrical responses in auditory cortical regions following perceived peripheral sound shifts. Moreover, higher-order cortical areas perform inferences on the sources of disparate audiovisual signals. Recent studies have also highlighted top-down influence in the ventriloquist illusion and postulated a governing function of neural oscillations for crossmodal processing. In this EEG study, we analyzed source-reconstructed neural oscillations to address the question of whether perceived sound shifts affect the laterality of auditory responses. Moreover, we investigated the modulation of neural oscillations related to the occurrence of the illusion more generally. With respect to the first question, we did not find evidence for significant changes in the laterality of auditory responses due to perceived sound shifts. However, we found a sustained reduction of mediofrontal theta-band power starting prior to stimulus onset when participants perceived the illusion compared to when they did not perceive the illusion. We suggest that this effect reflects a state of diminished cognitive control, leading to reliance on more readily discriminable visual information and increased crossmodal influence. We conclude that mediofrontal theta-band oscillations serve as a neural mechanism underlying top-down modulation of crossmodal processing in the ventriloquist illusion.
Collapse
Affiliation(s)
- Mathis Kaiser
- Department of Psychiatry and PsychotherapyCharité Universitätsmedizin—BerlinBerlinGermany
- Berlin School of Mind and BrainHumboldt Universität zu BerlinBerlinGermany
| | - Daniel Senkowski
- Department of Psychiatry and PsychotherapyCharité Universitätsmedizin—BerlinBerlinGermany
| | - Julian Keil
- Department of Psychiatry and PsychotherapyCharité Universitätsmedizin—BerlinBerlinGermany
- Biological PsychologyChristian‐Albrechts‐Universität zu KielKielGermany
| |
Collapse
|
35
|
Bidelman GM, Pearson C, Harrison A. Lexical Influences on Categorical Speech Perception Are Driven by a Temporoparietal Circuit. J Cogn Neurosci 2021; 33:840-852. [PMID: 33464162 DOI: 10.1162/jocn_a_01678] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Categorical judgments of otherwise identical phonemes are biased toward hearing words (i.e., "Ganong effect") suggesting lexical context influences perception of even basic speech primitives. Lexical biasing could manifest via late stage postperceptual mechanisms related to decision or, alternatively, top-down linguistic inference that acts on early perceptual coding. Here, we exploited the temporal sensitivity of EEG to resolve the spatiotemporal dynamics of these context-related influences on speech categorization. Listeners rapidly classified sounds from a /gɪ/-/kɪ/ gradient presented in opposing word-nonword contexts (GIFT-kift vs. giss-KISS), designed to bias perception toward lexical items. Phonetic perception shifted toward the direction of words, establishing a robust Ganong effect behaviorally. ERPs revealed a neural analog of lexical biasing emerging within ~200 msec. Source analyses uncovered a distributed neural network supporting the Ganong including middle temporal gyrus, inferior parietal lobe, and middle frontal cortex. Yet, among Ganong-sensitive regions, only left middle temporal gyrus and inferior parietal lobe predicted behavioral susceptibility to lexical influence. Our findings confirm lexical status rapidly constrains sublexical categorical representations for speech within several hundred milliseconds but likely does so outside the purview of canonical auditory-sensory brain areas.
Collapse
Affiliation(s)
- Gavin M Bidelman
- University of Memphis, TN.,University of Tennessee Health Sciences Center, Memphis, TN
| | | | | |
Collapse
|
36
|
Hausfeld L, Shiell M, Formisano E, Riecke L. Cortical processing of distracting speech in noisy auditory scenes depends on perceptual demand. Neuroimage 2020; 228:117670. [PMID: 33359352 DOI: 10.1016/j.neuroimage.2020.117670] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2020] [Revised: 12/13/2020] [Accepted: 12/14/2020] [Indexed: 11/15/2022] Open
Abstract
Selective attention is essential for the processing of multi-speaker auditory scenes because they require the perceptual segregation of the relevant speech ("target") from irrelevant speech ("distractors"). For simple sounds, it has been suggested that the processing of multiple distractor sounds depends on bottom-up factors affecting task performance. However, it remains unclear whether such dependency applies to naturalistic multi-speaker auditory scenes. In this study, we tested the hypothesis that increased perceptual demand (the processing requirement posed by the scene to separate the target speech) reduces the cortical processing of distractor speech thus decreasing their perceptual segregation. Human participants were presented with auditory scenes including three speakers and asked to selectively attend to one speaker while their EEG was acquired. The perceptual demand of this selective listening task was varied by introducing an auditory cue (interaural time differences, ITDs) for segregating the target from the distractor speakers, while acoustic differences between the distractors were matched in ITD and loudness. We obtained a quantitative measure of the cortical segregation of distractor speakers by assessing the difference in how accurately speech-envelope following EEG responses could be predicted by models of averaged distractor speech versus models of individual distractor speech. In agreement with our hypothesis, results show that interaural segregation cues led to improved behavioral word-recognition performance and stronger cortical segregation of the distractor speakers. The neural effect was strongest in the δ-band and at early delays (0 - 200 ms). Our results indicate that during low perceptual demand, the human cortex represents individual distractor speech signals as more segregated. This suggests that, in addition to purely acoustical properties, the cortical processing of distractor speakers depends on factors like perceptual demand.
Collapse
Affiliation(s)
- Lars Hausfeld
- Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, Maastricht University, P.O. Box 616, 6200MD Maastricht, The Netherlands; Maastricht Brain Imaging Centre, 6200MD Maastricht, The Netherlands.
| | - Martha Shiell
- Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, Maastricht University, P.O. Box 616, 6200MD Maastricht, The Netherlands; Maastricht Brain Imaging Centre, 6200MD Maastricht, The Netherlands
| | - Elia Formisano
- Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, Maastricht University, P.O. Box 616, 6200MD Maastricht, The Netherlands; Maastricht Brain Imaging Centre, 6200MD Maastricht, The Netherlands; Maastricht Centre for Systems Biology, 6200MD Maastricht, The Netherlands
| | - Lars Riecke
- Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, Maastricht University, P.O. Box 616, 6200MD Maastricht, The Netherlands; Maastricht Brain Imaging Centre, 6200MD Maastricht, The Netherlands
| |
Collapse
|
37
|
Ching ASM, Kim J, Davis C. Time course of the unmasked attentional blink. Psychophysiology 2020; 58:e13686. [PMID: 33141450 DOI: 10.1111/psyp.13686] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2020] [Revised: 07/27/2020] [Accepted: 08/17/2020] [Indexed: 11/30/2022]
Abstract
The Attentional Blink (AB) usually refers to the impaired report of a second target (T2) if it appears within 200-500 ms after a first target within a rapid sequence of distractors. The present study focused on a less studied AB variant known as the unmasked AB, where T2 is the last item of the sequence and T2 report is unaffected. This aspect of the unmasked AB holds promise for an experimental paradigm in which measures of on-going event-related processing are unconfounded by differences in late-stage processing. To fully characterize the unmasked AB paradigm, we used a randomization statistics approach to comprehensively examine the electroencephalographic signature of the unmasked AB. We examined the unmasked AB with auditory and visual T2s-participants attended to either the auditory or visual information within a sequence of paired auditory-visual stimuli, and reported targets within the attended modality stream while ignoring the other. As predicted, T2 report was unaffected by the unmasked AB. The visual AB was associated with delayed but intact N2 and P3 components, and a suppressed N1. We suggest that this N1 is linked to auditory processing of the distractor stream, and reflects the cognitive system prioritizing the processing of visual targets over auditory distractors in response to AB-related processing load. The auditory AB only indicated a delayed but intact P3. Collectively, these findings support the view that the AB limits the entry of information into consciousness via a late-stage modal bottleneck, and suggest an ongoing compensatory response at early latencies.
Collapse
Affiliation(s)
| | - Jeesun Kim
- The MARCS Institute, University of Western Sydney, Sydney, Australia
| | - Chris Davis
- The MARCS Institute, University of Western Sydney, Sydney, Australia
| |
Collapse
|
38
|
Talkington WJ, Donai J, Kadner AS, Layne ML, Forino A, Wen S, Gao S, Gray MM, Ashraf AJ, Valencia GN, Smith BD, Khoo SK, Gray SJ, Lass N, Brefczynski-Lewis JA, Engdahl S, Graham D, Frum CA, Lewis JW. Electrophysiological Evidence of Early Cortical Sensitivity to Human Conspecific Mimic Voice as a Distinct Category of Natural Sound. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2020; 63:3539-3559. [PMID: 32936717 PMCID: PMC8060013 DOI: 10.1044/2020_jslhr-20-00063] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/17/2020] [Revised: 04/29/2020] [Accepted: 07/01/2020] [Indexed: 06/11/2023]
Abstract
Purpose From an anthropological perspective of hominin communication, the human auditory system likely evolved to enable special sensitivity to sounds produced by the vocal tracts of human conspecifics whether attended or passively heard. While numerous electrophysiological studies have used stereotypical human-produced verbal (speech voice and singing voice) and nonverbal vocalizations to identify human voice-sensitive responses, controversy remains as to when (and where) processing of acoustic signal attributes characteristic of "human voiceness" per se initiate in the brain. Method To explore this, we used animal vocalizations and human-mimicked versions of those calls ("mimic voice") to examine late auditory evoked potential responses in humans. Results Here, we revealed an N1b component (96-120 ms poststimulus) during a nonattending listening condition showing significantly greater magnitude in response to mimics, beginning as early as primary auditory cortices, preceding the time window reported in previous studies that revealed species-specific vocalization processing initiating in the range of 147-219 ms. During a sound discrimination task, a P600 (500-700 ms poststimulus) component showed specificity for accurate discrimination of human mimic voice. Distinct acoustic signal attributes and features of the stimuli were used in a classifier model, which could distinguish most human from animal voice comparably to behavioral data-though none of these single features could adequately distinguish human voiceness. Conclusions These results provide novel ideas for algorithms used in neuromimetic hearing aids, as well as direct electrophysiological support for a neurocognitive model of natural sound processing that informs both neurodevelopmental and anthropological models regarding the establishment of auditory communication systems in humans. Supplemental Material https://doi.org/10.23641/asha.12903839.
Collapse
Affiliation(s)
- William J. Talkington
- Department of Neuroscience, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| | - Jeremy Donai
- Department of Communication Sciences and Disorders, College of Education and Human Services, West Virginia University, Morgantown
| | - Alexandra S. Kadner
- Department of Neuroscience, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| | - Molly L. Layne
- Department of Neuroscience, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| | - Andrew Forino
- Department of Neuroscience, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| | - Sijin Wen
- Department of Biostatistics, West Virginia University, Morgantown
| | - Si Gao
- Department of Biostatistics, West Virginia University, Morgantown
| | - Margeaux M. Gray
- Department of Biology, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| | - Alexandria J. Ashraf
- Department of Biology, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| | - Gabriela N. Valencia
- Department of Neuroscience, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| | - Brandon D. Smith
- Department of Biology, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| | - Stephanie K. Khoo
- Department of Biology, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| | - Stephen J. Gray
- Department of Neuroscience, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| | - Norman Lass
- Department of Communication Sciences and Disorders, College of Education and Human Services, West Virginia University, Morgantown
| | | | - Susannah Engdahl
- Department of Neuroscience, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| | - David Graham
- Department of Computer Science and Electrical Engineering, West Virginia University, Morgantown
| | - Chris A. Frum
- Department of Neuroscience, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| | - James W. Lewis
- Department of Neuroscience, Rockefeller Neuroscience Institute, West Virginia University, Morgantown
| |
Collapse
|
39
|
Mahmud MS, Ahmed F, Al-Fahad R, Moinuddin KA, Yeasin M, Alain C, Bidelman GM. Decoding Hearing-Related Changes in Older Adults' Spatiotemporal Neural Processing of Speech Using Machine Learning. Front Neurosci 2020; 14:748. [PMID: 32765215 PMCID: PMC7378401 DOI: 10.3389/fnins.2020.00748] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2019] [Accepted: 06/25/2020] [Indexed: 12/25/2022] Open
Abstract
Speech perception in noisy environments depends on complex interactions between sensory and cognitive systems. In older adults, such interactions may be affected, especially in those individuals who have more severe age-related hearing loss. Using a data-driven approach, we assessed the temporal (when in time) and spatial (where in the brain) characteristics of cortical speech-evoked responses that distinguish older adults with or without mild hearing loss. We performed source analyses to estimate cortical surface signals from the EEG recordings during a phoneme discrimination task conducted under clear and noise-degraded conditions. We computed source-level ERPs (i.e., mean activation within each ROI) from each of the 68 ROIs of the Desikan-Killiany (DK) atlas, averaged over a randomly chosen 100 trials without replacement to form feature vectors. We adopted a multivariate feature selection method called stability selection and control to choose features that are consistent over a range of model parameters. We use parameter optimized support vector machine (SVM) as a classifiers to investigate the time course and brain regions that segregate groups and speech clarity. For clear speech perception, whole-brain data revealed a classification accuracy of 81.50% [area under the curve (AUC) 80.73%; F1-score 82.00%], distinguishing groups within ∼60 ms after speech onset (i.e., as early as the P1 wave). We observed lower accuracy of 78.12% [AUC 77.64%; F1-score 78.00%] and delayed classification performance when speech was embedded in noise, with group segregation at 80 ms. Separate analysis using left (LH) and right hemisphere (RH) regions showed that LH speech activity was better at distinguishing hearing groups than activity measured in the RH. Moreover, stability selection analysis identified 12 brain regions (among 1428 total spatiotemporal features from 68 regions) where source activity segregated groups with >80% accuracy (clear speech); whereas 16 regions were critical for noise-degraded speech to achieve a comparable level of group segregation (78.7% accuracy). Our results identify critical time-courses and brain regions that distinguish mild hearing loss from normal hearing in older adults and confirm a larger number of active areas, particularly in RH, when processing noise-degraded speech information.
Collapse
Affiliation(s)
- Md Sultan Mahmud
- Department of Electrical and Computer Engineering, The University of Memphis, Memphis, TN, United States
| | - Faruk Ahmed
- Department of Electrical and Computer Engineering, The University of Memphis, Memphis, TN, United States
| | - Rakib Al-Fahad
- Department of Electrical and Computer Engineering, The University of Memphis, Memphis, TN, United States
| | - Kazi Ashraf Moinuddin
- Department of Electrical and Computer Engineering, The University of Memphis, Memphis, TN, United States
| | - Mohammed Yeasin
- Department of Electrical and Computer Engineering, The University of Memphis, Memphis, TN, United States
| | - Claude Alain
- Rotman Research Institute-Baycrest Centre for Geriatric Care, Toronto, ON, Canada.,Department of Psychology, University of Toronto, Toronto, ON, Canada.,Institute of Medical Sciences, University of Toronto, Toronto, ON, Canada
| | - Gavin M Bidelman
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, United States.,School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, United States.,Department of Anatomy and Neurobiology, University of Tennessee Health Science Center, Memphis, TN, United States
| |
Collapse
|
40
|
Jones KT, Johnson EL, Tauxe ZS, Rojas DC. Modulation of auditory gamma-band responses using transcranial electrical stimulation. J Neurophysiol 2020; 123:2504-2514. [PMID: 32459551 DOI: 10.1152/jn.00003.2020] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/06/2023] Open
Abstract
Auditory gamma-band (>30 Hz) activity is a biomarker of cortical excitation/inhibition (E/I) balance in autism, schizophrenia, and bipolar disorder. We provide a comprehensive account of the effects of transcranial alternating current stimulation (tACS) and transcranial direct current stimulation (tDCS) on gamma responses. Forty-five healthy young adults listened to 40-Hz auditory click trains while electroencephalography (EEG) data were collected to measure stimulus-related gamma activity immediately before and after 10 min of 1 mA tACS (40 Hz), tDCS, or sham stimulation to left auditory cortex. tACS, but not tDCS, increased gamma power and phase locking to the auditory stimulus. However, both tACS and tDCS strengthened the gamma phase connectome, and effects persisted beyond the stimulus. Finally, tDCS strengthened the coupling of gamma activity to alpha oscillations after termination of the stimulus. No effects were observed in prestimulus gamma power, the gamma amplitude connectome, or any band-limited alpha measure. Whereas both stimulation techniques synchronize gamma responses between regions, tACS also tunes the magnitude and timing of gamma responses to the stimulus. Results reveal dissociable neurophysiological changes following tACS and tDCS and demonstrate that clinical biomarkers can be altered with noninvasive neurostimulation, especially frequency-tuned tACS.NEW & NOTEWORTHY Gamma frequency-tuned transcranial alternating current stimulation (tACS) adjusts the magnitude and timing of auditory gamma responses, as compared with both sham stimulation and transcranial direct current stimulation (tDCS). However, both tACS and tDCS strengthen the gamma phase connectome, which is disrupted in numerous neurological and psychiatric disorders. These findings reveal dissociable neurophysiological changes following two noninvasive neurostimulation techniques commonly applied in clinical and research settings.
Collapse
Affiliation(s)
- Kevin T Jones
- Colorado State University, Department of Psychology, Fort Collins, Colorado.,University of California-San Francisco, Department of Neurology, Neuroscape, San Francisco, California
| | - Elizabeth L Johnson
- University of California-Berkeley, Helen Wills Neuroscience Institute, Berkeley, California.,Wayne State University, Institute of Gerontology, Life-Span Cognitive Neuroscience Program, Detroit, Michigan
| | - Zoe S Tauxe
- Colorado State University, Department of Psychology, Fort Collins, Colorado.,University of California-San Diego, Department of Psychology, San Diego, California
| | - Donald C Rojas
- Colorado State University, Department of Psychology, Fort Collins, Colorado
| |
Collapse
|
41
|
Zimmermann J, Ross B, Moscovitch M, Alain C. Neural dynamics supporting auditory long-term memory effects on target detection. Neuroimage 2020; 218:116979. [PMID: 32447014 DOI: 10.1016/j.neuroimage.2020.116979] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2019] [Revised: 05/15/2020] [Accepted: 05/18/2020] [Indexed: 12/31/2022] Open
Abstract
Auditory long-term memory has been shown to facilitate signal detection. However, the nature and timing of the cognitive processes supporting such benefits remain equivocal. We measured neuroelectric brain activity while young adults were presented with a contextual memory cue designed to assist with the detection of a faint pure tone target embedded in an audio clip of an everyday environmental scene (e.g., the soundtrack of a restaurant). During an initial familiarization task, participants heard such audio clips, half of which included a target sound (memory cue trials) at a specific time and location (left or right ear), as well as audio clips without a target (neutral trials). Following a 1-h or 24-h retention interval, the same audio clips were presented, but now all included a target. Participants were asked to press a button as soon as they heard the pure tone target. Overall, participants were faster and more accurate during memory than neutral cue trials. The auditory contextual memory effects on performance coincided with three temporally and spatially distinct neural modulations, which encompassed changes in the amplitude of event-related potential as well as changes in theta, alpha, beta and gamma power. Brain electrical source analyses revealed greater source activity in memory than neutral cue trials in the right superior temporal gyrus and left parietal cortex. Conversely, neutral trials were associated with greater source activity than memory cue trials in the left posterior medial temporal lobe. Target detection was associated with increased negativity (N2), and a late positive (P3b) wave at frontal and parietal sites, respectively. The effect of auditory contextual memory on brain activity preceding target onset showed little lateralization. Together, these results are consistent with contextual memory facilitating retrieval of target-context associations and deployment and management of auditory attentional resources to when the target occurred. The results also suggest that the auditory cortices, parietal cortex, and medial temporal lobe may be parts of a neural network enabling memory-guided attention during auditory scene analysis.
Collapse
Affiliation(s)
- Jacqueline Zimmermann
- Rotman Research Institute, Psychology, University of Toronto, Ontario, Canada; Department of Psychology, University of Toronto, Ontario, Canada
| | - Bernhard Ross
- Rotman Research Institute, Psychology, University of Toronto, Ontario, Canada; Department of Medical Biophysics, University of Toronto, Ontario, Canada; Institute of Medical Sciences, University of Toronto, Ontario, Canada
| | - Morris Moscovitch
- Rotman Research Institute, Psychology, University of Toronto, Ontario, Canada; Department of Psychology, University of Toronto, Ontario, Canada
| | - Claude Alain
- Rotman Research Institute, Psychology, University of Toronto, Ontario, Canada; Department of Psychology, University of Toronto, Ontario, Canada; Institute of Medical Sciences, University of Toronto, Ontario, Canada; Faculty of Music, University of Toronto, Ontario, Canada.
| |
Collapse
|
42
|
Foldal MD, Blenkmann AO, Llorens A, Knight RT, Solbakk AK, Endestad T. The brain tracks auditory rhythm predictability independent of selective attention. Sci Rep 2020; 10:7975. [PMID: 32409738 PMCID: PMC7224206 DOI: 10.1038/s41598-020-64758-y] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2019] [Accepted: 04/07/2020] [Indexed: 11/16/2022] Open
Abstract
The brain responds to violations of expected rhythms, due to extraction- and prediction of the temporal structure in auditory input. Yet, it is unknown how probability of rhythm violations affects the overall rhythm predictability. Another unresolved question is whether predictive processes are independent of attention processes. In this study, EEG was recorded while subjects listened to rhythmic sequences. Predictability was manipulated by changing the stimulus-onset-asynchrony (SOA deviants) for given tones in the rhythm. When SOA deviants were inserted rarely, predictability remained high, whereas predictability was lower with more frequent SOA deviants. Dichotic tone-presentation allowed for independent manipulation of attention, as specific tones of the rhythm were presented to separate ears. Attention was manipulated by instructing subjects to attend to tones in one ear only, while keeping the rhythmic structure of tones constant. The analyses of event-related potentials revealed an attenuated N1 for tones when rhythm predictability was high, while the N1 was enhanced by attention to tones. Bayesian statistics revealed no interaction between predictability and attention. A right-lateralization of attention effects, but not predictability effects, suggested potentially different cortical processes. This is the first study to show that probability of rhythm violation influences rhythm predictability, independent of attention.
Collapse
Affiliation(s)
- Maja D Foldal
- Department of Psychology, University of Oslo, Oslo, Norway. .,RITMO Centre for Interdisciplinary Studies in Rhythm, Time and Motion, University of Oslo, Oslo, Norway.
| | - Alejandro O Blenkmann
- Department of Psychology, University of Oslo, Oslo, Norway.,RITMO Centre for Interdisciplinary Studies in Rhythm, Time and Motion, University of Oslo, Oslo, Norway
| | - Anaïs Llorens
- Department of Psychology, University of Oslo, Oslo, Norway.,Department of Neurosurgery, Oslo University Hospital, Oslo, Norway.,Department of Psychology and the Helen Wills Neuroscience Institute, University of California Berkeley, Berkeley, USA
| | - Robert T Knight
- Department of Psychology and the Helen Wills Neuroscience Institute, University of California Berkeley, Berkeley, USA
| | - Anne-Kristin Solbakk
- Department of Psychology, University of Oslo, Oslo, Norway.,RITMO Centre for Interdisciplinary Studies in Rhythm, Time and Motion, University of Oslo, Oslo, Norway.,Department of Neurosurgery, Oslo University Hospital, Oslo, Norway.,Department of Neuropsychology, Helgeland Hospital, Mosjøen, Norway
| | - Tor Endestad
- Department of Psychology, University of Oslo, Oslo, Norway.,RITMO Centre for Interdisciplinary Studies in Rhythm, Time and Motion, University of Oslo, Oslo, Norway.,Department of Neuropsychology, Helgeland Hospital, Mosjøen, Norway
| |
Collapse
|
43
|
Spatiotemporal dynamics of auditory information processing in the insular cortex: an intracranial EEG study using an oddball paradigm. Brain Struct Funct 2020; 225:1537-1559. [DOI: 10.1007/s00429-020-02072-z] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2019] [Accepted: 04/13/2020] [Indexed: 12/27/2022]
|
44
|
The First 250 ms of Auditory Processing: No Evidence of Early Processing Negativity in the Go/NoGo Task. Sci Rep 2020; 10:4041. [PMID: 32132630 PMCID: PMC7055275 DOI: 10.1038/s41598-020-61060-9] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2019] [Accepted: 01/30/2020] [Indexed: 12/02/2022] Open
Abstract
Past evidence of an early Processing Negativity in auditory Go/NoGo event-related potential (ERP) data suggests that young adults proactively process sensory information in two-choice tasks. This study aimed to clarify the occurrence of Go/NoGo Processing Negativity and investigate the ERP component series related to the first 250 ms of auditory processing in two Go/NoGo tasks differing in target probability. ERP data related to each task were acquired from 60 healthy young adults (M = 20.4, SD = 3.1 years). Temporal principal components analyses were used to decompose ERP data in each task. Statistical analyses compared component amplitudes between stimulus type (Go vs. NoGo) and probability (High vs. Low). Neuronal source localisation was also conducted for each component. Processing Negativity was not evident; however, P1, N1a, N1b, and N1c were identified in each task, with Go P2 and NoGo N2b. The absence of Processing Negativity in this study indicated that young adults do not proactively process targets to complete the Go/NoGo task and/or questioned Processing Negativity’s conceptualisation. Additional analyses revealed stimulus-specific processing as early as P1, and outlined a complex network of active neuronal sources underlying each component, providing useful insight into Go and NoGo information processing in young adults.
Collapse
|
45
|
Bidelman GM, Bush LC, Boudreaux AM. Effects of Noise on the Behavioral and Neural Categorization of Speech. Front Neurosci 2020; 14:153. [PMID: 32180700 PMCID: PMC7057933 DOI: 10.3389/fnins.2020.00153] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2019] [Accepted: 02/10/2020] [Indexed: 02/02/2023] Open
Abstract
We investigated whether the categorical perception (CP) of speech might also provide a mechanism that aids its perception in noise. We varied signal-to-noise ratio (SNR) [clear, 0 dB, -5 dB] while listeners classified an acoustic-phonetic continuum (/u/ to /a/). Noise-related changes in behavioral categorization were only observed at the lowest SNR. Event-related brain potentials (ERPs) differentiated category vs. category-ambiguous speech by the P2 wave (~180-320 ms). Paralleling behavior, neural responses to speech with clear phonetic status (i.e., continuum endpoints) were robust to noise down to -5 dB SNR, whereas responses to ambiguous tokens declined with decreasing SNR. Results demonstrate that phonetic speech representations are more resistant to degradation than corresponding acoustic representations. Findings suggest the mere process of binning speech sounds into categories provides a robust mechanism to aid figure-ground speech perception by fortifying abstract categories from the acoustic signal and making the speech code more resistant to external interferences.
Collapse
Affiliation(s)
- Gavin M Bidelman
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, United States.,School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, United States.,Department of Anatomy and Neurobiology, University of Tennessee Health Sciences Center, Memphis, TN, United States
| | - Lauren C Bush
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, United States
| | - Alex M Boudreaux
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, United States
| |
Collapse
|
46
|
Silva DMR, Rothe-Neves R, Melges DB. Long-latency event-related responses to vowels: N1-P2 decomposition by two-step principal component analysis. Int J Psychophysiol 2019; 148:93-102. [PMID: 31863852 DOI: 10.1016/j.ijpsycho.2019.11.010] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/15/2019] [Revised: 11/16/2019] [Accepted: 11/18/2019] [Indexed: 11/26/2022]
Abstract
The N1-P2 complex of the auditory event-related potential (ERP) has been used to examine neural activity associated with speech sound perception. Since it is thought to reflect multiple generator processes, its functional significance is difficult to infer. In the present study, a temporospatial principal component analysis (PCA) was used to decompose the N1-P2 response into latent factors underlying covariance patterns in ERP data recorded during passive listening to pairs of successive vowels. In each trial, one of six sounds drawn from an /i/-/e/ vowel continuum was followed either by an identical sound, a different token of the same vowel category, or a token from the other category. Responses were examined as to how they were modulated by within- and across-category vowel differences and by adaptation (repetition suppression) effects. Five PCA factors were identified as corresponding to three well-known N1 subcomponents and two P2 subcomponents. Results added evidence that the N1 peak reflects both generators that are sensitive to spectral information and generators that are not. For later latency ranges, different patterns of sensitivity to vowel quality were found, including category-related effects. Particularly, a subcomponent identified as the Tb wave showed release from adaptation in response to an /i/ followed by an /e/ sound. A P2 subcomponent varied linearly with spectral shape along the vowel continuum, while the other was stronger the closer the vowel was to the category boundary, suggesting separate processing of continuous and category-related information. Thus, the PCA-based decomposition of the N1-P2 complex was functionally meaningful, revealing distinct underlying processes at work during speech sound perception.
Collapse
Affiliation(s)
- Daniel M R Silva
- Phonetics Lab, Faculty of Letters, Federal University of Minas Gerais, Belo Horizonte, Brazil
| | - Rui Rothe-Neves
- Phonetics Lab, Faculty of Letters, Federal University of Minas Gerais, Belo Horizonte, Brazil.
| | - Danilo B Melges
- Graduate Program in Electrical Engineering, Department of Electrical Engineering, Federal University of Minas Gerais
| |
Collapse
|
47
|
Auditory-frontal Channeling in α and β Bands is Altered by Age-related Hearing Loss and Relates to Speech Perception in Noise. Neuroscience 2019; 423:18-28. [PMID: 31705894 DOI: 10.1016/j.neuroscience.2019.10.044] [Citation(s) in RCA: 24] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2019] [Revised: 09/19/2019] [Accepted: 10/27/2019] [Indexed: 01/16/2023]
Abstract
Difficulty understanding speech-in-noise (SIN) is a pervasive problem faced by older adults particularly those with hearing loss. Previous studies have identified structural and functional changes in the brain that contribute to older adults' speech perception difficulties. Yet, many of these studies use neuroimaging techniques that evaluate only gross activation in isolated brain regions. Neural oscillations may provide further insight into the processes underlying SIN perception as well as the interaction between auditory cortex and prefrontal linguistic brain regions that mediate complex behaviors. We examined frequency-specific neural oscillations and functional connectivity of the EEG in older adults with and without hearing loss during an active SIN perception task. Brain-behavior correlations revealed listeners who were more resistant to the detrimental effects of noise also demonstrated greater modulation of α phase coherence between clean and noise-degraded speech, suggesting α desynchronization reflects release from inhibition and more flexible allocation of neural resources. Additionally, we found top-down β connectivity between prefrontal and auditory cortices strengthened with poorer hearing thresholds despite minimal behavioral differences. This is consistent with the proposal that linguistic brain areas may be recruited to compensate for impoverished auditory inputs through increased top-down predictions to assist SIN perception. Overall, these results emphasize the importance of top-down signaling in low-frequency brain rhythms that help compensate for hearing-related declines and facilitate efficient SIN processing.
Collapse
|
48
|
Bidelman GM, Mahmud MS, Yeasin M, Shen D, Arnott SR, Alain C. Age-related hearing loss increases full-brain connectivity while reversing directed signaling within the dorsal-ventral pathway for speech. Brain Struct Funct 2019; 224:2661-2676. [PMID: 31346715 PMCID: PMC6778722 DOI: 10.1007/s00429-019-01922-9] [Citation(s) in RCA: 26] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2019] [Accepted: 07/13/2019] [Indexed: 01/08/2023]
Abstract
Speech comprehension difficulties are ubiquitous to aging and hearing loss, particularly in noisy environments. Older adults' poorer speech-in-noise (SIN) comprehension has been related to abnormal neural representations within various nodes (regions) of the speech network, but how senescent changes in hearing alter the transmission of brain signals remains unspecified. We measured electroencephalograms in older adults with and without mild hearing loss during a SIN identification task. Using functional connectivity and graph-theoretic analyses, we show that hearing-impaired (HI) listeners have more extended (less integrated) communication pathways and less efficient information exchange among widespread brain regions (larger network eccentricity) than their normal-hearing (NH) peers. Parameter optimized support vector machine classifiers applied to EEG connectivity data showed hearing status could be decoded (> 85% accuracy) solely using network-level descriptions of brain activity, but classification was particularly robust using left hemisphere connections. Notably, we found a reversal in directed neural signaling in left hemisphere dependent on hearing status among specific connections within the dorsal-ventral speech pathways. NH listeners showed an overall net "bottom-up" signaling directed from auditory cortex (A1) to inferior frontal gyrus (IFG; Broca's area), whereas the HI group showed the reverse signal (i.e., "top-down" Broca's → A1). A similar flow reversal was noted between left IFG and motor cortex. Our full-brain connectivity results demonstrate that even mild forms of hearing loss alter how the brain routes information within the auditory-linguistic-motor loop.
Collapse
Affiliation(s)
- Gavin M Bidelman
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA.
- School of Communication Sciences and Disorders, University of Memphis, 4055 North Park Loop, Memphis, TN, 38152, USA.
- Department of Anatomy and Neurobiology, University of Tennessee Health Sciences Center, Memphis, TN, USA.
| | - Md Sultan Mahmud
- Department of Electrical and Computer Engineering, University of Memphis, Memphis, TN, USA
| | - Mohammed Yeasin
- Department of Electrical and Computer Engineering, University of Memphis, Memphis, TN, USA
| | - Dawei Shen
- Rotman Research Institute-Baycrest Centre for Geriatric Care, Toronto, ON, Canada
| | - Stephen R Arnott
- Rotman Research Institute-Baycrest Centre for Geriatric Care, Toronto, ON, Canada
| | - Claude Alain
- Rotman Research Institute-Baycrest Centre for Geriatric Care, Toronto, ON, Canada
- Department of Psychology, University of Toronto, Toronto, ON, Canada
- Institute of Medical Sciences, University of Toronto, Toronto, ON, Canada
| |
Collapse
|
49
|
Lee JY, Kang BC, Park JW, Park HJ. Changes in Cortical Auditory Evoked Potentials by Ipsilateral, Contralateral and Binaural Speech Stimulation in Normal-Hearing Adults. Clin Exp Otorhinolaryngol 2019; 13:133-140. [PMID: 31640335 PMCID: PMC7248601 DOI: 10.21053/ceo.2019.00801] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2019] [Accepted: 09/02/2019] [Indexed: 11/22/2022] Open
Abstract
Objectives Cortical auditory evoked potentials (CAEPs) have been used to examine auditory cortical development or changes in patients with hearing loss. However, there have been no studies analyzing CAEP responses to the different sound stimulation by different stimulation sides. We characterized changes in normal CAEP responses by stimulation sides in normal-hearing adults. Methods CAEPs from the right auditory cortex were recorded in 16 adults following unilateral (ipsilateral and contralateral) and bilateral sound stimulation using three speech sounds (/m/, /g/, and /t/). Amplitudes and latencies of the CAEP peaks in three conditions were compared. Results Contralateral stimulation elicited larger P2-N1 amplitudes (sum of P2 and N1 amplitudes) than ipsilateral stimulation regardless of the stimulation sounds, mostly due to the larger P2 amplitudes obtained, but elicited comparable P2-N1 amplitudes to bilateral stimulation. Although the P2-N1 amplitudes obtained with the three speech sounds were comparable following contralateral stimulation, the /m/ sound elicited the largest P2-N1 amplitude in ipsilateral stimulation condition due to the largest N1 amplitude obtained, whereas /t/ elicited larger a P2-N1 amplitude than /g/ in bilateral stimulation condition due to a larger P2 amplitude. Conclusion Spectrally different speech sounds and input sides are encoded differently at the cortical level in normal-hearing adults. Standardized speech stimuli, as well as specific input sides of speech, are needed to examine normal development or rehabilitation-related changes of the auditory cortex in the future.
Collapse
Affiliation(s)
- Jee Yeon Lee
- Department of Otorhinolaryngology-Head and Neck Surgery, Asan Medical Center, University of Ulsan College of Medicine, Seoul, Korea
| | - Byung Chul Kang
- Department of Otorhinolaryngology-Head and Neck Surgery, Ulsan University Hospital, University of Ulsan College of Medicine, Ulsan, Korea
| | - Jun Woo Park
- Department of Otorhinolaryngology-Head and Neck Surgery, Asan Medical Center, University of Ulsan College of Medicine, Seoul, Korea
| | - Hong Ju Park
- Department of Otorhinolaryngology-Head and Neck Surgery, Asan Medical Center, University of Ulsan College of Medicine, Seoul, Korea
| |
Collapse
|
50
|
Bidelman GM, Myers MH. Frontal cortex selectively overrides auditory processing to bias perception for looming sonic motion. Brain Res 2019; 1726:146507. [PMID: 31606413 DOI: 10.1016/j.brainres.2019.146507] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2019] [Revised: 10/02/2019] [Accepted: 10/09/2019] [Indexed: 12/18/2022]
Abstract
Rising intensity sounds signal approaching objects traveling toward an observer. A variety of species preferentially respond to looming over receding auditory motion, reflecting an evolutionary perceptual bias for recognizing approaching threats. We probed the neural origins of this stark perceptual anisotropy to reveal how the brain creates privilege for auditory looming events. While recording neural activity via electroencephalography (EEG), human listeners rapidly judged whether dynamic (intensity varying) tones were looming or receding in percept. Behaviorally, listeners responded faster to auditory looms confirming a perceptual bias for approaching signals. EEG source analysis revealed sensory activation localized to primary auditory cortex (PAC) and decision-related activity in prefrontal cortex (PFC) within 200 ms after sound onset followed by additional expansive PFC activation by 500 ms. Notably, early PFC (but not PAC) activity rapidly differentiated looming and receding stimuli and this effect roughly co-occurred with sound arrival in auditory cortex. Brain-behavior correlations revealed an association between PFC neural latencies and listeners' speed of sonic motion judgments. Directed functional connectivity revealed stronger information flow from PFC → PAC during looming vs. receding sounds. Our electrophysiological data reveal a critical, previously undocumented role of prefrontal cortex in judging dynamic sonic motion. Both faster neural bias and a functional override of obligatory sensory processing via selective, directional PFC signaling toward auditory system establish the perceptual privilege for approaching looming sounds.
Collapse
Affiliation(s)
- Gavin M Bidelman
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA; School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; University of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, USA.
| | - Mark H Myers
- University of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, USA
| |
Collapse
|