1
|
Polonenko MJ, Maddox RK. The Effect of Speech Masking on the Human Subcortical Response to Continuous Speech. eNeuro 2025; 12:ENEURO.0561-24.2025. [PMID: 40127932 PMCID: PMC11974362 DOI: 10.1523/eneuro.0561-24.2025] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2024] [Revised: 02/04/2025] [Accepted: 02/10/2025] [Indexed: 03/26/2025] Open
Abstract
Auditory masking-the interference of the encoding and processing of an acoustic stimulus imposed by one or more competing stimuli-is nearly omnipresent in daily life and presents a critical barrier to many listeners, including people with hearing loss, users of hearing aids and cochlear implants, and people with auditory processing disorders. The perceptual aspects of masking have been actively studied for several decades, and particular emphasis has been placed on masking of speech by other speech sounds. The neural effects of such masking, especially at the subcortical level, have been much less studied, in large part due to the technical limitations of making such measurements. Recent work has allowed estimation of the auditory brainstem response (ABR), whose characteristic waves are linked to specific subcortical areas, to naturalistic speech. In this study, we used those techniques to measure the encoding of speech stimuli that were masked by one or more simultaneous other speech stimuli. We presented listeners with simultaneous speech from one, two, three, or five simultaneous talkers, corresponding to a range of signal-to-noise ratios (clean, 0, -3, and -6 dB), and derived the ABR to each talker in the mixture. Each talker in a mixture was treated in turn as a target sound masked by other talkers, making the response quicker to acquire. We found consistently across listeners that ABR Wave V amplitudes decreased and latencies increased as the number of competing talkers increased.
Collapse
Affiliation(s)
- Melissa J Polonenko
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, Minnesota, 55455
- Departments of Biomedical Engineering and Neuroscience, University of Rochester, Rochester, New York, 14627
| | - Ross K Maddox
- Departments of Biomedical Engineering and Neuroscience, University of Rochester, Rochester, New York, 14627
- Kresge Hearing Research Institute, Department of Otolaryngology - Head and Neck Surgery, University of Michigan, Ann Arbor, Michigan, 48109
| |
Collapse
|
2
|
Reisinger P, Gillis M, Suess N, Vanthornhout J, Haider CL, Hartmann T, Hauswald A, Schwarz K, Francart T, Weisz N. Neural Speech Tracking Contribution of Lip Movements Predicts Behavioral Deterioration When the Speaker's Mouth Is Occluded. eNeuro 2025; 12:ENEURO.0368-24.2024. [PMID: 39819839 PMCID: PMC11801124 DOI: 10.1523/eneuro.0368-24.2024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2024] [Accepted: 12/09/2024] [Indexed: 01/19/2025] Open
Abstract
Observing lip movements of a speaker facilitates speech understanding, especially in challenging listening situations. Converging evidence from neuroscientific studies shows stronger neural responses to audiovisual stimuli compared with audio-only stimuli. However, the interindividual variability of this contribution of lip movement information and its consequences on behavior are unknown. We analyzed source-localized magnetoencephalographic responses from 29 normal-hearing participants (12 females) listening to audiovisual speech, both with and without the speaker wearing a surgical face mask, and in the presence or absence of a distractor speaker. Using temporal response functions to quantify neural speech tracking, we show that neural responses to lip movements are, in general, enhanced when speech is challenging. After controlling for speech acoustics, we show that lip movements contribute to enhanced neural speech tracking, particularly when a distractor speaker is present. However, the extent of this visual contribution to neural speech tracking varied greatly among participants. Probing the behavioral relevance, we demonstrate that individuals who show a higher contribution of lip movements in terms of neural speech tracking show a stronger drop in comprehension and an increase in perceived difficulty when the mouth is occluded by a surgical face mask. In contrast, no effect was found when the mouth was not occluded. We provide novel insights on how the contribution of lip movements in terms of neural speech tracking varies among individuals and its behavioral relevance, revealing negative consequences when visual speech is absent. Our results also offer potential implications for objective assessments of audiovisual speech perception.
Collapse
Affiliation(s)
- Patrick Reisinger
- Department of Psychology, Centre for Cognitive Neuroscience, Paris-Lodron-University of Salzburg, Salzburg 5020, Austria
| | - Marlies Gillis
- Experimental Oto-Rhino-Laryngology, Department of Neurosciences, Leuven Brain Institute, KU Leuven, Leuven 3000, Belgium
| | - Nina Suess
- Department of Psychology, Centre for Cognitive Neuroscience, Paris-Lodron-University of Salzburg, Salzburg 5020, Austria
| | - Jonas Vanthornhout
- Experimental Oto-Rhino-Laryngology, Department of Neurosciences, Leuven Brain Institute, KU Leuven, Leuven 3000, Belgium
| | - Chandra Leon Haider
- Department of Psychology, Centre for Cognitive Neuroscience, Paris-Lodron-University of Salzburg, Salzburg 5020, Austria
| | - Thomas Hartmann
- Department of Psychology, Centre for Cognitive Neuroscience, Paris-Lodron-University of Salzburg, Salzburg 5020, Austria
| | - Anne Hauswald
- Department of Psychology, Centre for Cognitive Neuroscience, Paris-Lodron-University of Salzburg, Salzburg 5020, Austria
| | | | - Tom Francart
- Experimental Oto-Rhino-Laryngology, Department of Neurosciences, Leuven Brain Institute, KU Leuven, Leuven 3000, Belgium
| | - Nathan Weisz
- Department of Psychology, Centre for Cognitive Neuroscience, Paris-Lodron-University of Salzburg, Salzburg 5020, Austria
- Neuroscience Institute, Christian Doppler University Hospital, Paracelsus Medical University Salzburg, Salzburg 5020, Austria
| |
Collapse
|
3
|
Carranante G, Cany C, Farri P, Giavazzi M, Varnet L. Mapping the spectrotemporal regions influencing perception of French stop consonants in noise. Sci Rep 2024; 14:27183. [PMID: 39516258 PMCID: PMC11549473 DOI: 10.1038/s41598-024-77634-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2024] [Accepted: 10/23/2024] [Indexed: 11/16/2024] Open
Abstract
Understanding how speech sounds are decoded into linguistic units has been a central research challenge over the last century. This study follows a reverse-correlation approach to reveal the acoustic cues listeners use to categorize French stop consonants in noise. Compared to previous methods, this approach ensures an unprecedented level of detail with only minimal theoretical assumptions. Thirty-two participants performed a speech-in-noise discrimination task based on natural /aCa/ utterances, with C = /b/, /d/, /g/, /p/, /t/, or /k/. The trial-by-trial analysis of their confusions enabled us to map the spectrotemporal information they relied on for their decisions. In place-of-articulation contrasts, the results confirmed the critical role of formant consonant-vowel transitions, used by all participants, and, to a lesser extent, vowel-consonant transitions and high-frequency release bursts. Similarly, for voicing contrasts, we validated the prominent role of the voicing bar cue, with some participants also using formant transitions and burst cues. This approach revealed that most listeners use a combination of several cues for each task, with significant variability within the participant group. These insights shed new light on decades-old debates regarding the relative importance of cues for phoneme perception and suggest that research on acoustic cues should not overlook individual variability in speech perception.
Collapse
Affiliation(s)
- Géraldine Carranante
- Laboratoire des systèmes perceptifs, Département d'études cognitives, École normale supérieure, PSL University, CNRS, 75005, Paris, France
| | - Charlotte Cany
- Laboratoire des systèmes perceptifs, Département d'études cognitives, École normale supérieure, PSL University, CNRS, 75005, Paris, France
| | - Pouria Farri
- Laboratoire des systèmes perceptifs, Département d'études cognitives, École normale supérieure, PSL University, CNRS, 75005, Paris, France
| | - Maria Giavazzi
- Laboratoire de Sciences Cognitives et Psycholinguistique, Département d'études cognitives, École normale supérieure, PSL University, CNRS, 75005, Paris, France
| | - Léo Varnet
- Laboratoire des systèmes perceptifs, Département d'études cognitives, École normale supérieure, PSL University, CNRS, 75005, Paris, France.
| |
Collapse
|
4
|
Osses A, Varnet L. A microscopic investigation of the effect of random envelope fluctuations on phoneme-in-noise perception. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 155:1469-1485. [PMID: 38364046 DOI: 10.1121/10.0024469] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/01/2023] [Accepted: 01/05/2024] [Indexed: 02/18/2024]
Abstract
In this study, we investigated the effect of specific noise realizations on the discrimination of two consonants, /b/ and /d/. For this purpose, we collected data from twelve participants, who listened to /aba/ or /ada/ embedded in one of three background noises. All noises had the same long-term spectrum but differed in the amount of random envelope fluctuations. The data were analyzed on a trial-by-trial basis using the reverse-correlation method. The results revealed that it is possible to predict the categorical responses with better-than-chance accuracy purely based on the spectro-temporal distribution of the random envelope fluctuations of the corresponding noises, without taking into account the actual targets or the signal-to-noise ratios used in the trials. The effect of the noise fluctuations explained on average 8.1% of the participants' responses in white noise, a proportion that increased up to 13.3% for noises with a larger amount of fluctuations. The estimated time-frequency weights revealed that the measured effect originated from confusions between noise fluctuations and relevant acoustic cues from the target sounds. Similar conclusions were obtained from simulations using an artificial listener.
Collapse
Affiliation(s)
- Alejandro Osses
- Laboratoire des Systèmes Perceptifs, Département d'Études Cognitives, École Normale Supérieure, PSL University, Centre National de la Recherche Scientifique, 75005 Paris, France
| | - Léo Varnet
- Laboratoire des Systèmes Perceptifs, Département d'Études Cognitives, École Normale Supérieure, PSL University, Centre National de la Recherche Scientifique, 75005 Paris, France
| |
Collapse
|
5
|
Koprowska A, Marozeau J, Dau T, Serman M. The effect of phoneme-based auditory training on speech intelligibility in hearing-aid users. Int J Audiol 2023; 62:1048-1058. [PMID: 36301675 DOI: 10.1080/14992027.2022.2135032] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2022] [Accepted: 10/04/2022] [Indexed: 11/05/2022]
Abstract
OBJECTIVE Hearing loss commonly causes difficulties in understanding speech in the presence of background noise. The benefits of hearing-aids in terms of speech intelligibility in challenging listening scenarios remain limited. The present study investigated if phoneme-in-noise discrimination training improves phoneme identification and sentence intelligibility in noise in hearing-aid users. DESIGN Two groups of participants received either a two-week training program or a control intervention. Three phoneme categories were trained: onset consonants (C1), vowels (V) and post-vowel consonants (C2) in C1-V-C2-/i/ logatomes from the Danish nonsense word corpus (DANOK). Phoneme identification test and hearing in noise test (HINT) were administered before and after the respective interventions and, for the training group only, after three months. STUDY SAMPLE Twenty 63-to-79 years old individuals with a mild-to-moderate sensorineural hearing loss and at least one year of experience using hearing-aids. RESULTS The training provided an improvement in phoneme identification scores for vowels and post-vowel consonants, which was retained over three months. No significant performance improvement in HINT was found. CONCLUSION The study demonstrates that the training induced a robust refinement of auditory perception at a phoneme level but provides no evidence for the generalisation to an untrained sentence intelligibility task.
Collapse
Affiliation(s)
- Aleksandra Koprowska
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Lyngby, Denmark
- Copenhagen Hearing and Balance Center, Rigshospitalet, Copenhagen, Denmark
| | - Jeremy Marozeau
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Lyngby, Denmark
- Copenhagen Hearing and Balance Center, Rigshospitalet, Copenhagen, Denmark
| | - Torsten Dau
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Lyngby, Denmark
- Copenhagen Hearing and Balance Center, Rigshospitalet, Copenhagen, Denmark
| | | |
Collapse
|
6
|
Mishra SK, Moore DR. Auditory Deprivation during Development Alters Efferent Neural Feedback and Perception. J Neurosci 2023; 43:4642-4649. [PMID: 37221095 PMCID: PMC10286938 DOI: 10.1523/jneurosci.2182-22.2023] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2022] [Revised: 05/09/2023] [Accepted: 05/14/2023] [Indexed: 05/25/2023] Open
Abstract
Auditory experience plays a critical role in hearing development. Developmental auditory deprivation because of otitis media, a common childhood disease, produces long-standing changes in the central auditory system, even after the middle ear pathology is resolved. The effects of sound deprivation because of otitis media have been mostly studied in the ascending auditory system but remain to be examined in the descending pathway that runs from the auditory cortex to the cochlea via the brainstem. Alterations in the efferent neural system could be important because the descending olivocochlear pathway influences the neural representation of transient sounds in noise in the afferent auditory system and is thought to be involved in auditory learning. Here, we show that the inhibitory strength of the medial olivocochlear efferents is weaker in children with a documented history of otitis media relative to controls; both boys and girls were included in the study. In addition, children with otitis media history required a higher signal-to-noise ratio on a sentence-in-noise recognition task than controls to achieve the same criterion performance level. Poorer speech-in-noise recognition, a hallmark of impaired central auditory processing, was related to efferent inhibition, and could not be attributed to the middle ear or cochlear mechanics.SIGNIFICANCE STATEMENT Otitis media is the second most common reason children go to the doctor. Previously, degraded auditory experience because of otitis media has been associated with reorganized ascending neural pathways, even after middle ear pathology resolved. Here, we show that altered afferent auditory input because of otitis media during childhood is also associated with long-lasting reduced descending neural pathway function and poorer speech-in-noise recognition. These novel, efferent findings may be important for the detection and treatment of childhood otitis media.
Collapse
Affiliation(s)
- Srikanta K Mishra
- Department of Speech, Language and Hearing Sciences, University of Texas at Austin, Austin, Texas 78712
| | - David R Moore
- Communication Sciences Research Center, Cincinnati Children's Hospital, Cincinnati, Ohio 45229
- Department of Otolaryngology, College of Medicine, University of Cincinnati, Cincinnati, Ohio 45267
- Manchester Centre for Audiology and Deafness, University of Manchester, Manchester, M13 9PL, United Kingdom
| |
Collapse
|
7
|
Shen Y, Langley L. Spectral weighting for sentence recognition in steady-state and amplitude-modulated noise. JASA EXPRESS LETTERS 2023; 3:2887651. [PMID: 37125871 PMCID: PMC10155216 DOI: 10.1121/10.0017934] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/24/2022] [Accepted: 04/06/2023] [Indexed: 05/03/2023]
Abstract
Spectral weights in octave-frequency bands from 0.25 to 4 kHz were estimated for speech-in-noise recognition using two sentence materials (i.e., the IEEE and AzBio sentences). The masking noise was either unmodulated or sinusoidally amplitude-modulated at 8 Hz. The estimated spectral weights did not vary significantly across two test sessions and were similar for the two sentence materials. Amplitude-modulating the masker increased the weight at 2 kHz and decreased the weight at 0.25 kHz, which may support an upward shift in spectral weights for temporally fluctuating maskers.
Collapse
Affiliation(s)
- Yi Shen
- Department of Speech and Hearing Sciences, University of Washington, 1417 Northeast 42nd Street, Seattle, Washington 98105-6246, ,
| | - Lauren Langley
- Department of Speech and Hearing Sciences, University of Washington, 1417 Northeast 42nd Street, Seattle, Washington 98105-6246, ,
| |
Collapse
|
8
|
Yoon YS, Morgan D. Dichotic spectral integration range for consonant recognition in listeners with normal hearing. Front Psychol 2022; 13:1009463. [PMID: 36337493 PMCID: PMC9633255 DOI: 10.3389/fpsyg.2022.1009463] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2022] [Accepted: 10/05/2022] [Indexed: 01/16/2023] Open
Abstract
Dichotic spectral integration range, or DSIR, was measured for consonant recognition with normal-hearing listeners. DSIR is defined as a frequency range needed from 0 to 8,000 Hz band in one ear for consonant recognition when low-frequency information of the same consonant was presented to the opposite ear. DSIR was measured under the three signal processing conditions: (1) unprocessed, (2) target: intensified target spectro-temporal regions by 6 dB responsible for consonant recognition, and (3) target minus conflicting: intensified target regions minus spectro-temporal regions that increase confusion. Each consonant was low-pass filtered with a cutoff frequency of 250, 500, 750, and 1,000 Hz, and then was presented in the left ear or low-frequency (LF) ear. To create dichotic listening, the same consonant was simultaneously presented to the right ear or high-frequency (HF) ear. This was high-pass filtered with an initial cutoff frequency of 7,000 Hz, which was adjusted using an adaptive procedure to find the maximum high-pass cutoff for 99.99% correct consonant recognition. Mean DSIRs spanned from 3,198-8,000 Hz to 4,668-8,000 Hz (i.e., mid-to-high frequencies were unnecessary), depending on low-frequency information in the LF ear. DSIRs narrowed (i.e., required less frequency information) with increasing low-frequency information in the LF ear. However, the mean DSIRs were not significantly affected by the signal processing except at the low-pass cutoff frequency of 250 Hz. The individual consonant analyses revealed that /ta/, /da/, /sa/, and /za/ required the smallest DSIR, while /ka/, /ga/, /fa/, and /va/ required the largest DSIRs. DSIRs also narrowed with increasing low-frequency information for the two signal processing conditions except for 250 vs. 1,000 Hz under the target-conflicting condition. The results suggest that consonant recognition is possible with large amounts of spectral information missing if complementary spectral information is integrated across ears. DSIR is consonant-specific and relatively consistent, regardless of signal processing. The results will help determine the minimum spectral range needed in one ear for consonant recognition if limited low spectral information is available in the opposite ear.
Collapse
|
9
|
Alkharabsheh A, Aboudi O, Abdulbaqi K, Garadat S. The effect of wearing face mask on speech intelligibility in listeners with sensorineural hearing loss and normal hearing sensitivity. Int J Audiol 2022; 62:328-333. [PMID: 35235469 DOI: 10.1080/14992027.2022.2045366] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/27/2022]
Abstract
OBJECTIVE This study aimed to investigate the effect of wearing a face mask on word recognition in hearing-impaired listeners. DESIGN Word recognition scores were obtained in quiet and in different signal-to-noise ratios (SNRs 0, +5, and +10 dB) using two listening conditions (with N95 mask and with no-mask). STUDY SAMPLE Participants were forty-six listeners with normal hearing sensitivity and thirty-nine listeners with mild to moderately severe sensorineural hearing loss. RESULTS Results showed that wearing a face mask affected word recognition and that this effect was greater for listeners with hearing impairment than that observed for listeners with normal hearing sensitivity. The extent of this effect was also dependent on the SNR conditions such that the effect of wearing a face mask was worse in adverse listening conditions. CONCLUSIONS Overall, this study demonstrated that face mask can disrupt speech intelligibility possibly by degrading some acoustical features which may pose substantial difficulties for those with hearing impairment.
Collapse
Affiliation(s)
- Ana'am Alkharabsheh
- Department of Hearing and Speech Sciences, Faculty of Rehabilitation Sciences, University of Jordan, Amman, Jordan
| | - Ola Aboudi
- Ear Nose and Throat Clinic, University of Jordan Hospital, University of Jordan, Amman, Jordan
| | - Khader Abdulbaqi
- Department of Hearing and Speech Sciences, Faculty of Rehabilitation Sciences, University of Jordan, Amman, Jordan.,Ear Nose and Throat Clinic, University of Jordan Hospital, University of Jordan, Amman, Jordan
| | - Soha Garadat
- Department of Hearing and Speech Sciences, Faculty of Rehabilitation Sciences, University of Jordan, Amman, Jordan
| |
Collapse
|
10
|
Diao T, Ma X, Zhang J, Duan M, Yu L. The Correlation Between Hearing Loss, Especially High-Frequency Hearing Loss and Cognitive Decline Among the Elderly. Front Neurosci 2021; 15:750874. [PMID: 34867162 PMCID: PMC8634596 DOI: 10.3389/fnins.2021.750874] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2021] [Accepted: 10/18/2021] [Indexed: 11/29/2022] Open
Abstract
Objective: The relation between cognition and hearing loss has been increasingly paid high attention, however, few studies have focused on the role of high-frequency hearing loss in cognitive decline. This study is oriented to role of hearing loss especially high-frequency hearing loss in cognitive impairment among elderly people (age ≥ 60 years). Methods: The Montreal Cognitive Assessment Scale (MoCA) and pure tone audiometry were used to investigate the hearing loss and cognitive function of 201 elderly people older than 60 years. Factors possibly related to cognitive impairment including age, years of education, occupation, living conditions, history of otologic diseases, and high blood pressure were registered. This study consisted of two parts. First, univariate analysis and multiple linear regressions were performed to analyze the possible influencing factors of cognitive function among the 201 elderly people. Second, average hearing thresholds of low frequencies (250, 500 Hz), intermediate frequencies (1 k, 2 kHz), and high frequencies (4 k, 8 kHz) were calculated to screen out 40 cases with high-frequency hearing loss alone and 18 cases with normal hearing. Univariate analysis was used to compare the general condition, cognitive function, and each cognitive domain between the two groups, analyzing the relation between high-frequency hearing loss and cognitive function. Result: We found that age, years of education, pure tone average (PTA), occupation, living condition, history of otologic diseases, years of self-reported hearing loss, and hypertension history were related to cognitive function. Furthermore, age, education experience, duration of self-reported hearing loss, and hypertension were independent factors (p < 0.05). PTA was negatively related with attention, orientation, and general cognition (p < 0.05). There were only 18 cases (9.0%) with normal hearing, and 40 cases (19.9%) with abnormal high-frequency hearing alone. The overall cognitive function showed no significant difference between them (p > 0.05); in contrast, the speech and abstract ability were significantly decreased in cases with high-frequency hearing loss (p < 0.05). Conclusion: The increase of PTA among the elderly may affect the overall cognition by reducing attention and orientation. High-frequency hearing loss alone can affect the language and abstract ability to a certain extent, which is worthy of more attention.
Collapse
Affiliation(s)
- Tongxiang Diao
- Department of Otolaryngology, Head and Neck Surgery, People's Hospital, Peking University, Beijing, China
| | - Xin Ma
- Department of Otolaryngology, Head and Neck Surgery, People's Hospital, Peking University, Beijing, China
| | - Junbo Zhang
- Department of Otolaryngology, Head and Neck Surgery, Peking University First Hospital, Beijing, China
| | - Maoli Duan
- Department of Clinical Science, Intervention and Technology, Karolinska Institute, Stockholm, Sweden.,Department of Otolaryngology, Head and Neck Surgery & Audiology and Neurotology, Karolinska University Hospital, Karolinska Institute, Stockholm, Sweden
| | - Lisheng Yu
- Department of Otolaryngology, Head and Neck Surgery, People's Hospital, Peking University, Beijing, China
| |
Collapse
|
11
|
Roverud E, Dubno JR, Richards VM, Kidd G. Cross-frequency weights in normal and impaired hearing: Stimulus factors, stimulus dimensions, and associations with speech recognition. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 150:2327. [PMID: 34717459 PMCID: PMC8637742 DOI: 10.1121/10.0006450] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/09/2021] [Revised: 09/02/2021] [Accepted: 09/08/2021] [Indexed: 06/13/2023]
Abstract
Previous studies of level discrimination reported that listeners with high-frequency sensorineural hearing loss (SNHL) place greater weight on high frequencies than normal-hearing (NH) listeners. It is not clear whether these results are influenced by stimulus factors (e.g., group differences in presentation levels, cross-frequency discriminability of level differences used to measure weights) and whether such weights generalize to other tasks. Here, NH and SNHL weights were measured for level, duration, and frequency discrimination of two-tone complexes after measuring discriminability just-noticeable differences for each frequency and stimulus dimension. Stimuli were presented at equal sensation level (SL) or equal sound pressure level (SPL). Results showed that weights could change depending on which frequency contained the more discriminable level difference with uncontrolled cross-frequency discriminability. When cross-frequency discriminability was controlled, weights were consistent for level and duration discrimination, but not for frequency discrimination. Comparing equal SL and equal SPL weights indicated greater weight on the higher-level tone for level and duration discrimination. Weights were unrelated to improvements in recognition of low-pass-filtered speech with increasing cutoff frequency. These results suggest that cross-frequency weights and NH and SNHL weighting differences are influenced by stimulus factors and may not generalize to the use of speech cues in specific frequency regions.
Collapse
Affiliation(s)
- Elin Roverud
- Department of Speech, Language, and Hearing Sciences, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| | - Judy R Dubno
- Department of Otolaryngology-Head and Neck Surgery, Medical University of South Carolina, 135 Rutledge Avenue, MSC 550, Charleston, South Carolina 29425-5500, USA
| | - Virginia M Richards
- Department of Cognitive Sciences, 2201 Social and Behavioral Sciences Gateway, University of California-Irvine, Irvine, California 92697-5100, USA
| | - Gerald Kidd
- Department of Speech, Language, and Hearing Sciences, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| |
Collapse
|
12
|
Varnet L, Léger AC, Boucher S, Bonnet C, Petit C, Lorenzi C. Contributions of Age-Related and Audibility-Related Deficits to Aided Consonant Identification in Presbycusis: A Causal-Inference Analysis. Front Aging Neurosci 2021; 13:640522. [PMID: 33732140 PMCID: PMC7956988 DOI: 10.3389/fnagi.2021.640522] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/11/2020] [Accepted: 02/08/2021] [Indexed: 12/05/2022] Open
Abstract
The decline of speech intelligibility in presbycusis can be regarded as resulting from the combined contribution of two main groups of factors: (1) audibility-related factors and (2) age-related factors. In particular, there is now an abundant scientific literature on the crucial role of suprathreshold auditory abilities and cognitive functions, which have been found to decline with age even in the absence of audiometric hearing loss. However, researchers investigating the direct effect of aging in presbycusis have to deal with the methodological issue that age and peripheral hearing loss covary to a large extent. In the present study, we analyzed a dataset of consonant-identification scores measured in quiet and in noise for a large cohort (n = 459, age = 42-92) of hearing-impaired (HI) and normal-hearing (NH) listeners. HI listeners were provided with a frequency-dependent amplification adjusted to their audiometric profile. Their scores in the two conditions were predicted from their pure-tone average (PTA) and age, as well as from their Extended Speech Intelligibility Index (ESII), a measure of the impact of audibility loss on speech intelligibility. We relied on a causal-inference approach combined with Bayesian modeling to disentangle the direct causal effects of age and audibility on intelligibility from the indirect effect of age on hearing loss. The analysis revealed that the direct effect of PTA on HI intelligibility scores was 5 times higher than the effect of age. This overwhelming effect of PTA was not due to a residual audibility loss despite amplification, as confirmed by a ESII-based model. More plausibly, the marginal role of age could be a consequence of the relatively little cognitively-demanding task used in this study. Furthermore, the amount of variance in intelligibility scores was smaller for NH than HI listeners, even after accounting for age and audibility, reflecting the presence of additional suprathreshold deficits in the latter group. Although the non-sense-syllable materials and the particular amplification settings used in this study potentially restrict the generalization of the findings, we think that these promising results call for a wider use of causal-inference analysis in audiology, e.g., as a way to disentangle the influence of the various cognitive factors and suprathreshold deficits associated to presbycusis.
Collapse
Affiliation(s)
- Léo Varnet
- Laboratoire des Systèmes Perceptifs, UMR CNRS 8248, Département d'Études Cognitives, École normale supérieure, Université Paris Sciences & Lettres, Paris, France
| | - Agnès C. Léger
- Manchester Centre for Audiology and Deafness, Division of Human Communication, Development & Hearing, School of Health Sciences, Faculty of Biology, Medicine and Health, Manchester Academic Health Science Centre, University of Manchester, Manchester, United Kingdom
| | - Sophie Boucher
- Complexité du Vivant, Sorbonne Universités, Université Pierre et Marie Curie, Université Paris VI, Paris, France
- Institut de l'Audition, Institut Pasteur, INSERM, Paris, France
- Centre Hospitalier Universitaire d'Angers, Angers, France
| | - Crystel Bonnet
- Complexité du Vivant, Sorbonne Universités, Université Pierre et Marie Curie, Université Paris VI, Paris, France
- Institut de l'Audition, Institut Pasteur, INSERM, Paris, France
| | - Christine Petit
- Institut de l'Audition, Institut Pasteur, INSERM, Paris, France
- Collège de France, Paris, France
| | - Christian Lorenzi
- Laboratoire des Systèmes Perceptifs, UMR CNRS 8248, Département d'Études Cognitives, École normale supérieure, Université Paris Sciences & Lettres, Paris, France
| |
Collapse
|
13
|
Shen Y, Yun D, Liu Y. Individualized estimation of the Speech Intelligibility Index for short sentences: Test-retest reliability. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 148:1647. [PMID: 33003860 PMCID: PMC7511242 DOI: 10.1121/10.0001994] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/17/2023]
Abstract
The speech intelligibility index (SII) model was modified to allow individualized parameters. These parameters included the relative weights of speech cues in five octave-frequency bands ranging from 0.25 to 4 kHz, i.e., the band importance function, and the transfer function that allows the SII to generate predictions on speech-recognition scores. A Bayesian adaptive procedure, the quick-band-importance-function (qBIF) procedure, was utilized to enable efficient estimation of the SII parameters from individual listeners. In two experiments, the SII parameters were estimated for 30 normal-hearing adults using Institute of Electrical and Electronics Engineers (IEEE) sentences at speech levels of 55, 65, and 75 dB sound pressure level (in Experiment I) and for 15 hearing-impaired (HI) adult listeners using amplified IEEE or AzBio sentences (in Experiment II). In both experiments, even without prior training, the estimated model parameters showed satisfactory reliability between two runs of the qBIF procedure at least one week apart. For the HI listeners, inter-listener variability in most estimated SII parameters was larger than intra-listener variability of the qBIF procedure.
Collapse
Affiliation(s)
- Yi Shen
- Department of Speech and Hearing Sciences, University of Washington, 1417 Northeast 42nd Street, Seattle, Washington 98105-6246, USA
| | - Donghyeon Yun
- Department of Speech, Language and Hearing Sciences, Indiana University Bloomington, 200 South Jordan Avenue, Bloomington, Indiana 47405, USA
| | - Yi Liu
- Department of Speech, Language and Hearing Sciences, Indiana University Bloomington, 200 South Jordan Avenue, Bloomington, Indiana 47405, USA
| |
Collapse
|
14
|
Roverud E, Dubno JR, Kidd G. Hearing-Impaired Listeners Show Reduced Attention to High-Frequency Information in the Presence of Low-Frequency Information. Trends Hear 2020; 24:2331216520945516. [PMID: 32853117 PMCID: PMC7557677 DOI: 10.1177/2331216520945516] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2019] [Revised: 06/30/2020] [Accepted: 07/07/2020] [Indexed: 11/29/2022] Open
Abstract
Many listeners with sensorineural hearing loss have uneven hearing sensitivity across frequencies. This study addressed whether this uneven hearing loss leads to a biasing of attention to different frequency regions. Normal-hearing (NH) and hearing-impaired (HI) listeners performed a pattern discrimination task at two distant center frequencies (CFs): 750 and 3500 Hz. The patterns were sequences of pure tones in which each successive tonal element was randomly selected from one of two possible frequencies surrounding a CF. The stimuli were presented at equal sensation levels to ensure equal audibility. In addition, the frequency separation of the tonal elements within a pattern was adjusted for each listener so that equal pattern discrimination performance was obtained for each CF in quiet. After these adjustments, the pattern discrimination task was performed under conditions in which independent patterns were presented at both CFs simultaneously. The listeners were instructed to attend to the low or high CF before the stimulus (assessing selective attention to frequency with instruction) or after the stimulus (divided attention, assessing inherent frequency biases). NH listeners demonstrated approximately equal performance decrements (re: quiet) between the two CFs. HI listeners demonstrated much larger performance decrements at the 3500 Hz CF than at the 750 Hz CF in combined-presentation conditions for both selective and divided attention conditions, indicating a low-frequency attentional bias that is apparently not under subject control. Surprisingly, the magnitude of this frequency bias was not related to the degree of asymmetry in thresholds at the two CFs.
Collapse
Affiliation(s)
- Elin Roverud
- Department of Speech, Language & Hearing Sciences, Boston University
| | - Judy R. Dubno
- Department of Otolaryngology-Head and Neck Surgery, Medical University of South Carolina
| | - Gerald Kidd
- Department of Speech, Language & Hearing Sciences, Boston University
| |
Collapse
|