1
|
Compression and amplification algorithms in hearing aids impair the selectivity of neural responses to speech. Nat Biomed Eng 2021; 6:717-730. [PMID: 33941898 PMCID: PMC7612903 DOI: 10.1038/s41551-021-00707-y] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2020] [Accepted: 02/25/2021] [Indexed: 02/07/2023]
Abstract
In quiet environments, hearing aids improve the perception of low-intensity sounds. However, for high-intensity sounds in background noise, the aids often fail to provide a benefit to the wearer. Here, by using large-scale single-neuron recordings from hearing-impaired gerbils — an established animal model of human hearing — we show that hearing aids restore the sensitivity of neural responses to speech, but not their selectivity. Rather than reflecting a deficit in supra-threshold auditory processing, the low selectivity is a consequence of hearing-aid compression (which decreases the spectral and temporal contrasts of incoming sound) and of amplification (which distorts neural responses, regardless of whether hearing is impaired). Processing strategies that avoid the trade-off between neural sensitivity and selectivity should improve the performance of hearing aids.
Collapse
|
2
|
Apeksha K, Kumar AU. Effect of Quiet and Noise on P300 Response in Individuals with Auditory Neuropathy Spectrum Disorder. Int Arch Otorhinolaryngol 2020; 24:e462-e471. [PMID: 33101512 PMCID: PMC7575403 DOI: 10.1055/s-0039-3402441] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2019] [Accepted: 11/04/2019] [Indexed: 10/31/2022] Open
Abstract
Introduction Auditory neuropathy spectrum disorder (ANSD) is a clinical condition in which individuals have normal cochlear responses and abnormal neural responses. There is a lack of evidence in the literature regarding the neural discrimination skill in individuals with ANSD, especially when the signal is presented in the presence of noise. Objectives The present study was performed with the aim to investigate auditory discrimination skill, in quiet and in the presence of noise, in individuals with ANSD and to compare the findings with normal-hearing individuals. Methods A total of 30 individuals with normal hearing sensitivity and 30 individuals with ANSD in the age range of 15 to 55 years old, with the mean age of 27.86 years old, were the participants. P300 response was recorded from both groups using syllable pairs /ba/-/da/ in oddball paradigm and the syllable /da/ in repetitive paradigm in quiet and at +10 dB signal-to-noise ratio (SNR). Results There was significant prolongation in latency and reaction time, and reduction in amplitude of P300 response and sensitivity in both groups with the addition of noise. The topographic pattern analysis showed activation of the central-parietal-occipital region of the brain in individuals with ANSD, whereas activation of the central-parietal region was observed in individuals with normal hearing. The activation was more diffused in individuals with ANSD compared with that of individuals with normal hearing. Conclusion The individuals with ANSD showed a significantly more adverse effect of noise on the neural discrimination skill than the normal counterpart.
Collapse
Affiliation(s)
- Kumari Apeksha
- Department of Speech and Hearing, JSS Institute of Speech & Hearing, Mysuru, India
| | - Ajith U. Kumar
- Department of Audiology, All India Institute of Speech & Hearing, Mysuru, India
| |
Collapse
|
3
|
Relaño-Iborra H, Zaar J, Dau T. A speech-based computational auditory signal processing and perception model. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:3306. [PMID: 31795663 DOI: 10.1121/1.5129114] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/09/2019] [Accepted: 09/21/2019] [Indexed: 06/10/2023]
Abstract
A new speech intelligibility prediction model is presented which is based on the Computational Auditory Signal Processing and Perception model (CASP) of Jepsen, Ewert, and Dau [(2008). J. Acoust. Soc. Am. 124(1), 422-438]. The model combines a non-linear auditory-inspired preprocessing with a backend based on the cross-correlation between the clean and the degraded speech representations in the modulation envelope domain. Several speech degradation and speech enhancement algorithms were considered to study the ability of the model to predict data from normal-hearing listeners. Degradations of speech intelligibility due to additive noise, phase-jitter distortion, and single-channel noise reduction as well as improved speech intelligibility due to ideal binary mask processing are shown to be successfully accounted for by the model. Furthermore, the model reflects stimulus-level dependent effects of auditory perception, including audibility limitations at low levels and degraded speech intelligibility at high levels. Given its realistic non-linear auditory processing frontend, the speech-based computational auditory signal processing and perception model may provide a valuable computational framework for studying the effects of sensorineural hearing impairment on speech intelligibility.
Collapse
Affiliation(s)
- Helia Relaño-Iborra
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, DK-2800 Kgs. Lyngby, Denmark
| | - Johannes Zaar
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, DK-2800 Kgs. Lyngby, Denmark
| | - Torsten Dau
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, DK-2800 Kgs. Lyngby, Denmark
| |
Collapse
|
4
|
Liu C, Jin SH. Psychometric Functions of Vowel Detection and Identification in Long-Term Speech-Shaped Noise. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:1473-1485. [PMID: 30995169 DOI: 10.1044/2018_jslhr-h-18-0320] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
Purpose The goal of this study was to investigate vowel detection and identification in noise and provide baseline data regarding how vowel perception changed with signal-to-noise ratios. Psychometric functions of vowel detection and identification for 12 American English isolated vowels in long-term speech-shaped noise were examined for young listeners with normal hearing in this study. Method Vowel detection was measured at sensation levels from -10 to +5 dB (re: thresholds of vowel detection from the study of Liu and Eddins, 2008a ) with a 4-interval forced-choice procedure. Thresholds of vowel detection were computed for each listener as the speech level at which 70.7% correct performance was reached. Vowel identification was then examined at sensation levels from 0 to 12 dB relative to detection thresholds for each listener. Thresholds of vowel identification were calculated as the speech level with vowel identifiability ( d') equals to 1. Results Thresholds of vowel detection and identification were significantly affected by vowel category. Slopes of psychometric functions of vowel identification were significantly dependent on vowel category, whereas slopes of psychometric functions of vowel detection were not. Conclusions These results suggest that, given the same sensation levels, especially at low sensation levels, vowel sounds are not equally perceivable in terms of identifiability.
Collapse
Affiliation(s)
- Chang Liu
- Department of Communication Sciences and Disorders, University of Texas at Austin
| | - Su-Hyun Jin
- Department of Communication Sciences and Disorders, California State University, Northridge
| |
Collapse
|
5
|
Electrophysiology and Perception of Speech in Noise in Older Listeners: Effects of Hearing Impairment and Age. Ear Hear 2016; 36:710-22. [PMID: 26502191 DOI: 10.1097/aud.0000000000000191] [Citation(s) in RCA: 51] [Impact Index Per Article: 6.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVES Speech perception in background noise is difficult for many individuals, and there is considerable performance variability across listeners. The combination of physiological and behavioral measures may help to understand sources of this variability for individuals and groups and prove useful clinically with hard-to-test populations. The purpose of this study was threefold: (1) determine the effect of signal-to-noise ratio (SNR) and signal level on cortical auditory evoked potentials (CAEPs) and sentence-level perception in older normal-hearing (ONH) and older hearing-impaired (OHI) individuals, (2) determine the effects of hearing impairment and age on CAEPs and perception, and (3) explore how well CAEPs correlate with and predict speech perception in noise. DESIGN Two groups of older participants (15 ONH and 15 OHI) were tested using speech-in-noise stimuli to measure CAEPs and sentence-level perception of speech. The syllable /ba/, used to evoke CAEPs, and sentences were presented in speech-spectrum background noise at four signal levels (50, 60, 70, and 80 dB SPL) and up to seven SNRs (-10, -5, 0, 5, 15, 25, and 35 dB). These data were compared between groups to reveal the hearing impairment effect and then combined with previously published data for 15 young normal-hearing individuals to determine the aging effect. RESULTS Robust effects of SNR were found for perception and CAEPs. Small but significant effects of signal level were found for perception, primarily at poor SNRs and high signal levels, and in some limited instances for CAEPs. Significant effects of age were seen for both CAEPs and perception, while hearing impairment effects were only found with perception measures. CAEPs correlate well with perception and can predict SNR50s to within 2 dB for ONH. However, prediction error is much larger for OHI and varies widely (from 6 to 12 dB) depending on the model that was used for prediction. CONCLUSIONS When background noise is present, SNR dominates both perception-in-noise testing and cortical electrophysiological testing, with smaller and sometimes significant contributions from signal level. A mismatch between behavioral and electrophysiological results was found (hearing impairment effects were primarily only seen for behavioral data), illustrating the possible contributions of higher order cognitive processes on behavior. It is interesting that the hearing impairment effect size was more than five times larger than the aging effect size for CAEPs and perception. Sentence-level perception can be predicted well in normal-hearing individuals; however, additional research is needed to explore improved prediction methods for older individuals with hearing impairment.
Collapse
|
6
|
Hossain ME, Jassim WA, Zilany MSA. Reference-Free Assessment of Speech Intelligibility Using Bispectrum of an Auditory Neurogram. PLoS One 2016; 11:e0150415. [PMID: 26967160 PMCID: PMC4788356 DOI: 10.1371/journal.pone.0150415] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2015] [Accepted: 02/12/2016] [Indexed: 11/19/2022] Open
Abstract
Sensorineural hearing loss occurs due to damage to the inner and outer hair cells of the peripheral auditory system. Hearing loss can cause decreases in audibility, dynamic range, frequency and temporal resolution of the auditory system, and all of these effects are known to affect speech intelligibility. In this study, a new reference-free speech intelligibility metric is proposed using 2-D neurograms constructed from the output of a computational model of the auditory periphery. The responses of the auditory-nerve fibers with a wide range of characteristic frequencies were simulated to construct neurograms. The features of the neurograms were extracted using third-order statistics referred to as bispectrum. The phase coupling of neurogram bispectrum provides a unique insight for the presence (or deficit) of supra-threshold nonlinearities beyond audibility for listeners with normal hearing (or hearing loss). The speech intelligibility scores predicted by the proposed method were compared to the behavioral scores for listeners with normal hearing and hearing loss both in quiet and under noisy background conditions. The results were also compared to the performance of some existing methods. The predicted results showed a good fit with a small error suggesting that the subjective scores can be estimated reliably using the proposed neural-response-based metric. The proposed metric also had a wide dynamic range, and the predicted scores were well-separated as a function of hearing loss. The proposed metric successfully captures the effects of hearing loss and supra-threshold nonlinearities on speech intelligibility. This metric could be applied to evaluate the performance of various speech-processing algorithms designed for hearing aids and cochlear implants.
Collapse
Affiliation(s)
- Mohammad E. Hossain
- Department of Biomedical Engineering, Faculty of Engineering, University of Malaya, Kuala Lumpur, Malaysia
| | - Wissam A. Jassim
- Department of Biomedical Engineering, Faculty of Engineering, University of Malaya, Kuala Lumpur, Malaysia
| | - Muhammad S. A. Zilany
- Department of Biomedical Engineering, Faculty of Engineering, University of Malaya, Kuala Lumpur, Malaysia
- * E-mail:
| |
Collapse
|
7
|
Calandruccio L, Buss E, Doherty KA. The effect of presentation level on spectral weights for sentences. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 139:466-71. [PMID: 26827040 PMCID: PMC4723404 DOI: 10.1121/1.4940211] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/28/2015] [Revised: 12/17/2015] [Accepted: 01/07/2016] [Indexed: 05/17/2023]
Abstract
Psychophysical data indicate that spectral weights tend to increase with increasing presentation level at high frequencies. The present study examined whether spectral weights for speech perception are similarly affected by presentation level. Stimuli were sentences filtered into five contiguous frequency bands and presented at each of two levels (75 and 95 dB sound pressure level [SPL]). For the highest band (2807-10,000 Hz), normal-hearing listeners' weights were higher for the higher presentation level. Weights for the 95-dB-SPL level resembled those previously estimated for hearing-impaired listeners tested at comparably high levels, suggesting that hearing loss itself may not play a large role in spectral weighting for a sentence recognition task.
Collapse
Affiliation(s)
- Lauren Calandruccio
- Department of Psychological Sciences, Case Western Reserve University, Cleveland, Ohio 44106, USA
| | - Emily Buss
- Department of Otolaryngology/Head and Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina 27599, USA
| | - Karen A Doherty
- Department of Communication Sciences and Disorders, Institute for Sensory Research, Syracuse University, Syracuse, New York 13244, USA
| |
Collapse
|
8
|
|
9
|
Predicting perception in noise using cortical auditory evoked potentials. J Assoc Res Otolaryngol 2013; 14:891-903. [PMID: 24030818 DOI: 10.1007/s10162-013-0415-y] [Citation(s) in RCA: 60] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2013] [Accepted: 08/21/2013] [Indexed: 10/26/2022] Open
Abstract
Speech perception in background noise is a common challenge across individuals and health conditions (e.g., hearing impairment, aging, etc.). Both behavioral and physiological measures have been used to understand the important factors that contribute to perception-in-noise abilities. The addition of a physiological measure provides additional information about signal-in-noise encoding in the auditory system and may be useful in clarifying some of the variability in perception-in-noise abilities across individuals. Fifteen young normal-hearing individuals were tested using both electrophysiology and behavioral methods as a means to determine (1) the effects of signal-to-noise ratio (SNR) and signal level and (2) how well cortical auditory evoked potentials (CAEPs) can predict perception in noise. Three correlation/regression approaches were used to determine how well CAEPs predicted behavior. Main effects of SNR were found for both electrophysiology and speech perception measures, while signal level effects were found generally only for speech testing. These results demonstrate that when signals are presented in noise, sensitivity to SNR cues obscures any encoding of signal level cues. Electrophysiology and behavioral measures were strongly correlated. The best physiological predictors (e.g., latency, amplitude, and area of CAEP waves) of behavior (SNR at which 50 % of the sentence is understood) were N1 latency and N1 amplitude measures. In addition, behavior was best predicted by the 70-dB signal/5-dB SNR CAEP condition. It will be important in future studies to determine the relationship of electrophysiology and behavior in populations who experience difficulty understanding speech in noise such as those with hearing impairment or age-related deficits.
Collapse
|
10
|
Sensitivity of offset and onset cortical auditory evoked potentials to signals in noise. Clin Neurophysiol 2013; 125:370-80. [PMID: 24007688 DOI: 10.1016/j.clinph.2013.08.003] [Citation(s) in RCA: 24] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2013] [Revised: 07/03/2013] [Accepted: 08/05/2013] [Indexed: 11/21/2022]
Abstract
OBJECTIVE The purpose of this study was to determine the effects of SNR and signal level on the offset response of the cortical auditory evoked potential (CAEP). Successful listening often depends on how well the auditory system can extract target signals from competing background noise. Both signal onsets and offsets are encoded neurally and contribute to successful listening in noise. Neural onset responses to signals in noise demonstrate a strong sensitivity to signal-to-noise ratio (SNR) rather than signal level; however, the sensitivity of neural offset responses to these cues is not known. METHODS We analyzed the offset response from two previously published datasets for which only the onset response was reported. For both datasets, CAEPs were recorded from young normal-hearing adults in response to a 1000-Hz tone. For the first dataset, tones were presented at seven different signal levels without background noise, while the second dataset varied both signal level and SNR. RESULTS Offset responses demonstrated sensitivity to absolute signal level in quiet, SNR, and to absolute signal level in noise. CONCLUSIONS Offset sensitivity to signal level when presented in noise contrasts with previously published onset results. SIGNIFICANCE This sensitivity suggests a potential clinical measure of cortical encoding of signal level in noise.
Collapse
|
11
|
Dubno JR, Ahlstrom JB, Wang X, Horwitz AR. Level-dependent changes in perception of speech envelope cues. J Assoc Res Otolaryngol 2012; 13:835-52. [PMID: 22872414 DOI: 10.1007/s10162-012-0343-2] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/17/2012] [Accepted: 07/16/2012] [Indexed: 11/28/2022] Open
Abstract
Level-dependent changes in temporal envelope fluctuations in speech and related changes in speech recognition may reveal effects of basilar-membrane nonlinearities. As a result of compression in the basilar-membrane response, the "effective" magnitude of envelope fluctuations may be reduced as speech level increases from lower level (more linear) to mid-level (more compressive) regions. With further increases to a more linear region, speech envelope fluctuations may become more pronounced. To assess these effects, recognition of consonants and key words in sentences was measured as a function of speech level for younger adults with normal hearing. Consonant-vowel syllables and sentences were spectrally degraded using "noise vocoder" processing to maximize perceptual effects of changes to the speech envelope. Broadband noise at a fixed signal-to-noise ratio maintained constant audibility as speech level increased. Results revealed significant increases in scores and envelope-dependent feature transmission from 45 to 60 dB SPL and decreasing scores and feature transmission from 60 to 85 dB SPL. This quadratic pattern, with speech recognition maximized at mid levels and poorer at lower and higher levels, is consistent with a role of cochlear nonlinearities in perception of speech envelope cues.
Collapse
Affiliation(s)
- Judy R Dubno
- Department of Otolaryngology-Head and Neck Surgery, Medical University of South Carolina, 135 Rutledge Avenue, MSC 550, Charleston, SC 29425-5500, USA.
| | | | | | | |
Collapse
|
12
|
Effects of degree and configuration of hearing loss on the contribution of high- and low-frequency speech information to bilateral speech understanding. Ear Hear 2012; 32:543-55. [PMID: 21336138 DOI: 10.1097/aud.0b013e31820e5028] [Citation(s) in RCA: 48] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
Abstract
OBJECTIVES The purpose of this study was to examine the effects of degree and configuration of hearing loss on the use of, and benefit from, information in amplified high- and low-frequency speech presented in background noise. DESIGN Sixty-two adults with a wide range of high- and low-frequency sensorineural hearing loss (5 to 115+ dB HL) participated in the study. To examine the contribution of speech information in different frequency regions, speech understanding in noise was assessed in multiple low- and high-pass filter conditions, as well as a band-pass (713 to 3534 Hz) and wideband (143 to 8976 Hz) condition. To increase audibility over a wide frequency range, speech and noise were amplified based on each individual's hearing loss. A stepwise multiple linear regression approach was used to examine the contribution of several factors to (1) absolute performance in each filter condition and (2) the change in performance with the addition of amplified high- and low-frequency speech components. RESULTS Results from the regression analysis showed that degree of hearing loss was the strongest predictor of absolute performance for low- and high-pass filtered speech materials. In addition, configuration of hearing loss affected both absolute performance for severely low-pass filtered speech and benefit from extending high-frequency (3534 to 8976 Hz) bandwidth. Specifically, individuals with steeply sloping high-frequency losses made better use of low-pass filtered speech information than individuals with similar low-frequency thresholds but less high-frequency loss. In contrast, given similar high-frequency thresholds, individuals with flat hearing losses received more benefit from extending high-frequency bandwidth than individuals with more sloping losses. CONCLUSIONS Consistent with previous work, benefit from speech information in a given frequency region generally decreases as degree of hearing loss in that frequency region increases. However, given a similar degree of loss, the configuration of hearing loss also affects the ability to use speech information in different frequency regions. Except for individuals with steeply sloping high-frequency losses, providing high-frequency amplification (3534 to 8976 Hz) had either a beneficial effect on, or did not significantly degrade, speech understanding. These findings highlight the importance of extended high-frequency amplification for listeners with a wide range of high-frequency hearing losses, when seeking to maximize intelligibility.
Collapse
|
13
|
Effects of audibility and multichannel wide dynamic range compression on consonant recognition for listeners with severe hearing loss. Ear Hear 2009; 30:494-504. [PMID: 19633563 DOI: 10.1097/aud.0b013e3181aec5bc] [Citation(s) in RCA: 24] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
Abstract
OBJECTIVE This study examined the effects of multichannel wide-dynamic range compression (WDRC) amplification and stimulus audibility on consonant recognition and error patterns. DESIGN Listeners had either severe or mild to moderate sensorineural hearing loss. Each listener was monaurally fit with a wearable hearing aid using typical clinical procedures, frequency-gain parameters, and a hybrid of clinically prescribed compression ratios for desired sensation level () and NAL-NL (). Consonant-vowel nonsense syllables were presented in soundfield at multiple input levels (50, 65, 80 dB SPL). Test conditions were four-channel fast-acting WDRC amplification and a control compression limiting (CL) amplification condition. Listeners identified the stimulus heard from choices presented on an on-screen display. A between-subject repeated measures design was used to evaluate consonant recognition and consonant confusion patterns. RESULTS Fast-acting WDRC provided a considerable audibility advantage at 50 dB SPL, especially for listeners with severe hearing loss. Listeners with mild to moderate hearing loss received less audibility improvement from the fast-acting WDRC amplification, for conversational and high level speech, when compared with listeners with severe hearing loss. Analysis of WDRC benefit scores revealed that listeners had slightly lower scores with fast-acting WDRC amplification (relative to CL) when WDRC provided minimal improvement in audibility. The negative effect was greater for listeners with mild to moderate hearing loss compared with their counterparts with severe hearing loss. CONCLUSIONS All listeners, but particularly the severe loss group, benefited from fast-acting WDRC amplification for low-level speech. For conversational and higher speech levels (i.e., when WDRC does not confer a significant audibility advantage), fast-acting WDRC amplification seems to slightly degrade performance. Listeners' consonant confusion patterns suggest that this negative effect may be partly due to fast-acting WDRC-induced distortions, which alter specific consonant features. In support of this view, audibility accounted for a greater percentage of the variance in listeners' performance with CL amplification compared with fast-acting WDRC amplification.
Collapse
|
14
|
Kewley-Port D, Burkle TZ, Lee JH. Contribution of consonant versus vowel information to sentence intelligibility for young normal-hearing and elderly hearing-impaired listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 122:2365-75. [PMID: 17902871 DOI: 10.1121/1.2773986] [Citation(s) in RCA: 56] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2023]
Abstract
The purpose of this study was to examine the contribution of information provided by vowels versus consonants to sentence intelligibility in young normal-hearing (YNH) and typical elderly hearing-impaired (EHI) listeners. Sentences were presented in three conditions, unaltered or with either the vowels or the consonants replaced with speech shaped noise. Sentences from male and female talkers in the TIMIT database were selected. Baseline performance was established at a 70 dB SPL level using YNH listeners. Subsequently EHI and YNH participants listened at 95 dB SPL. Participants listened to each sentence twice and were asked to repeat the entire sentence after each presentation. Words were scored correct if identified exactly. Average performance for unaltered sentences was greater than 94%. Overall, EHI listeners performed more poorly than YNH listeners. However, vowel-only sentences were always significantly more intelligible than consonant-only sentences, usually by a ratio of 2:1 across groups. In contrast to written English or words spoken in isolation, these results demonstrated that for spoken sentences, vowels carry more information about sentence intelligibility than consonants for both young normal-hearing and elderly hearing-impaired listeners.
Collapse
Affiliation(s)
- Diane Kewley-Port
- Department of Speech and Hearing Sciences, Indiana University, Bloomington, Indiana 47405, USA.
| | | | | |
Collapse
|
15
|
Summers V, Cord MT. Intelligibility of speech in noise at high presentation levels: effects of hearing loss and frequency region. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 122:1130-7. [PMID: 17672659 DOI: 10.1121/1.2751251] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2023]
Abstract
These experiments examined how high presentation levels influence speech recognition for high- and low-frequency stimuli in noise. Normally hearing (NH) and hearing-impaired (HI) listeners were tested. In Experiment 1, high- and low-frequency bandwidths yielding 70%-correct word recognition in quiet were determined at levels associated with broadband speech at 75 dB SPL. In Experiment 2, broadband and band-limited sentences (based on passbands measured in Experiment 1) were presented at this level in speech-shaped noise filtered to the same frequency bandwidths as targets. Noise levels were adjusted to produce approximately 30%-correct word recognition. Frequency bandwidths and signal-to-noise ratios supporting criterion performance in Experiment 2 were tested at 75, 87.5, and 100 dB SPL in Experiment 3. Performance tended to decrease as levels increased. For NH listeners, this "rollover" effect was greater for high-frequency and broadband materials than for low-frequency stimuli. For HI listeners, the 75- to 87.5-dB increase improved signal audibility for high-frequency stimuli and rollover was not observed. However, the 87.5- to 100-dB increase produced qualitatively similar results for both groups: scores decreased most for high-frequency stimuli and least for low-frequency materials. Predictions of speech intelligibility by quantitative methods such as the Speech Intelligibility Index may be improved if rollover effects are modeled as frequency dependent.
Collapse
Affiliation(s)
- Van Summers
- Army Audiology and Speech Center, Walter Reed Army Medical Center, Washington, DC 20307-5001, USA.
| | | |
Collapse
|
16
|
Başkent D. Speech recognition in normal hearing and sensorineural hearing loss as a function of the number of spectral channels. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2006; 120:2908-25. [PMID: 17139748 DOI: 10.1121/1.2354017] [Citation(s) in RCA: 55] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/12/2023]
Abstract
Speech recognition by normal-hearing listeners improves as a function of the number of spectral channels when tested with a noiseband vocoder simulating cochlear implant signal processing. Speech recognition by the best cochlear implant users, however, saturates around eight channels and does not improve when more electrodes are activated, presumably due to reduced frequency selectivity caused by channel interactions. Listeners with sensorineural hearing loss may also have reduced frequency selectivity due to cochlear damage and the resulting reduction in the nonlinear cochlear mechanisms. The present study investigates whether such a limitation in spectral information transmission would be observed with hearing-impaired listeners, similar to implant users. To test the hypothesis, hearing-impaired subjects were selected from a population of patients with moderate hearing loss of cochlear origin, where the frequency selectivity would be expected to be poorer compared to normal hearing. Hearing-impaired subjects were tested for vowel and consonant recognition in steady-state background noise of varying levels using a noiseband vocoder and as a function of the number of spectral channels. For comparison, normal-hearing subjects were tested with the same stimuli at different presentation levels. In quiet and low background noise, performance by normal-hearing and hearing-impaired subjects was similar. In higher background noise, performance by hearing-impaired subjects saturated around eight channels, while performance by normal-hearing subjects continued to increase up to 12-16 channels with vowels, and 10-12 channels with consonants. A similar trend was observed for most of the presentation levels at which the normal-hearing subjects were tested. Therefore, it is unlikely that the effects observed with hearing-impaired subjects were due to insufficient audibility or high presentation levels. Consequently, the results with hearing-impaired subjects were similar to previous results obtained with implant users, but only for background noise conditions.
Collapse
Affiliation(s)
- Deniz Başkent
- House Ear Institute, Department of Auditory Implants, 2100 West Third Street, Los Angeles, California 90057, USA.
| |
Collapse
|
17
|
Kaplan-Neeman R, Kishon-Rabin L, Henkin Y, Muchnik C. Identification of syllables in noise: electrophysiological and behavioral correlates. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2006; 120:926-33. [PMID: 16938980 DOI: 10.1121/1.2217567] [Citation(s) in RCA: 47] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
This study was designed to characterize the effect of background noise on the identification of syllables using behavioral and electrophysiological measures. Twenty normal-hearing adults (18-30 years) performed an identification task in a two-alternative forced-choice paradigm. Stimuli consisted of naturally produced syllables [da] and [ga] embedded in white noise. The noise was initiated 1000 ms before the onset of the speech stimuli in order to separate the auditory event related potentials (AERP) response to noise onset from that to the speech. Syllables were presented in quiet and in five SNRs: +15, +3, 0, -3, and -6 dB. Results show that (1) performance accuracy, d', and reaction time were affected by the noise, more so for reaction time; (2) both N1 and P3 latency were prolonged as noise levels increased, more so for P3; (3) [ga] was better identified than [da], in all noise conditions; and (4) P3 latency was longer for [da] than for [ga] for SNR 0 through -6 dB, while N1 latency was longer for [ga] than for [da] in most listening conditions. In conclusion, the unique stimuli structure utilized in this study demonstrated the effects of noise on speech recognition at both the physical and the perceptual processing levels.
Collapse
Affiliation(s)
- Ricky Kaplan-Neeman
- Department of Communication Disorders, Sackler Faculty of Medicine, Tel-Aviv University, Israel.
| | | | | | | |
Collapse
|
18
|
Dubno JR, Horwitz AR, Ahlstrom JB. Spectral and threshold effects on recognition of speech at higher-than-normal levels. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2006; 120:310-20. [PMID: 16875228 DOI: 10.1121/1.2206508] [Citation(s) in RCA: 14] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
To examine spectral and threshold effects for speech and noise at high levels, recognition of nonsense syllables was assessed for low-pass-filtered speech and speech-shaped maskers and high-pass-filtered speech and speech-shaped maskers at three speech levels, with signal-to-noise ratio held constant. Subjects were younger adults with normal hearing and older adults with normal hearing but significantly higher average quiet thresholds. A broadband masker was always present to minimize audibility differences between subject groups and across presentation levels. For subjects with lower thresholds, the declines in recognition of low-frequency syllables in low-frequency maskers were attributed to nonlinear growth of masking which reduced "effective" signal-to-noise ratio at high levels, whereas the decline for subjects with higher thresholds was not fully explained by nonlinear masking growth. For all subjects, masking growth did not entirely account for declines in recognition of high-frequency syllables in high-frequency maskers at high levels. Relative to younger subjects with normal hearing and lower quiet thresholds, older subjects with normal hearing and higher quiet thresholds had poorer consonant recognition in noise, especially for high-frequency speech in high-frequency maskers. Age-related effects on thresholds and task proficiency may be determining factors in the recognition of speech in noise at high levels.
Collapse
Affiliation(s)
- Judy R Dubno
- Department of Otolaryngology-Head and Neck Surgery, Medical University of South Carolina, 135 Rutledge Avenue, P.O. Box 250550, Charleston, South Carolina 29425, USA.
| | | | | |
Collapse
|
19
|
Hornsby BWY, Ricketts TA. The effects of hearing loss on the contribution of high- and low-frequency speech information to speech understanding. II. Sloping hearing loss. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2006; 119:1752-63. [PMID: 16583917 PMCID: PMC2735822 DOI: 10.1121/1.2161432] [Citation(s) in RCA: 43] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/08/2023]
Abstract
The speech understanding of persons with sloping high-frequency (HF) hearing impairment (HI) was compared to normal hearing (NH) controls and previous research on persons with "flat" losses [Hornsby and Ricketts (2003). J. Acoust. Soc. Am. 113, 1706-1717] to examine how hearing loss configuration affects the contribution of speech information in various frequency regions. Speech understanding was assessed at multiple low- and high-pass filter cutoff frequencies. Crossover frequencies, defined as the cutoff frequencies at which low- and high-pass filtering yielded equivalent performance, were significantly lower for the sloping HI, compared to NH, group suggesting that HF HI limits the utility of HF speech information. Speech intelligibility index calculations suggest this limited utility was not due simply to reduced audibility but also to the negative effects of high presentation levels and a poorer-than-normal use of speech information in the frequency region with the greatest hearing loss (the HF regions). This deficit was comparable, however, to that seen in low-frequency regions of persons with similar HF thresholds and "flat" hearing losses suggesting that sensorineural HI results in a "uniform," rather than frequency-specific, deficit in speech understanding, at least for persons with HF thresholds up to 60-80 dB HL.
Collapse
Affiliation(s)
- Benjamin W Y Hornsby
- Department of Hearing and Speech Science, Dan Maddox Hearing Aid Research Laboratory, Vanderbilt University, Vanderbilt Bill Wilkerson Center Nashville, Tennessee 37232-8242, USA.
| | | |
Collapse
|