1
|
Kuk F, Slugocki C, Korhonen P. Performance of older normal-hearing listeners on the tracking of noise tolerance (TNT) test. Int J Audiol 2024; 63:393-400. [PMID: 37335169 DOI: 10.1080/14992027.2023.2223760] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2022] [Revised: 05/17/2023] [Accepted: 06/05/2023] [Indexed: 06/21/2023]
Abstract
OBJECTIVE To gather preliminary reference data on older normal-hearing (NH) adults for the refined Tracking of Noise Tolerance (TNT) test. DESIGN Within-subject repeated measures. Participants were tested on the TNT in the sound-field and under headphones. In the sound-field, speech stimuli were presented at 75 dB SPL and 82 dB SPL from 0° with a speech-shaped noise presented either from 0° or 180° at a level controlled by the participants. The order of signal level, mode of presentation, noise azimuth, and TNT passages were counterbalanced across listeners. Testing was repeated for one condition after 1-3 weeks to estimate within-session and between-session reliability. STUDY SAMPLE Twenty-five NH listeners (51-82 yrs of age). RESULTS Mean TNT scores (TNTAve) were about 4 dB at a speech input of 75 dB SPL and 3 dB at 82 dB SPL. The TNTAve was similar between the headphone and sound-field presentations in the co-located noise. TNTAve scores measured with noise-back were about 1 dB better than those measured from the front. The 95% confidence intervals of absolute test-retest differences were about 1.2 dB within-session and 2.0 dB between sessions. CONCLUSIONS The refined TNT may be a reliable tool to measure noise acceptance and subjective speech intelligibility.
Collapse
Affiliation(s)
- Francis Kuk
- WSA Office of Research in Clinical Amplification (ORCA-USA), IL, USA
| | | | - Petri Korhonen
- WSA Office of Research in Clinical Amplification (ORCA-USA), IL, USA
| |
Collapse
|
2
|
Margolis RH, Rao A, Wilson RH, Saly GL. Non-linguistic auditory speech processing. Int J Audiol 2023; 62:217-226. [PMID: 35369837 DOI: 10.1080/14992027.2022.2055654] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
Abstract
OBJECTIVES A method for testing auditory processing of non-linguistic speech-like stimuli was developed and evaluated. DESIGN Monosyllabic words were temporally reversed and distorted. Stimuli were matched for spectrum and level. Listeners discriminated between distorted and undistorted stimuli. STUDY SAMPLE Three groups were tested. The Normal group was comprised of 12 normal-hearing participants. The Senior group was comprised of 12 seniors. The Hearing Loss group was comprised of 12 participants with thresholds of at least 35 dB HL at one or more frequencies. RESULTS The Senior group scored lower than the Normal group, and the Hearing Loss group scored lower than the Senior group. Scores for forward compressed speech were slightly higher than backward compressed speech but the difference was not statistically significant. Retest scores were slightly higher than scores on the first test, but the difference was not statistically significant. CONCLUSIONS Large differences in discrimination of distorted speech were observed among the three groups. Age and hearing loss separately affected performance. The depressed performance of the Senior group may be a result of "hidden hearing loss" that is attributed to cochlear synaptopathy. The backward-distorted speech task may be a useful non-linguistic test of speech processing that is language independent.
Collapse
Affiliation(s)
- Robert H Margolis
- Speech and Hearing Science, College of Health Solutions, Arizona State University, Tempe, Arizona.,Audiology Incorporated, Arden Hills, Minnesota, USA
| | - Aparna Rao
- Speech and Hearing Science, College of Health Solutions, Arizona State University, Tempe, Arizona
| | - Richard H Wilson
- Speech and Hearing Science, College of Health Solutions, Arizona State University, Tempe, Arizona
| | | |
Collapse
|
3
|
Rönnberg J, Signoret C, Andin J, Holmer E. The cognitive hearing science perspective on perceiving, understanding, and remembering language: The ELU model. Front Psychol 2022; 13:967260. [PMID: 36118435 PMCID: PMC9477118 DOI: 10.3389/fpsyg.2022.967260] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2022] [Accepted: 08/08/2022] [Indexed: 11/13/2022] Open
Abstract
The review gives an introductory description of the successive development of data patterns based on comparisons between hearing-impaired and normal hearing participants' speech understanding skills, later prompting the formulation of the Ease of Language Understanding (ELU) model. The model builds on the interaction between an input buffer (RAMBPHO, Rapid Automatic Multimodal Binding of PHOnology) and three memory systems: working memory (WM), semantic long-term memory (SLTM), and episodic long-term memory (ELTM). RAMBPHO input may either match or mismatch multimodal SLTM representations. Given a match, lexical access is accomplished rapidly and implicitly within approximately 100-400 ms. Given a mismatch, the prediction is that WM is engaged explicitly to repair the meaning of the input - in interaction with SLTM and ELTM - taking seconds rather than milliseconds. The multimodal and multilevel nature of representations held in WM and LTM are at the center of the review, being integral parts of the prediction and postdiction components of language understanding. Finally, some hypotheses based on a selective use-disuse of memory systems mechanism are described in relation to mild cognitive impairment and dementia. Alternative speech perception and WM models are evaluated, and recent developments and generalisations, ELU model tests, and boundaries are discussed.
Collapse
Affiliation(s)
- Jerker Rönnberg
- Linnaeus Centre HEAD, Department of Behavioural Sciences and Learning, Linköping University, Linköping, Sweden
| | | | | | | |
Collapse
|
4
|
Distracting Linguistic Information Impairs Neural Tracking of Attended Speech. CURRENT RESEARCH IN NEUROBIOLOGY 2022; 3:100043. [DOI: 10.1016/j.crneur.2022.100043] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2021] [Revised: 04/27/2022] [Accepted: 05/24/2022] [Indexed: 11/20/2022] Open
|
5
|
Scheuregger O, Hjortkjær J, Dau T. Identification and Discrimination of Sound Textures in Hearing-Impaired and Older Listeners. Trends Hear 2021; 25:23312165211065608. [PMID: 34939472 PMCID: PMC8721370 DOI: 10.1177/23312165211065608] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
Sound textures are a broad class of sounds defined by their homogeneous temporal structure. It has been suggested that sound texture perception is mediated by time-averaged summary statistics measured from early stages of the auditory system. The ability of young normal-hearing (NH) listeners to identify synthetic sound textures increases as the statistics of the synthetic texture approach those of its real-world counterpart. In sound texture discrimination, young NH listeners utilize the fine temporal stimulus information for short-duration stimuli, whereas they switch to a time-averaged statistical representation as the stimulus' duration increases. The present study investigated how younger and older listeners with a sensorineural hearing impairment perform in the corresponding texture identification and discrimination tasks in which the stimuli were amplified to compensate for the individual listeners' loss of audibility. In both hearing impaired (HI) listeners and NH controls, sound texture identification performance increased as the number of statistics imposed during the synthesis stage increased, but hearing impairment was accompanied by a significant reduction in overall identification accuracy. Sound texture discrimination performance was measured across listener groups categorized by age and hearing loss. Sound texture discrimination performance was unaffected by hearing loss at all excerpt durations. The older listeners' sound texture and exemplar discrimination performance decreased for signals of short excerpt duration, with older HI listeners performing better than older NH listeners. The results suggest that the time-averaged statistic representations of sound textures provide listeners with cues which are robust to the effects of age and sensorineural hearing loss.
Collapse
Affiliation(s)
- Oliver Scheuregger
- Hearing Systems Section, Department of Health Technology, 5205Technical University of Denmark, Kongens Lyngby, Denmark
| | - Jens Hjortkjær
- Hearing Systems Section, Department of Health Technology, 5205Technical University of Denmark, Kongens Lyngby, Denmark.,Danish Research Centre for Magnetic Resonance, Centre for Functional and Diagnostic Imaging and Research, Copenhagen University Hospital Hvidovre, Kettegård Allé 30, DK-2650 Hvidovre, Denmark
| | - Torsten Dau
- Hearing Systems Section, Department of Health Technology, 5205Technical University of Denmark, Kongens Lyngby, Denmark
| |
Collapse
|
6
|
Riley JR, Borland MS, Tamaoki Y, Skipton SK, Engineer CT. Auditory Brainstem Responses Predict Behavioral Deficits in Rats with Varying Levels of Noise-Induced Hearing Loss. Neuroscience 2021; 477:63-75. [PMID: 34634426 DOI: 10.1016/j.neuroscience.2021.10.003] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2021] [Revised: 09/30/2021] [Accepted: 10/04/2021] [Indexed: 11/30/2022]
Abstract
Intense noise exposure is a leading cause of hearing loss, which results in degraded speech sound discrimination ability, particularly in noisy environments. The development of an animal model of speech discrimination deficits due to noise induced hearing loss (NIHL) would enable testing of potential therapies to improve speech sound processing. Rats can accurately detect and discriminate human speech sounds in the presence of quiet and background noise. Further, it is known that profound hearing loss results in functional deafness in rats. In this study, we generated rats with a range of impairments which model the large range of hearing impairments observed in patients with NIHL. One month after noise exposure, we stratified rats into three distinct deficit groups based on their auditory brainstem response (ABR) thresholds. These groups exhibited markedly different behavioral outcomes across a range of tasks. Rats with moderate hearing loss (30 dB shifts in ABR threshold) were not impaired in speech sound detection or discrimination. Rats with severe hearing loss (55 dB shifts) were impaired at discriminating speech sounds in the presence of background noise. Rats with profound hearing loss (70 dB shifts) were unable to detect and discriminate speech sounds above chance level performance. Across groups, ABR threshold accurately predicted behavioral performance on all tasks. This model of long-term impaired speech discrimination in noise, demonstrated by the severe group, mimics the most common clinical presentation of NIHL and represents a useful tool for developing and improving interventions to target restoration of hearing.
Collapse
Affiliation(s)
- Jonathan R Riley
- The University of Texas at Dallas, Texas Biomedical Device Center, 800 West Campbell Road BSB11, Richardson, TX 75080, USA; The University of Texas at Dallas, School of Behavioral and Brain Sciences, 800 West Campbell Road BSB11, Richardson, TX 75080, USA.
| | - Michael S Borland
- The University of Texas at Dallas, Texas Biomedical Device Center, 800 West Campbell Road BSB11, Richardson, TX 75080, USA; The University of Texas at Dallas, School of Behavioral and Brain Sciences, 800 West Campbell Road BSB11, Richardson, TX 75080, USA
| | - Yuko Tamaoki
- The University of Texas at Dallas, Texas Biomedical Device Center, 800 West Campbell Road BSB11, Richardson, TX 75080, USA; The University of Texas at Dallas, School of Behavioral and Brain Sciences, 800 West Campbell Road BSB11, Richardson, TX 75080, USA
| | - Samantha K Skipton
- The University of Texas at Dallas, Texas Biomedical Device Center, 800 West Campbell Road BSB11, Richardson, TX 75080, USA; The University of Texas at Dallas, School of Behavioral and Brain Sciences, 800 West Campbell Road BSB11, Richardson, TX 75080, USA
| | - Crystal T Engineer
- The University of Texas at Dallas, Texas Biomedical Device Center, 800 West Campbell Road BSB11, Richardson, TX 75080, USA; The University of Texas at Dallas, School of Behavioral and Brain Sciences, 800 West Campbell Road BSB11, Richardson, TX 75080, USA
| |
Collapse
|
7
|
Effects of Simulated and Profound Unilateral Sensorineural Hearing Loss on Recognition of Speech in Competing Speech. Ear Hear 2021; 41:411-419. [PMID: 31356386 DOI: 10.1097/aud.0000000000000764] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Unilateral hearing loss (UHL) is a condition as common as bilateral hearing loss in adults. Because of the unilaterally reduced audibility associated with UHL, binaural processing of sounds may be disrupted. As a consequence, daily tasks such as listening to speech in a background of spatially distinct competing sounds may be challenging. A growing body of subjective and objective data suggests that spatial hearing is negatively affected by UHL. However, the type and degree of UHL vary considerably in previous studies. The aim here was to determine the effect of a profound sensorineural UHL, and of a simulated UHL, on recognition of speech in competing speech, and the binaural and monaural contributions to spatial release from masking, in a demanding multisource listening environment. DESIGN Nine subjects (25 to 61 years) with profound sensorineural UHL [mean pure-tone average (PTA) across 0.5, 1, 2, and 4 kHz = 105 dB HL] and normal contralateral hearing (mean PTA = 7.2 dB HL) were included based on the criterion that the target and competing speech were inaudible in the ear with hearing loss. Thirteen subjects with normal hearing (19 to 60 years; mean left PTA = 4.1 dB HL; mean right PTA = 5.5 dB HL) contributed data in normal and simulated "mild-to-moderate" UHL conditions (PTA = 38.6 dB HL). The main outcome measure was the threshold for 40% correct speech recognition in colocated (0°) and spatially and symmetrically separated (±30° and ±150°) competing speech conditions. Spatial release from masking was quantified as the threshold difference between colocated and separated conditions. RESULTS Thresholds in profound UHL were higher (worse) than normal hearing in separated and colocated conditions, and comparable to simulated UHL. Monaural spatial release from masking, that is, the spatial release achieved by subjects with profound UHL, was significantly different from zero and 49% of the magnitude of the spatial release from masking achieved by subjects with normal hearing. There were subjects with profound UHL who showed negative spatial release, whereas subjects with normal hearing consistently showed positive spatial release from masking in the normal condition. The simulated UHL had a larger effect on the speech recognition threshold for separated than for colocated conditions, resulting in decreased spatial release from masking. The difference in spatial release between normal-hearing and simulated UHL conditions increased with age. CONCLUSIONS The results demonstrate that while recognition of speech in colocated and separated competing speech is impaired for profound sensorineural UHL, spatial release from masking may be possible when competing speech is symmetrically distributed around the listener. A "mild-to-moderate" simulated UHL decreases spatial release from masking compared with normal-hearing conditions and interacts with age, indicating that small amounts of residual hearing in the UHL ear may be more beneficial for separated than for colocated interferer conditions for young listeners.
Collapse
|
8
|
Versfeld NJ, Lie S, Kramer SE, Zekveld AA. Informational masking with speech-on-speech intelligibility: Pupil response and time-course of learning. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:2353. [PMID: 33940918 DOI: 10.1121/10.0003952] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/21/2020] [Accepted: 03/09/2021] [Indexed: 06/12/2023]
Abstract
Previous research has shown a learning effect on speech perception in nonstationary maskers. The present study addressed the time-course of this learning effect and the role of informational masking. To that end, speech reception thresholds (SRTs) were measured for speech in either a stationary noise masker, an interrupted noise masker, or a single-talker masker. The utterance of the single talker was either time-forward (intelligible) or time-reversed (unintelligible), and the sample of the utterance was either frozen (same utterance at each presentation) or random (different utterance at each presentation but from the same speaker). Simultaneously, the pupil dilation response was measured to assess differences in the listening effort between conditions and to track changes in the listening effort over time within each condition. The results showed a learning effect for all conditions but the stationary noise condition-that is, improvement in SRT over time while maintaining equal pupil responses. There were no significant differences in pupil responses between conditions despite large differences in the SRT. Time reversal of the frozen speech affected neither the SRT nor pupil responses.
Collapse
Affiliation(s)
- Niek J Versfeld
- Amsterdam Universitair Medisch Centrum, Vrije Universiteit Amsterdam, Otolaryngology Head and Neck Surgery, Ear and Hearing, Amsterdam Public Health Research Institute, Amsterdam, The Netherlands
| | - Sisi Lie
- Amsterdam Universitair Medisch Centrum, Vrije Universiteit Amsterdam, Otolaryngology Head and Neck Surgery, Ear and Hearing, Amsterdam Public Health Research Institute, Amsterdam, The Netherlands
| | - Sophia E Kramer
- Amsterdam Universitair Medisch Centrum, Vrije Universiteit Amsterdam, Otolaryngology Head and Neck Surgery, Ear and Hearing, Amsterdam Public Health Research Institute, Amsterdam, The Netherlands
| | - Adriana A Zekveld
- Amsterdam Universitair Medisch Centrum, Vrije Universiteit Amsterdam, Otolaryngology Head and Neck Surgery, Ear and Hearing, Amsterdam Public Health Research Institute, Amsterdam, The Netherlands
| |
Collapse
|
9
|
Lu H, McKinney MF, Zhang T, Oxenham AJ. Investigating age, hearing loss, and background noise effects on speaker-targeted head and eye movements in three-way conversations. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:1889. [PMID: 33765809 DOI: 10.1121/10.0003707] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/08/2020] [Accepted: 02/19/2021] [Indexed: 06/12/2023]
Abstract
Although beamforming algorithms for hearing aids can enhance performance, the wearer's head may not always face the target talker, potentially limiting real-world benefits. This study aimed to determine the extent to which eye tracking improves the accuracy of locating the current talker in three-way conversations and to test the hypothesis that eye movements become more likely to track the target talker with increasing background noise levels, particularly in older and/or hearing-impaired listeners. Conversations between a participant and two confederates were held around a small table in quiet and with background noise levels of 50, 60, and 70 dB sound pressure level, while the participant's eye and head movements were recorded. Ten young normal-hearing listeners were tested, along with ten older normal-hearing listeners and eight hearing-impaired listeners. Head movements generally undershot the talker's position by 10°-15°, but head and eye movements together predicted the talker's position well. Contrary to our original hypothesis, no major differences in listening behavior were observed between the groups or between noise levels, although the hearing-impaired listeners tended to spend less time looking at the current talker than the other groups, especially at the highest noise level.
Collapse
Affiliation(s)
- Hao Lu
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| | - Martin F McKinney
- Starkey Hearing Technologies, 6700 Washington Avenue South, Eden Prairie, Minnesota 55344, USA
| | - Tao Zhang
- Starkey Hearing Technologies, 6700 Washington Avenue South, Eden Prairie, Minnesota 55344, USA
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| |
Collapse
|
10
|
Speech Enhancement for Hearing Aids with Deep Learning on Environmental Noises. APPLIED SCIENCES-BASEL 2020. [DOI: 10.3390/app10176077] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Hearing aids are small electronic devices designed to improve hearing for persons with impaired hearing, using sophisticated audio signal processing algorithms and technologies. In general, the speech enhancement algorithms in hearing aids remove the environmental noise and enhance speech while still giving consideration to hearing characteristics and the environmental surroundings. In this study, a speech enhancement algorithm was proposed to improve speech quality in a hearing aid environment by applying noise reduction algorithms with deep neural network learning based on noise classification. In order to evaluate the speech enhancement in an actual hearing aid environment, ten types of noise were self-recorded and classified using convolutional neural networks. In addition, noise reduction for speech enhancement in the hearing aid were applied by deep neural networks based on the noise classification. As a result, the speech quality based on the speech enhancements removed using the deep neural networks—and associated environmental noise classification—exhibited a significant improvement over that of the conventional hearing aid algorithm. The improved speech quality was also evaluated by objective measure through the perceptual evaluation of speech quality score, the short-time objective intelligibility score, the overall quality composite measure, and the log likelihood ratio score.
Collapse
|
11
|
Park G, Lee S. Environmental Noise Classification Using Convolutional Neural Networks with Input Transform for Hearing Aids. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2020; 17:ijerph17072270. [PMID: 32230966 PMCID: PMC7178286 DOI: 10.3390/ijerph17072270] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/31/2020] [Revised: 03/24/2020] [Accepted: 03/25/2020] [Indexed: 11/16/2022]
Abstract
Hearing aids are essential for people with hearing loss, and noise estimation and classification are some of the most important technologies used in devices. This paper presents an environmental noise classification algorithm for hearing aids that uses convolutional neural networks (CNNs) and image signals transformed from sound signals. The algorithm was developed using the data of ten types of noise acquired from living environments where such noises occur. Spectrogram images transformed from sound data are used as the input of the CNNs after processing of the images by a sharpening mask and median filter. The classification results of the proposed algorithm were compared with those of other noise classification methods. A maximum correct classification accuracy of 99.25% was achieved by the proposed algorithm for a spectrogram time length of 1 s, with the correct classification accuracy decreasing with increasing spectrogram time length up to 8 s. For a spectrogram time length of 8 s and using the sharpening mask and median filter, the classification accuracy was 98.73%, which is comparable with the 98.79% achieved by the conventional method for a time length of 1 s. The proposed hearing aid noise classification algorithm thus offers less computational complexity without compromising on performance.
Collapse
|
12
|
Goldsworthy RL, Markle KL. Pediatric Hearing Loss and Speech Recognition in Quiet and in Different Types of Background Noise. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:758-767. [PMID: 30950727 PMCID: PMC9907566 DOI: 10.1044/2018_jslhr-h-17-0389] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/19/2017] [Revised: 04/23/2018] [Accepted: 10/12/2018] [Indexed: 05/27/2023]
Abstract
Purpose Speech recognition deteriorates with hearing loss, particularly in fluctuating background noise. This study examined how hearing loss affects speech recognition in different types of noise to clarify how characteristics of the noise interact with the benefits listeners receive when listening in fluctuating compared to steady-state noise. Method Speech reception thresholds were measured for a closed set of spondee words in children (ages 5-17 years) in quiet, speech-spectrum noise, 2-talker babble, and instrumental music. Twenty children with normal hearing and 43 children with hearing loss participated; children with hearing loss were subdivided into groups with cochlear implant (18 children) and hearing aid (25 children) groups. A cohort of adults with normal hearing was included for comparison. Results Hearing loss had a large effect on speech recognition for each condition, but the effect of hearing loss was largest in 2-talker babble and smallest in speech-spectrum noise. Children with normal hearing had better speech recognition in 2-talker babble than in speech-spectrum noise, whereas children with hearing loss had worse recognition in 2-talker babble than in speech-spectrum noise. Almost all subjects had better speech recognition in instrumental music compared to speech-spectrum noise, but with less of a difference observed for children with hearing loss. Conclusions Speech recognition is more sensitive to the effects of hearing loss when measured in fluctuating compared to steady-state noise. Speech recognition measured in fluctuating noise depends on an interaction of hearing loss with characteristics of the background noise; specifically, children with hearing loss were able to derive a substantial benefit for listening in fluctuating noise when measured in instrumental music compared to 2-talker babble.
Collapse
|
13
|
Toward a more comprehensive understanding of the impact of masker type and signal-to-noise ratio on the pupillary response while performing a speech-in-noise test. Hear Res 2018; 369:67-78. [DOI: 10.1016/j.heares.2018.05.006] [Citation(s) in RCA: 40] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/03/2017] [Revised: 04/07/2018] [Accepted: 05/09/2018] [Indexed: 11/19/2022]
|
14
|
Kidd G, Colburn HS. Informational Masking in Speech Recognition. SPRINGER HANDBOOK OF AUDITORY RESEARCH 2017. [DOI: 10.1007/978-3-319-51662-2_4] [Citation(s) in RCA: 52] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/04/2022]
|
15
|
Kidd G, Mason CR, Swaminathan J, Roverud E, Clayton KK, Best V. Determining the energetic and informational components of speech-on-speech masking. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 140:132. [PMID: 27475139 PMCID: PMC5392100 DOI: 10.1121/1.4954748] [Citation(s) in RCA: 73] [Impact Index Per Article: 8.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/13/2023]
Abstract
Identification of target speech was studied under masked conditions consisting of two or four independent speech maskers. In the reference conditions, the maskers were colocated with the target, the masker talkers were the same sex as the target, and the masker speech was intelligible. The comparison conditions, intended to provide release from masking, included different-sex target and masker talkers, time-reversal of the masker speech, and spatial separation of the maskers from the target. Significant release from masking was found for all comparison conditions. To determine whether these reductions in masking could be attributed to differences in energetic masking, ideal time-frequency segregation (ITFS) processing was applied so that the time-frequency units where the masker energy dominated the target energy were removed. The remaining target-dominated "glimpses" were reassembled as the stimulus. Speech reception thresholds measured using these resynthesized ITFS-processed stimuli were the same for the reference and comparison conditions supporting the conclusion that the amount of energetic masking across conditions was the same. These results indicated that the large release from masking found under all comparison conditions was due primarily to a reduction in informational masking. Furthermore, the large individual differences observed generally were correlated across the three masking release conditions.
Collapse
Affiliation(s)
- Gerald Kidd
- Department of Speech, Language and Hearing Sciences and Hearing Research Center, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| | - Christine R Mason
- Department of Speech, Language and Hearing Sciences and Hearing Research Center, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| | - Jayaganesh Swaminathan
- Department of Speech, Language and Hearing Sciences and Hearing Research Center, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| | - Elin Roverud
- Department of Speech, Language and Hearing Sciences and Hearing Research Center, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| | - Kameron K Clayton
- Department of Speech, Language and Hearing Sciences and Hearing Research Center, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| | - Virginia Best
- Department of Speech, Language and Hearing Sciences and Hearing Research Center, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| |
Collapse
|
16
|
Jalilvand H, Pourbakht A, Jalaee S. The relationship between hearing aid frequency response and acceptable noise level in patients with sensorineural hearing loss. Adv Biomed Res 2015; 4:256. [PMID: 26918238 PMCID: PMC4746943 DOI: 10.4103/2277-9175.170681] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/04/2014] [Accepted: 01/13/2015] [Indexed: 11/29/2022] Open
Abstract
Background: When fitting hearing aid as a compensatory device for an impaired cochlea in a patient with sensorineural hearing loss (HL), it is needed to the effective and efficient frequency response would be selected regarding providing the patient's perfect speech perception. There is not any research about the effects of frequency modifications on speech perception in patients with HL regarding the cochlear desensitization. The effect (s) of modifications in frequency response of hearing aid amplification on the results of acceptable noise level (ANL) test is the main aim of this study. Materials and Methods: The amounts of ANL in two conditions of linear amplification (high frequency emphasis [HFE] and mid frequency emphasis [MFE]) were measured. Thirty-two male subjects who participated in this study had the moderate to severe sensorineural HL. Results: There was not any significant difference between ANL in linear amplification of hearing aid with HFE frequency response and ANL in linear amplification of hearing aid with MFE frequency response. Conclusion: The gain modification of frequency response not only does not affect the patient's performance of speech intelligibility in ANL test. This indicates that we need to note to the cochlear desensitization phenomenon when fitting hearing aid as a compensatory device for an impaired cochlea in a patient. The cochlear desensitization has not been considered properly in hearing aid fitting formula which is needed to be explored more about the bio-mechanisms of impaired cochlea.
Collapse
Affiliation(s)
- Hamid Jalilvand
- Department of Audiology, School of Rehabilitation Sciences, Iran University of Medical Sciences, Tehran, Iran
| | - Akram Pourbakht
- Department of Audiology, School of Rehabilitation Sciences, Iran University of Medical Sciences, Tehran, Iran
| | - Shohreh Jalaee
- Department of Statistics, School of Rehabilitation, Tehran University of Medical Sciences, Tehran, Iran
| |
Collapse
|
17
|
Larsson M, Ekström SR, Ranjbar P. Effects of sounds of locomotion on speech perception. Noise Health 2015; 17:227-32. [PMID: 26168953 PMCID: PMC4900485 DOI: 10.4103/1463-1741.160711] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022] Open
Abstract
Human locomotion typically creates noise, a possible consequence of which is the masking of sound signals originating in the surroundings. When walking side by side, people often subconsciously synchronize their steps. The neurophysiological and evolutionary background of this behavior is unclear. The present study investigated the potential of sound created by walking to mask perception of speech and compared the masking produced by walking in step with that produced by unsynchronized walking. The masking sound (footsteps on gravel) and the target sound (speech) were presented through the same speaker to 15 normal-hearing subjects. The original recorded walking sound was modified to mimic the sound of two individuals walking in pace or walking out of synchrony. The participants were instructed to adjust the sound level of the target sound until they could just comprehend the speech signal (“just follow conversation” or JFC level) when presented simultaneously with synchronized or unsynchronized walking sound at 40 dBA, 50 dBA, 60 dBA, or 70 dBA. Synchronized walking sounds produced slightly less masking of speech than did unsynchronized sound. The median JFC threshold in the synchronized condition was 38.5 dBA, while the corresponding value for the unsynchronized condition was 41.2 dBA. Combined results at all sound pressure levels showed an improvement in the signal-to-noise ratio (SNR) for synchronized footsteps; the median difference was 2.7 dB and the mean difference was 1.2 dB [P < 0.001, repeated-measures analysis of variance (RM-ANOVA)]. The difference was significant for masker levels of 50 dBA and 60 dBA, but not for 40 dBA or 70 dBA. This study provides evidence that synchronized walking may reduce the masking potential of footsteps.
Collapse
Affiliation(s)
- Matz Larsson
- The Cardiology--Lung Clinic; School of Health and Medical Sciences, Örebro University; Institute of Environmental Medicine, Karolinska Institutet, Örebro, Stockholm, Sweden
| | | | | |
Collapse
|
18
|
Keidser G, Best V, Freeston K, Boyce A. Cognitive spare capacity: evaluation data and its association with comprehension of dynamic conversations. Front Psychol 2015; 6:597. [PMID: 25999904 PMCID: PMC4422016 DOI: 10.3389/fpsyg.2015.00597] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2015] [Accepted: 04/22/2015] [Indexed: 11/29/2022] Open
Abstract
It is well-established that communication involves the working memory system, which becomes increasingly engaged in understanding speech as the input signal degrades. The more resources allocated to recovering a degraded input signal, the fewer resources, referred to as cognitive spare capacity (CSC), remain for higher-level processing of speech. Using simulated natural listening environments, the aims of this paper were to (1) evaluate an English version of a recently introduced auditory test to measure CSC that targets the updating process of the executive function, (2) investigate if the test predicts speech comprehension better than the reading span test (RST) commonly used to measure working memory capacity, and (3) determine if the test is sensitive to increasing the number of attended locations during listening. In Experiment I, the CSC test was presented using a male and a female talker, in quiet and in spatially separated babble- and cafeteria-noises, in an audio-only and in an audio-visual mode. Data collected on 21 listeners with normal and impaired hearing confirmed that the English version of the CSC test is sensitive to population group, noise condition, and clarity of speech, but not presentation modality. In Experiment II, performance by 27 normal-hearing listeners on a novel speech comprehension test presented in noise was significantly associated with working memory capacity, but not with CSC. Moreover, this group showed no significant difference in CSC as the number of talker locations in the test increased. There was no consistent association between the CSC test and the RST. It is recommended that future studies investigate the psychometric properties of the CSC test, and examine its sensitivity to the complexity of the listening environment in participants with both normal and impaired hearing.
Collapse
Affiliation(s)
- Gitte Keidser
- National Acoustic LaboratoriesSydney, NSW, Australia
| | - Virginia Best
- Department of Speech, Language, and Hearing Sciences, Boston UniversityBoston, MA, USA
| | | | - Alexandra Boyce
- Department of Audiology, Macquarie UniversitySydney, NSW, Australia
| |
Collapse
|
19
|
Abstract
OBJECTIVE To develop and apply a pedagogical method focusing on Empowerment, Empathy, Competence, and Counselling: the EC programme, and to present an initial evaluation. DESIGN The EC programme was gradually developed within a study circle framework and in dialogue with study circle leaders and participants (clients) with hearing impairment (HI). An evaluation was carried out with the study circle leaders. STUDY SAMPLE Seventeen upper secondary school students with HI took part in the development of the programme. Eighteen study circle leaders responded to a questionnaire. RESULTS The EC programme developed consisted of films, CD, and DVD productions to increase insight into one's own hearing ability, to demonstrate for others what HI means, strategies to evaluate situations, and help to act constructively in social situations. The study circle leaders found most of the course material appropriate and easy to use, as a whole or in parts. The leaders' evaluations indicated that the clients had increased their knowledge about how the HI affected themselves and others. The clients had improved their self-confidence and their empathic view of others. CONCLUSION The EC programme can be used in its entirety or in part. Participation may lead to increased empowerment, empathy, competence and counselling ability.
Collapse
Affiliation(s)
- Erik Borg
- Audiological Research Centre, Örebro University Hospital , Örebro , Sweden
| | | |
Collapse
|
20
|
Lidestam B, Holgersson J, Moradi S. Comparison of informational vs. energetic masking effects on speechreading performance. Front Psychol 2014; 5:639. [PMID: 25009520 PMCID: PMC4068195 DOI: 10.3389/fpsyg.2014.00639] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2014] [Accepted: 06/04/2014] [Indexed: 11/13/2022] Open
Abstract
The effects of two types of auditory distracters (steady-state noise vs. four-talker babble) on visual-only speechreading accuracy were tested against a baseline (silence) in 23 participants with above-average speechreading ability. Their task was to speechread high frequency Swedish words. They were asked to rate their own performance and effort, and report how distracting each type of auditory distracter was. Only four-talker babble impeded speechreading accuracy. This suggests competition for phonological processing, since the four-talker babble demands phonological processing, which is also required for the speechreading task. Better accuracy was associated with lower self-rated effort in silence; no other correlations were found.
Collapse
Affiliation(s)
- Björn Lidestam
- Department of Behavioral Sciences and Learning, Linköping University Linköping, Sweden
| | - Johan Holgersson
- Department of Behavioral Sciences and Learning, Linköping University Linköping, Sweden
| | - Shahram Moradi
- Linnaeus Centre HEAD, Swedish Institute for Disability Research, Department of Behavioral Sciences and Learning, Linköping University Linköping, Sweden
| |
Collapse
|
21
|
Mishra S, Lunner T, Stenfelt S, Rönnberg J, Rudner M. Seeing the talker's face supports executive processing of speech in steady state noise. Front Syst Neurosci 2013; 7:96. [PMID: 24324411 PMCID: PMC3840300 DOI: 10.3389/fnsys.2013.00096] [Citation(s) in RCA: 39] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2013] [Accepted: 11/09/2013] [Indexed: 11/21/2022] Open
Abstract
Listening to speech in noise depletes cognitive resources, affecting speech processing. The present study investigated how remaining resources or cognitive spare capacity (CSC) can be deployed by young adults with normal hearing. We administered a test of CSC (CSCT; Mishra et al., 2013) along with a battery of established cognitive tests to 20 participants with normal hearing. In the CSCT, lists of two-digit numbers were presented with and without visual cues in quiet, as well as in steady-state and speech-like noise at a high intelligibility level. In low load conditions, two numbers were recalled according to instructions inducing executive processing (updating, inhibition) and in high load conditions the participants were additionally instructed to recall one extra number, which was the always the first item in the list. In line with previous findings, results showed that CSC was sensitive to memory load and executive function but generally not related to working memory capacity (WMC). Furthermore, CSCT scores in quiet were lowered by visual cues, probably due to distraction. In steady-state noise, the presence of visual cues improved CSCT scores, probably by enabling better encoding. Contrary to our expectation, CSCT performance was disrupted more in steady-state than speech-like noise, although only without visual cues, possibly because selective attention could be used to ignore the speech-like background and provide an enriched representation of target items in working memory similar to that obtained in quiet. This interpretation is supported by a consistent association between CSCT scores and updating skills.
Collapse
Affiliation(s)
- Sushmit Mishra
- Linnaeus Centre HEAD, Swedish Institute for Disability Research, Department of Behavioural Sciences and Learning, Linköping University Linköping, Sweden
| | | | | | | | | |
Collapse
|
22
|
Rönnberg J, Lunner T, Zekveld A, Sörqvist P, Danielsson H, Lyxell B, Dahlström O, Signoret C, Stenfelt S, Pichora-Fuller MK, Rudner M. The Ease of Language Understanding (ELU) model: theoretical, empirical, and clinical advances. Front Syst Neurosci 2013; 7:31. [PMID: 23874273 PMCID: PMC3710434 DOI: 10.3389/fnsys.2013.00031] [Citation(s) in RCA: 608] [Impact Index Per Article: 50.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2013] [Accepted: 06/24/2013] [Indexed: 12/28/2022] Open
Abstract
Working memory is important for online language processing during conversation. We use it to maintain relevant information, to inhibit or ignore irrelevant information, and to attend to conversation selectively. Working memory helps us to keep track of and actively participate in conversation, including taking turns and following the gist. This paper examines the Ease of Language Understanding model (i.e., the ELU model, Rönnberg, 2003; Rönnberg et al., 2008) in light of new behavioral and neural findings concerning the role of working memory capacity (WMC) in uni-modal and bimodal language processing. The new ELU model is a meaning prediction system that depends on phonological and semantic interactions in rapid implicit and slower explicit processing mechanisms that both depend on WMC albeit in different ways. It is based on findings that address the relationship between WMC and (a) early attention processes in listening to speech, (b) signal processing in hearing aids and its effects on short-term memory, (c) inhibition of speech maskers and its effect on episodic long-term memory, (d) the effects of hearing impairment on episodic and semantic long-term memory, and finally, (e) listening effort. New predictions and clinical implications are outlined. Comparisons with other WMC and speech perception models are made.
Collapse
Affiliation(s)
- Jerker Rönnberg
- Department of Behavioural Sciences and Learning, Linköping University Linköping, Sweden ; Linnaeus Centre HEAD, Swedish Institute for Disability Research, Linköping University Linköping, Sweden
| | | | | | | | | | | | | | | | | | | | | |
Collapse
|
23
|
Seo S, Yook S, Nam KW, Han J, Kwon SY, Hong SH, Kim D, Lee S, Jang DP, Kim IY. Real Time Environmental Classification Algorithm Using Neural Network for Hearing Aids. ACTA ACUST UNITED AC 2013. [DOI: 10.9718/jber.2013.34.1.8] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
24
|
Sörqvist P, Rönnberg J. Episodic long-term memory of spoken discourse masked by speech: what is the role for working memory capacity? JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2012; 55:210-218. [PMID: 22199182 DOI: 10.1044/1092-4388(2011/10-0353)] [Citation(s) in RCA: 48] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
PURPOSE To investigate whether working memory capacity (WMC) modulates the effects of to-be-ignored speech on the memory of materials conveyed by to-be-attended speech. METHOD Two tasks (reading span, Daneman & Carpenter, 1980; Rönnberg et al., 2008; and size-comparison span, Sörqvist, Ljungberg, & Ljung, 2010) were used to measure individual differences in WMC. Episodic long-term memory of spoken discourse was measured by requesting participants to listen to stories masked either by normal speech or by a rotated version of that speech and to subsequently answer questions on the content of the stories. RESULTS Normal speech impaired performance on the episodic long-term memory test, and both WMC tasks were negatively related to this effect, indicating that individuals with high WMC are less susceptible to disruption. Moreover, further analyses revealed that size-comparison span (a task that requires resolution of semantic confusion by inhibition processes) is a stronger predictor of the effect than is reading span. CONCLUSIONS Cognitive control processes support listening in adverse conditions. In particular, inhibition processes acting to resolve semantic confusion seem to underlie the relationship between WMC and susceptibility to distraction from masking speech.
Collapse
|
25
|
Li G, Lutman ME, Wang S, Bleeck S. Relationship between speech recognition in noise and sparseness. Int J Audiol 2011; 51:75-82. [DOI: 10.3109/14992027.2011.625984] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
Affiliation(s)
- Guoping Li
- Institute of Sound and Vibration Research, University of Southampton, Southampton, UK
| | | | | | | |
Collapse
|
26
|
Shetake JA, Wolf JT, Cheung RJ, Engineer CT, Ram SK, Kilgard MP. Cortical activity patterns predict robust speech discrimination ability in noise. Eur J Neurosci 2011; 34:1823-38. [PMID: 22098331 DOI: 10.1111/j.1460-9568.2011.07887.x] [Citation(s) in RCA: 51] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
Abstract
The neural mechanisms that support speech discrimination in noisy conditions are poorly understood. In quiet conditions, spike timing information appears to be used in the discrimination of speech sounds. In this study, we evaluated the hypothesis that spike timing is also used to distinguish between speech sounds in noisy conditions that significantly degrade neural responses to speech sounds. We tested speech sound discrimination in rats and recorded primary auditory cortex (A1) responses to speech sounds in background noise of different intensities and spectral compositions. Our behavioral results indicate that rats, like humans, are able to accurately discriminate consonant sounds even in the presence of background noise that is as loud as the speech signal. Our neural recordings confirm that speech sounds evoke degraded but detectable responses in noise. Finally, we developed a novel neural classifier that mimics behavioral discrimination. The classifier discriminates between speech sounds by comparing the A1 spatiotemporal activity patterns evoked on single trials with the average spatiotemporal patterns evoked by known sounds. Unlike classifiers in most previous studies, this classifier is not provided with the stimulus onset time. Neural activity analyzed with the use of relative spike timing was well correlated with behavioral speech discrimination in quiet and in noise. Spike timing information integrated over longer intervals was required to accurately predict rat behavioral speech discrimination in noisy conditions. The similarity of neural and behavioral discrimination of speech in noise suggests that humans and rats may employ similar brain mechanisms to solve this problem.
Collapse
Affiliation(s)
- Jai A Shetake
- The University of Texas at Dallas, School of Behavioral Brain Sciences, 800 West Campbell Road, GR41 Richardson, TX 75080-3021, USA
| | | | | | | | | | | |
Collapse
|
27
|
Castillo M. Sounds of silence. AJNR Am J Neuroradiol 2010; 32:1155-7. [PMID: 21051519 DOI: 10.3174/ajnr.a2283] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
28
|
Abstract
Self-assessment of classroom assistive listening devices (ALDs) based on induction loop systems was carried out in Swedish classes for hearing-impaired students. A questionnaire was developed and completed by 25 students (bilateral hearing aid users, 10-20 years old). Responses for hearing aid microphone mode (M) and telecoil mode (T) were collected. Two attributes, audibility and awareness, were identified and assigned to either mode. Better audibility was achieved in T-mode. Students with severe hearing loss benefited more using T-mode when compared to the better hearing students, especially in more difficult listening situations. Better awareness was achieved in M-mode; students could better hear, locate and segregate sounds in the environment around them. Depending on the situation, students make different choices between audibility and awareness. Self-assessment is a promising approach for determining what combination of ALD design and function that will best benefit the students.
Collapse
Affiliation(s)
- Johan Odelius
- Department of Human Work Sciences, Luleå University of Technology, Sweden.
| | | |
Collapse
|
29
|
Borg E, Gustafsson D, Bergkvist C, Wikström C. On the problem of listening while talking. LOGOP PHONIATR VOCO 2009; 34:218-23. [PMID: 19883169 DOI: 10.3109/14015430903193228] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
Abstract
The Gauffin and Sundberg technique of assessing masking during vocalization was modified and tested on 22 normal-hearing and 20 hearing-impaired subjects. The masking effect of the vocalized [a:] on narrow-band noise pulses (250-8,000 Hz) and on test reading was studied. The results showed that the female voice was about 4 dB more efficient in masking external speech compared to the male voice and that the female voice had a high-frequency bias of masking the narrow-band noise, whereas the male voice had a low-frequency bias. Subjects with hearing impairment in the high frequencies were particularly impaired by the masking caused by their own voice. The implications for multilogue conversations and auditory rehabilitation are discussed.
Collapse
Affiliation(s)
- Erik Borg
- Audiological Research Centre (Ahlse'ns), Orebro, Sweden.
| | | | | | | |
Collapse
|
30
|
Infants' listening in multitalker environments: effect of the number of background talkers. Atten Percept Psychophys 2009; 71:822-36. [PMID: 19429961 DOI: 10.3758/app.71.4.822] [Citation(s) in RCA: 27] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Infants are often spoken to in the presence of background sounds, including speech from other talkers. In the present study, we compared 5- and 8.5-month-olds' abilities to recognize their own names in the context of three different types of background speech: that of a single talker, multitalker babble, and that of a single talker played backward. Infants recognized their names at a 10-dB signal-to-noise ratio in the multiple-voice condition but not in the single-voice (nonreversed) condition, a pattern opposite to that of typical adult performance. Infants similarly failed to recognize their names when the background talker's voice was reversed--that is, unintelligible, but with speech-like acoustic properties. These data suggest that infants may have difficulty segregating the components of different speech streams when those streams are acoustically too similar. Alternatively, infants' attention may be drawn to the time-varying acoustic properties associated with a single talker's speech, causing difficulties when a single talker is the competing sound.
Collapse
|
31
|
Nord L, Hammarberg B, Lundström E. Laryngectomee speech in noise—voice effort, speech rate and intelligibility. ACTA ACUST UNITED AC 2009. [DOI: 10.3109/14015439509099191] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
|
32
|
Hällgren M, Larsby B, Lyxell B, Arlinger S. Speech understanding in quiet and noise, with and without hearing aids. Int J Audiol 2009; 44:574-83. [PMID: 16315448 DOI: 10.1080/14992020500190011] [Citation(s) in RCA: 57] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
Abstract
Speech recognition and cognitive functions important for speech understanding were evaluated by objective measures and by scores of perceived effort, with and without hearing aids. The tests were performed in silence, and with background conditions of speech spectrum random noise and ordinary speech. One young and one elderly group of twelve hearing-impaired subjects each participated. Hearing aid use improved speech recognition in silence (7 dB) and in the condition with speech as background (2.5 dB S/N), but did not change the perceived effort scores. In the cognitive tests no hearing aid benefit was seen in objective measures, while there was an effect of hearing aid use in scores of perceived effort, subjects reported less effort. There were no age effects on hearing aid benefit. In conclusion, hearing aid use may result in reduced effort in listening tasks that is not associated with improvement in objective scores.
Collapse
Affiliation(s)
- Mathias Hällgren
- Division of Technical Audiology, Department of Neuroscience and Locomotion, Linköping University, Sweden.
| | | | | | | |
Collapse
|
33
|
Rönnberg J, Rudner M, Foo C, Lunner T. Cognition counts: A working memory system for ease of language understanding (ELU). Int J Audiol 2009; 47 Suppl 2:S99-105. [PMID: 19012117 DOI: 10.1080/14992020802301167] [Citation(s) in RCA: 322] [Impact Index Per Article: 20.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
34
|
Larsby B, Hällgren M, Lyxell B, Arlinger S. Cognitive performance and perceived effort in speech processing tasks: effects of different noise backgrounds in normal-hearing and hearing-impaired subjects Desempeño cognitivo y percepción del esfuerzo en tareas de procesamiento del lenguaje: Efectos de las diferentes condiciones de fondo en sujetos normales e hipoacúsicos. Int J Audiol 2009; 44:131-43. [PMID: 15916114 DOI: 10.1080/14992020500057244] [Citation(s) in RCA: 99] [Impact Index Per Article: 6.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Abstract
Cognitive tests of speech understanding were administered (presented as text, or in auditory or audiovisual modality) and perceived effort was rated. This was done in four background conditions: in silence, and in three types of noise (S/N=+10 dB) varying in temporal structure and meaningfulness. Four groups of 12 subjects each (young/elderly with normal hearing and young/elderly with hearing impairment) participated. The presence of noise had a negative effect on accuracy and speed of performance in the speech processing tasks, and resulted in higher scores of perceived effort, even when the stimuli were presented as text. Differences in performance between noise conditions existed. In the subjective scores, the noise with temporal variations, but without meaningful content, was the most disruptive of the three noise conditions. In the objective scores the hearing-impaired subjects showed poorer results in noise with temporal variations. The elderly subjects were more distracted by noise with temporal variations, and especially by noise with meaningful content. In noise, all subjects, particularly those with impaired hearing, were more dependent upon visual cues than in the quiet condition.
Collapse
Affiliation(s)
- Birgitta Larsby
- Department of Neuroscience and Locomotion, Division of Technical Audiology, University Hospital, S-581 85 Linkoping, Sweden.
| | | | | | | |
Collapse
|
35
|
Wagener KC, Brand T. Sentence intelligibility in noise for listeners with normal hearing and hearing impairment: Influence of measurement procedure and masking parameters La inteligibilidad de frases en silencio para sujetos con audición normal y con hipoacusia: la influencia del procedimiento de medición y de los parámetros de enmascaramiento. Int J Audiol 2009; 44:144-56. [PMID: 15916115 DOI: 10.1080/14992020500057517] [Citation(s) in RCA: 106] [Impact Index Per Article: 6.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Abstract
Speech intelligibility measurements strongly depend on several procedural parameters. In order to obtain comparable results from different test procedures, these parameters must be investigated as to which should be standardized and which could be set freely. This study investigates the influence of noise level, noise type, and presentation mode on speech reception thresholds (SRTs), and intelligibility function slopes in noise for normal-hearing and hearing-impaired subjects. The noise presentation level had no significant influence on either SRTs or slope values, provided that the presentation level exceeded hearing threshold. Two stationary, speech-shaped noises produced identical results. Speech-simulating fluctuating noise yielded about 14 dB lower SRTs for normal-hearing subjects and about 10 dB lower SRTs for 20% of the heating-impaired subjects. Of the hearing-impaired subjects, 30% did not benefit from the modulations and showed similar SRTs as for stationary noise. Using continuous noise yielded lower SRTs compared to gated noise. However, the difference between the results in continuous and gated noise was not significant for the hearing-impaired subjects. A presentation level of 65 dB SPL (normal-hearing subjects) or 80 dB SPL (hearing-impaired subjects) and an interfering noise with a spectrum similar to the mean long-term average speech spectrum (LTASS) is suggested for comparable adaptive measurement procedures. A fluctuating, speech-shaped noise is recommended to differentiate between subjects.
Collapse
Affiliation(s)
- Kirsten Carola Wagener
- Carl von Ossietzky Universität, Oldenburg, Medizinische Physik, Fakultät V/Institut für Physik, Oldenburg, Germany.
| | | |
Collapse
|
36
|
Bernstein JGW, Grant KW. Auditory and auditory-visual intelligibility of speech in fluctuating maskers for normal-hearing and hearing-impaired listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2009; 125:3358-72. [PMID: 19425676 DOI: 10.1121/1.3110132] [Citation(s) in RCA: 172] [Impact Index Per Article: 10.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/15/2023]
Abstract
Speech intelligibility for audio-alone and audiovisual (AV) sentences was estimated as a function of signal-to-noise ratio (SNR) for a female target talker presented in a stationary noise, an interfering male talker, or a speech-modulated noise background, for eight hearing-impaired (HI) and five normal-hearing (NH) listeners. At the 50% keywords-correct performance level, HI listeners showed 7-12 dB less fluctuating-masker benefit (FMB) than NH listeners, consistent with previous results. Both groups showed significantly more FMB under AV than audio-alone conditions. When compared at the same stationary-noise SNR, FMB differences between listener groups and modalities were substantially smaller, suggesting that most of the FMB differences at the 50% performance level may reflect a SNR dependence of the FMB. Still, 1-5 dB of the FMB difference between listener groups remained, indicating a possible role for reduced audibility, limited spectral or temporal resolution, or an inability to use auditory source-segregation cues, in directly limiting the ability to listen in the dips of a fluctuating masker. A modified version of the extended speech-intelligibility index that predicts a larger FMB at less favorable SNRs accounted for most of the FMB differences between listener groups and modalities. Overall, these data suggest that HI listeners retain more of an ability to listen in the dips of a fluctuating masker than previously thought. Instead, the fluctuating-masker difficulties exhibited by HI listeners may derive from the reduced FMB associated with the more favorable SNRs they require to identify a reasonable proportion of the target speech.
Collapse
Affiliation(s)
- Joshua G W Bernstein
- Army Audiology and Speech Center, Walter Reed Army Medical Center, Washington, DC 20307, USA.
| | | |
Collapse
|
37
|
Wang D, Kjems U, Pedersen MS, Boldt JB, Lunner T. Speech intelligibility in background noise with ideal binary time-frequency masking. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2009; 125:2336-47. [PMID: 19354408 DOI: 10.1121/1.3083233] [Citation(s) in RCA: 45] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/15/2023]
Abstract
Ideal binary time-frequency masking is a signal separation technique that retains mixture energy in time-frequency units where local signal-to-noise ratio exceeds a certain threshold and rejects mixture energy in other time-frequency units. Two experiments were designed to assess the effects of ideal binary masking on speech intelligibility of both normal-hearing (NH) and hearing-impaired (HI) listeners in different kinds of background interference. The results from Experiment 1 demonstrate that ideal binary masking leads to substantial reductions in speech-reception threshold for both NH and HI listeners, and the reduction is greater in a cafeteria background than in a speech-shaped noise. Furthermore, listeners with hearing loss benefit more than listeners with normal hearing, particularly for cafeteria noise, and ideal masking nearly equalizes the speech intelligibility performances of NH and HI listeners in noisy backgrounds. The results from Experiment 2 suggest that ideal binary masking in the low-frequency range yields larger intelligibility improvements than in the high-frequency range, especially for listeners with hearing loss. The findings from the two experiments have major implications for understanding speech perception in noise, computational auditory scene analysis, speech enhancement, and hearing aid design.
Collapse
Affiliation(s)
- DeLiang Wang
- Department of Computer Science & Engineering and Center for Cognitive Science, The Ohio State University, Columbus, Ohio 43210, USA
| | | | | | | | | |
Collapse
|
38
|
Rudner M, Foo C, Sundewall-Thorén E, Lunner T, Rönnberg J. Phonological mismatch and explicit cognitive processing in a sample of 102 hearing-aid users. Int J Audiol 2009; 47 Suppl 2:S91-8. [PMID: 19012116 DOI: 10.1080/14992020802304393] [Citation(s) in RCA: 47] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
Abstract
Rudner et al (2008) showed that when compression release settings are manipulated in the hearing instruments of Swedish habitual users, the resulting mismatch between the phonological form of the input speech signal and representations stored in long-term memory leads to greater engagement of explicit cognitive processing under taxing listening conditions. The mismatch effect is manifest in significant correlations between performance on cognitive tests and aided-speech-recognition performance in modulated noise and/or with fast compression release settings. This effect is predicted by the ELU model (Rönnberg et al, 2008). In order to test whether the mismatch effect can be generalized across languages, we examined two sets of aided speech recognition data collected from a Danish population where two cognitive tests, reading span and letter monitoring, had been administered. A reanalysis of all three datasets, including 102 participants, demonstrated the mismatch effect. These findings suggest that the effect of phonological mismatch, as predicted by the ELU model (Rönnberg et al, this issue) and tapped by the reading span test, is a stable phenomenon across these two Scandinavian languages.
Collapse
Affiliation(s)
- Mary Rudner
- Department of Behavioural Sciences and Learning, Linköping University, Linköping, Sweden
| | | | | | | | | |
Collapse
|
39
|
Larsby B, Hällgren M, Lyxell B. The interference of different background noises on speech processing in elderly hearing impaired subjects. Int J Audiol 2009; 47 Suppl 2:S83-90. [PMID: 19012115 DOI: 10.1080/14992020802301159] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
Abstract
The objective of the investigation is to study the interference of different background noises on speech processing. For this purpose speech recognition with the Hagerman test and a test battery with speech comprehension tasks (SVIPS) were performed in speech-weighted background noises varying in temporal structure, signal-to-noise ratio (SNR), and meaningfulness. With different test criteria and a score of perceived effort, the aim was to get a more complete picture of speech comprehension under adverse listening situations. Twenty-four subjects, aged 56-83 years, with a bilateral sensorineural hearing impairment, participated in the study. Differences in performance between the different background noises varied depending on the speech processing task, SNR, and on quantitative versus qualitative outcome measures. Age effects were seen in the Hagerman test and especially in background conditions of modulated noises (speech and reversed speech). Findings are discussed in relation to a hypothesis suggesting that masking and distraction interference from background noises on speech processing at peripheral, central auditory, and cognitive levels depends on the SNR used and the noise type and the listening task.
Collapse
Affiliation(s)
- Birgitta Larsby
- Department of Clinical and Experimental Medicine, Division of Technical Audiology, Linköping University, Linköping, Sweden.
| | | | | |
Collapse
|
40
|
The Benefit Obtained from Visually Displayed Text from an Automatic Speech Recognizer During Listening to Speech Presented in Noise. Ear Hear 2008; 29:838-52. [DOI: 10.1097/aud.0b013e31818005bd] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
41
|
Gordon-Hickey S, Moore RE. Acceptance of noise with intelligible, reversed, and unfamiliar primary discourse. Am J Audiol 2008; 17:129-35. [PMID: 18840706 DOI: 10.1044/1059-0889(2008/06-0018)] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022] Open
Abstract
PURPOSE The purpose of this study was to investigate the effects of intelligible and unintelligible primary discourse on acceptance of noise. Of particular interest was the effect of intelligibility on the most comfortable loudness (MCL) component of acceptable noise level (ANL). METHOD ANLs were measured for 30 participants using an intelligible discourse, a reversed discourse, and an unfamiliar primary discourse. For each discourse, MCL and background noise level (BNL) were found. The ANL was then computed by subtracting the mean BNL from the MCL. RESULTS The intelligibility of the primary discourse did not affect MCL. The ANL was significantly different for the intelligible versus reversed condition and the intelligible versus unfamiliar (Chinese) condition. CONCLUSION Results indicate that ANL may change as speech intelligibility changes and/or speech recognition ability decreases in adults with normal hearing.
Collapse
|
42
|
SEQUEIRA SARAHDOSSANTOS, SPECHT KARSTEN, HMLINEN HEIKKI, HUGDAHL KENNETH. The effects of different intensity levels of background noise on dichotic listening to consonant-vowel syllables. Scand J Psychol 2008; 49:305-10. [DOI: 10.1111/j.1467-9450.2008.00664.x] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
43
|
Abstract
OBJECTIVES A common complaint of many older adults is difficulty communicating in situations where they must focus on one talker in the presence of other people speaking. In listening environments containing multiple talkers, age-related changes may be caused by increased sensitivity to energetic masking, increased susceptibility to informational masking (e.g., confusion between the target voice and masking voices), and/or cognitive deficits. The purpose of the present study was to tease out these contributions to the difficulties that older adults experience in speech-on-speech masking situations. DESIGN Groups of younger, normal-hearing individuals and older adults with varying degrees of hearing sensitivity (n = 12 per group) participated in a study of sentence recognition in the presence of four types of maskers: a two-talker masker consisting of voices of the same sex as the target voice, a two-talker masker of voices of the opposite sex as the target, a signal-envelope-modulated noise derived from the two-talker complex, and a speech-shaped steady noise. Subjects also completed a voice discrimination task to determine the extent to which they were able to incidentally learn to tell apart the target voice from the same-sex masking voices and to examine whether this ability influenced speech-on-speech masking. RESULTS Results showed that older adults had significantly poorer performance in the presence of all four types of maskers, with the largest absolute difference for the same-sex masking condition. When the data were analyzed in terms of relative group differences (i.e., adjusting for absolute performance) the greatest effect was found for the opposite-sex masker. Degree of hearing loss was significantly related to performance in several listening conditions. Some older subjects demonstrated a reduced ability to discriminate between the masking and target voices; performance on this task was not related to speech recognition ability. CONCLUSIONS The overall pattern of results suggests that although amount of informational masking does not seem to differ between older and younger listeners, older adults (particularly those with hearing loss) evidence a deficit in the ability to selectively attend to a target voice, even when the masking voices are from talkers of the opposite sex. Possible explanations for these findings include problems understanding speech in the presence of a masker with temporal and spectral fluctuations and/or age-related changes in cognitive function.
Collapse
Affiliation(s)
- Karen S Helfer
- Department of Communication Disorders, University of Massachusetts, Amherst, Massachusetts 01003, USA.
| | | |
Collapse
|
44
|
Tye-Murray N, Sommers MS, Spehar B. Audiovisual Integration and Lipreading Abilities of Older Adults with Normal and Impaired Hearing. Ear Hear 2007; 28:656-68. [PMID: 17804980 DOI: 10.1097/aud.0b013e31812f7185] [Citation(s) in RCA: 90] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVE The purpose of the current study was to examine how age-related hearing impairment affects lipreading and auditory-visual integration. The working hypothesis for the investigation was that presbycusic hearing loss would increase reliance on visual speech information, resulting in better lipreading and auditory-visual integration in older persons who have hearing impairment, compared with older persons who have normal hearing. DESIGN This study compared the performance of 53 adults with normal hearing (above age 65) and 24 adults with mild-to-moderate hearing impairment (above age 65) on auditory-only (A), visual-only (V), and auditory-visual (AV) speech perception, using consonants, words, and sentences as stimuli. All testing was conducted in the presence of multi-talker background babble, set individually for each participant and each type of stimulus, to obtain approximately equivalent A performance across the groups. In addition, we compared the two groups of participants on measures of auditory enhancement, visual enhancement, and auditory-visual integration that were derived from the A, V and AV performance scores. RESULTS In general, the two groups of participants performed similarly on measures of V and AV speech perception. The one exception to this finding was that the participants with hearing impairment performed significantly better than the participants with normal hearing on V identification of words. Measures of visual enhancement, auditory enhancement, and auditory-visual integration did not differ as a function of hearing status. CONCLUSIONS Overall, the results of the current study suggest that despite increased reliance on visual speech information, older adults who have hearing impairment do not exhibit better V speech perception or auditory-visual integration than age-matched individuals who have normal hearing. These findings indicate that inclusion of V and AV speech perception measures can provide important information for designing maximally effective audiological rehabilitation strategies.
Collapse
Affiliation(s)
- Nancy Tye-Murray
- Department of Otolaryngology, Washington University, St. Louis, Missouri 63110, USA.
| | | | | |
Collapse
|
45
|
George ELJ, Festen JM, Houtgast T. Factors affecting masking release for speech in modulated noise for normal-hearing and hearing-impaired listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2006; 120:2295-311. [PMID: 17069325 DOI: 10.1121/1.2266530] [Citation(s) in RCA: 140] [Impact Index Per Article: 7.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/12/2023]
Abstract
The Speech Reception Threshold for sentences in stationary noise and in several amplitude-modulated noises was measured for 8 normal-hearing listeners, 29 sensorineural hearing-impaired listeners, and 16 normal-hearing listeners with simulated hearing loss. This approach makes it possible to determine whether the reduced benefit from masker modulations, as often observed for hearing-impaired listeners, is due to a loss of signal audibility, or due to suprathreshold deficits, such as reduced spectral and temporal resolution, which were measured in four separate psychophysical tasks. Results show that the reduced masking release can only partly be accounted for by reduced audibility, and that, when considering suprathreshold deficits, the normal effects associated with a raised presentation level should be taken into account. In this perspective, reduced spectral resolution does not appear to qualify as an actual suprathreshold deficit, while reduced temporal resolution does. Temporal resolution and age are shown to be the main factors governing masking release for speech in modulated noise, accounting for more than half of the intersubject variance. Their influence appears to be related to the processing of mainly the higher stimulus frequencies. Results based on calculations of the Speech Intelligibility Index in modulated noise confirm these conclusions.
Collapse
Affiliation(s)
- Erwin L J George
- ENT/Audiology, VU University Medical Center, P.O. Box 7057, 1007 MB Amsterdam, The Netherlands.
| | | | | |
Collapse
|
46
|
Hollich G, Newman RS, Jusczyk PW. Infants' use of synchronized visual information to separate streams of speech. Child Dev 2005; 76:598-613. [PMID: 15892781 DOI: 10.1111/j.1467-8624.2005.00866.x] [Citation(s) in RCA: 81] [Impact Index Per Article: 4.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
Abstract
In 4 studies, 7.5-month-olds used synchronized visual-auditory correlations to separate a target speech stream when a distractor passage was presented at equal loudness. Infants succeeded in a segmentation task (using the head-turn preference procedure with video familiarization) when a video of the talker's face was synchronized with the target passage (Experiment 1, N = 30). Infants did not succeed in this task when an unsynchronized (Experiment 2, N = 30) or static (Experiment 3, N = 30) face was presented during familiarization. Infants also succeeded when viewing a synchronized oscilloscope pattern (Experiment 4, N = 26), suggesting that their ability to use visual information is related to domain-general sensitivities to any synchronized auditory-visual correspondence.
Collapse
Affiliation(s)
- George Hollich
- Department of Psychological Sciences, Purdue University, West Lafayette, IN 47907, USA.
| | | | | |
Collapse
|
47
|
Rhebergen KS, Versfeld NJ. A Speech Intelligibility Index-based approach to predict the speech reception threshold for sentences in fluctuating noise for normal-hearing listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2005; 117:2181-92. [PMID: 15898659 DOI: 10.1121/1.1861713] [Citation(s) in RCA: 53] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/02/2023]
Abstract
The SII model in its present form (ANSI S3.5-1997, American National Standards Institute, New York) can accurately describe intelligibility for speech in stationary noise but fails to do so for nonstationary noise maskers. Here, an extension to the SII model is proposed with the aim to predict the speech intelligibility in both stationary and fluctuating noise. The basic principle of the present approach is that both speech and noise signal are partitioned into small time frames. Within each time frame the conventional SII is determined, yielding the speech information available to the listener at that time frame. Next, the SII values of these time frames are averaged, resulting in the SII for that particular condition. Using speech reception threshold (SRT) data from the literature, the extension to the present SII model can give a good account for SRTs in stationary noise, fluctuating speech noise, interrupted noise, and multiple-talker noise. The predictions for sinusoidally intensity modulated (SIM) noise and real speech or speech-like maskers are better than with the original SII model, but are still not accurate. For the latter type of maskers, informational masking may play a role.
Collapse
Affiliation(s)
- Koenraad S Rhebergen
- Department of Clinical and Experimental Audiology, Academic Medical Center Room D2-223, Meibergdreef 9, 1105 AZ Amsterdam, The Netherlands.
| | | |
Collapse
|
48
|
Kozou H, Kujala T, Shtyrov Y, Toppila E, Starck J, Alku P, Näätänen R. The effect of different noise types on the speech and non-speech elicited mismatch negativity. Hear Res 2005; 199:31-9. [PMID: 15574298 DOI: 10.1016/j.heares.2004.07.010] [Citation(s) in RCA: 53] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/27/2004] [Accepted: 07/09/2004] [Indexed: 11/18/2022]
Abstract
The effect of different types of real-life noise on the central auditory processing of speech and non-speech sounds was evaluated by the means of mismatch negativity and behavioral responses. Subjects (19-34 years old; 6 males, 4 females) were presented, in separate conditions, with either speech or non-speech stimuli of approximately equal complexity in five background conditions: babble noise, industrial noise, traffic noise, wide band noise, and silent condition. Whereas there were no effects of stimuli or noise on the behavioral responses, the MMN results revealed that speech and non-speech sounds are processed differently both in silent and noisy conditions. Speech processing was more affected than non-speech processing in all noise conditions. Moreover, different noise types had a differential effect on the pre-attentive discrimination, as reflected in MMN, on speech and non-speech sounds. Babble and industrial noises dramatically reduced the MMN amplitudes for both stimulus types, while traffic noise affected only speech stimuli.
Collapse
Affiliation(s)
- H Kozou
- Department of Psychology, Cognitive Brain Research Unit, University of Helsinki, P.O. Box 9, FIN-00014 Helsinki, Finland
| | | | | | | | | | | | | |
Collapse
|
49
|
Alcántara JI, Weisblatt EJL, Moore BCJ, Bolton PF. Speech-in-noise perception in high-functioning individuals with autism or Asperger's syndrome. J Child Psychol Psychiatry 2004; 45:1107-14. [PMID: 15257667 DOI: 10.1111/j.1469-7610.2004.t01-1-00303.x] [Citation(s) in RCA: 189] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Abstract
BACKGROUND High-functioning individuals with autism (HFA) or Asperger's syndrome (AS) commonly report difficulties understanding speech in situations where there is background speech or noise. The objective of this study was threefold: (1) to verify the validity of these reports; (2) to quantify the difficulties experienced; and (3) to propose possible mechanisms to explain the perceptual deficits described. METHOD Speech-in-noise perception abilities were measured using speech reception thresholds (SRTs), defined as the speech-to-noise ratio (SNR) at which approximately 50% of the speech is correctly identified. SRTs were measured for 11 individuals with HFA/AS and 9 age/IQ-matched normal-hearing control subjects, using an adaptive procedure, in a non-reverberant sound-attenuating chamber. The speech materials were standardised lists of everyday sentences spoken by a British male speaker. The background sounds were: (1) a single female talker; (2) a steady speech-shaped noise; (3) a speech-shaped noise with temporal dips; (4) a steady speech-shaped noise with regularly spaced spectral dips; and (5) a speech-shaped noise with temporal and spectral dips. RESULTS SRTs for the HFA/AS group were generally higher (worse) than those for the controls, across the five background sounds. A statistically significant difference in SRTs between the subject groups was found only for those background sounds that contained temporal or spectro-temporal dips. SRTs for the HFA/AS individuals were 2 to 3.5 dB higher than for the controls, equivalent to a substantial decrease in speech recognition. Expressed another way, the HFA/AS individuals required a higher SNR, whenever there were temporal dips in the background sound, to perform at the same level as the controls. CONCLUSIONS The results suggest that the speech-in-noise perception difficulties experienced by individuals with autism may be due, in part, to a reduced ability to integrate information from glimpses present in the temporal dips in the noise.
Collapse
Affiliation(s)
- José I Alcántara
- Department of Experimental Psychology, University of Cambridge, UK.
| | | | | | | |
Collapse
|
50
|
Stickney GS, Zeng FG, Litovsky R, Assmann P. Cochlear implant speech recognition with speech maskers. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2004; 116:1081-91. [PMID: 15376674 DOI: 10.1121/1.1772399] [Citation(s) in RCA: 255] [Impact Index Per Article: 12.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/08/2023]
Abstract
Speech recognition performance was measured in normal-hearing and cochlear-implant listeners with maskers consisting of either steady-state speech-spectrum-shaped noise or a competing sentence. Target sentences from a male talker were presented in the presence of one of three competing talkers (same male, different male, or female) or speech-spectrum-shaped noise generated from this talker at several target-to-masker ratios. For the normal-hearing listeners, target-masker combinations were processed through a noise-excited vocoder designed to simulate a cochlear implant. With unprocessed stimuli, a normal-hearing control group maintained high levels of intelligibility down to target-to-masker ratios as low as 0 dB and showed a release from masking, producing better performance with single-talker maskers than with steady-state noise. In contrast, no masking release was observed in either implant or normal-hearing subjects listening through an implant simulation. The performance of the simulation and implant groups did not improve when the single-talker masker was a different talker compared to the same talker as the target speech, as was found in the normal-hearing control. These results are interpreted as evidence for a significant role of informational masking and modulation interference in cochlear implant speech recognition with fluctuating maskers. This informational masking may originate from increased target-masker similarity when spectral resolution is reduced.
Collapse
Affiliation(s)
- Ginger S Stickney
- University of California, Department of Otolaryngology--Head and Neck Surgery, Irvine, California 92697-1275, USA.
| | | | | | | |
Collapse
|