1
|
Best V, Roverud E. Externalization of Speech When Listening With Hearing Aids. Trends Hear 2024; 28:23312165241229572. [PMID: 38347733 PMCID: PMC10865954 DOI: 10.1177/23312165241229572] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2023] [Revised: 01/08/2024] [Accepted: 01/15/2024] [Indexed: 02/15/2024] Open
Abstract
Subjective reports indicate that hearing aids can disrupt sound externalization and/or reduce the perceived distance of sounds. Here we conducted an experiment to explore this phenomenon and to quantify how frequently it occurs for different hearing-aid styles. Of particular interest were the effects of microphone position (behind the ear vs. in the ear) and dome type (closed vs. open). Participants were young adults with normal hearing or with bilateral hearing loss, who were fitted with hearing aids that allowed variations in the microphone position and the dome type. They were seated in a large sound-treated booth and presented with monosyllabic words from loudspeakers at a distance of 1.5 m. Their task was to rate the perceived externalization of each word using a rating scale that ranged from 10 (at the loudspeaker in front) to 0 (in the head) to -10 (behind the listener). On average, compared to unaided listening, hearing aids tended to reduce perceived distance and lead to more in-the-head responses. This was especially true for closed domes in combination with behind-the-ear microphones. The behavioral data along with acoustical recordings made in the ear canals of a manikin suggest that increased low-frequency ear-canal levels (with closed domes) and ambiguous spatial cues (with behind-the-ear microphones) may both contribute to breakdowns of externalization.
Collapse
Affiliation(s)
- Virginia Best
- Department of Speech, Language and Hearing Sciences, Boston University, Boston, MA 02215, USA
| | - Elin Roverud
- Department of Speech, Language and Hearing Sciences, Boston University, Boston, MA 02215, USA
| |
Collapse
|
2
|
Nisha KV, Parmar A, Shivaiah C, Prabhu P. Differential advantages of musical backgrounds on binaural integration and interaction skills in instrumentalists, vocalists, and non-musicians. J Otol 2023; 18:185-192. [PMID: 37877069 PMCID: PMC10593578 DOI: 10.1016/j.joto.2023.06.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2023] [Revised: 06/22/2023] [Accepted: 06/26/2023] [Indexed: 10/26/2023] Open
Abstract
Background Musical perception requires a host of skills. Instrumental musicians place greater emphasis on motor coordination, whereas vocal musicians rehearse vocal sounds. The study explored the differential advantages of musical background on binaural integration and interaction in musicians (instrumentalists, vocalists) and compared them with age-matched non-musicians. Methods Eight six participants aged 20-40 y with normal hearing sensitivity were subjected to binaural tests using a standard group comparison research design. The participants were segregated into three groups - Group 1 included instrumentalists (n = 26, mean age: 17.73 ± 2.83 y), while Group 2 and Group 3 consisted of vocalists (n = 30, mean age: 19.30 ± 2.47 y) and non-musicians (n = 30, mean age: 18.20 ± 3.02 y) respectively. The binaural processes namely integration (Dichotic syllable test, DST; and virtual acoustic space identification - VASI) and interaction (Interaural difference thresholds for time and level: ITD & ILD), were administered on all the participants. Results Statistical analyses showed the main effect of musicianship. Bonferroni pair-wise test revealed that the musicians (instrumentalists and vocalists) outperformed (p < 0.05) non-musicians in all the tests. The differential advantage of the musical background was seen on the binaural integration test with instrumentalists performing better in the VASI test compared to vocalists, and vice-versa for DST. No difference was observed in interaction tasks (ITD & ILD) between vocalists and instrumentalists (p > 0.05). Conclusion Musical background-induced differential advantages can be reasonably noted in the binaural skills of instrumentalists and vocalists (compared to non-musicians).
Collapse
Affiliation(s)
- Kavassery Venkateswaran Nisha
- Department of Audiology, All India Institute of Speech and Hearing (AIISH), Naimisham Campus, Manasagangothri, Mysore, 570006, India
| | - Adya Parmar
- Department of Audiology, All India Institute of Speech and Hearing (AIISH), Naimisham Campus, Manasagangothri, Mysore, 570006, India
| | - Chandana Shivaiah
- Department of Audiology, All India Institute of Speech and Hearing (AIISH), Naimisham Campus, Manasagangothri, Mysore, 570006, India
| | - Prashanth Prabhu
- Department of Audiology, All India Institute of Speech and Hearing (AIISH), Naimisham Campus, Manasagangothri, Mysore, 570006, India
| |
Collapse
|
3
|
Nisha KV, Uppunda AK, Konadath S. Effects of Maturation and Chronological Aging on Auditory Spatial Processing: A Cross-Sectional Study Across Life Span. Am J Audiol 2023; 32:119-134. [PMID: 36548963 DOI: 10.1044/2022_aja-22-00113] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/24/2022] Open
Abstract
OBJECTIVE The primary aim of the research was to document spatial acuity changes across the life span using a battery of psychoacoustical and perceptual tests. The secondary aim was to identify the optimal metric for measuring spatial processing changes across the life span (ages 10-70 years). DESIGN AND STUDY SAMPLE A cross-sectional study comprising 115 participants with clinically normal hearing was conducted. Purposive sampling was adopted to recruit participants in the study, who were divided into six groups based on their chronological age. METHOD Temporal, intensity, spectral, and composite correlates of spatial acuity were assessed using psychoacoustic measures and perceptual questionnaires. The temporal (interaural time difference [ITD]) and intensity correlates (interaural level difference [ILD]) of spatial perception were obtained using a MATLAB (v 2020a), whereas the composite correlate (virtual auditory space identification scores [VASIs]) and perceptual ratings of spatial processing were measured using Paradigm software and speech spatial and qualities in Kannada (SSQ-K). RESULTS Results across all tests (multivariate analyses variance: 6 age groups × 4 tests, followed by post hoc tests) consistently demonstrate poor ITD and ILD thresholds and overall lower spatial accuracy (VASI, SSQ-K) with increasing age. Discriminant function analyses (DFAs) revealed that VASI had a higher predictive power in capturing age-related changes in auditory spatial processing. The group segregation on spatial performance in DFA became evident after 50 years. CONCLUSION This study provides evidence of gradual change in all three correlates of spatial processing, with statistically demonstrable deficits appearing from fourth decade of life on VASI and fifth decade of life on binaural processing.
Collapse
Affiliation(s)
| | - Ajith Kumar Uppunda
- Department of Audiology, All India Institute of Speech and Hearing, Manasagangothri, Mysuru
| | - Sreeraj Konadath
- Department of Audiology, All India Institute of Speech and Hearing, Manasagangothri, Mysuru
| |
Collapse
|
4
|
Nisha KV, Uppunda AK, Kumar RT. Spatial rehabilitation using virtual auditory space training paradigm in individuals with sensorineural hearing impairment. Front Neurosci 2023; 16:1080398. [PMID: 36733923 PMCID: PMC9887142 DOI: 10.3389/fnins.2022.1080398] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2022] [Accepted: 12/20/2022] [Indexed: 01/18/2023] Open
Abstract
Purpose The present study aimed to quantify the effects of spatial training using virtual sources on a battery of spatial acuity measures in listeners with sensorineural hearing impairment (SNHI). Methods An intervention-based time-series comparison design involving 82 participants divided into three groups was adopted. Group I (n = 27, SNHI-spatially trained) and group II (n = 25, SNHI-untrained) consisted of SNHI listeners, while group III (n = 30) had listeners with normal hearing (NH). The study was conducted in three phases. In the pre-training phase, all the participants underwent a comprehensive assessment of their spatial processing abilities using a battery of tests including spatial acuity in free-field and closed-field scenarios, tests for binaural processing abilities (interaural time threshold [ITD] and level difference threshold [ILD]), and subjective ratings. While spatial acuity in the free field was assessed using a loudspeaker-based localization test, the closed-field source identification test was performed using virtual stimuli delivered through headphones. The ITD and ILD thresholds were obtained using a MATLAB psychoacoustic toolbox, while the participant ratings on the spatial subsection of speech, spatial, and qualities questionnaire in Kannada were used for the subjective ratings. Group I listeners underwent virtual auditory spatial training (VAST), following pre-evaluation assessments. All tests were re-administered on the group I listeners halfway through training (mid-training evaluation phase) and after training completion (post-training evaluation phase), whereas group II underwent these tests without any training at the same time intervals. Results and discussion Statistical analysis showed the main effect of groups in all tests at the pre-training evaluation phase, with post hoc comparisons that revealed group equivalency in spatial performance of both SNHI groups (groups I and II). The effect of VAST in group I was evident on all the tests, with the localization test showing the highest predictive power for capturing VAST-related changes on Fischer discriminant analysis (FDA). In contrast, group II demonstrated no changes in spatial acuity across timelines of measurements. FDA revealed increased errors in the categorization of NH as SNHI-trained at post-training evaluation compared to pre-training evaluation, as the spatial performance of the latter improved with VAST in the post-training phase. Conclusion The study demonstrated positive outcomes of spatial training using VAST in listeners with SNHI. The utility of this training program can be extended to other clinical population with spatial auditory processing deficits such as auditory neuropathy spectrum disorder, cochlear implants, central auditory processing disorders etc.
Collapse
|
5
|
Snapp HA, Millet B, Schaefer-Solle N, Rajguru SM, Ausili SA. The effects of hearing protection devices on spatial awareness in complex listening environments. PLoS One 2023; 18:e0280240. [PMID: 36634110 PMCID: PMC9836314 DOI: 10.1371/journal.pone.0280240] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2022] [Accepted: 12/26/2022] [Indexed: 01/13/2023] Open
Abstract
Hearing protection devices (HPDs) remain the first line of defense against hazardous noise exposure and noise-induced hearing loss (NIHL). Despite the increased awareness of NIHL as a major occupational health hazard, implementation of effective hearing protection interventions remains challenging in at-risk occupational groups including those in public safety that provide fire, emergency medical, or law enforcement services. A reduction of situational awareness has been reported as a primary barrier to including HPDs as routine personal protective equipment. This study examined the effects of hearing protection and simulated NIHL on spatial awareness in ten normal hearing subjects. In a sound-attenuating booth and using a head-orientation tracker, speech intelligibility and localization accuracy were collected from these subjects under multiple listening conditions. Results demonstrate that the use of HPDs disrupts spatial hearing as expected, specifically localization performance and monitoring of speech signals. There was a significant interaction between hemifield and signal-to-noise ratio (SNR), with speech intelligibility significantly affected when signals were presented from behind at reduced SNR. Results also suggest greater spatial hearing disruption using over-the-ear HPDs when compared to the removal of high frequency cues typically associated with NIHL through low-pass filtering. These results are consistent with reduced situational awareness as a self-reported barrier to routine HPD use, and was evidenced in our study by decreased ability to make accurate decisions about source location in a controlled dual-task localization experiment.
Collapse
Affiliation(s)
- Hillary A. Snapp
- Department of Otolaryngology, University of Miami, Miami, FL, United States of America
- * E-mail:
| | - Barbara Millet
- Department of Interactive Media, University of Miami, Miami, FL, United States of America
| | | | - Suhrud M. Rajguru
- Department of Biomedical Engineering, University of Miami, Miami, FL, United States of America
| | - Sebastian A. Ausili
- Department of Otolaryngology, University of Miami, Miami, FL, United States of America
| |
Collapse
|
6
|
Gessa E, Giovanelli E, Spinella D, Verdelet G, Farnè A, Frau GN, Pavani F, Valzolgher C. Spontaneous head-movements improve sound localization in aging adults with hearing loss. Front Hum Neurosci 2022; 16:1026056. [PMID: 36310849 PMCID: PMC9609159 DOI: 10.3389/fnhum.2022.1026056] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2022] [Accepted: 09/21/2022] [Indexed: 11/04/2023] Open
Abstract
Moving the head while a sound is playing improves its localization in human listeners, in children and adults, with or without hearing problems. It remains to be ascertained if this benefit can also extend to aging adults with hearing-loss, a population in which spatial hearing difficulties are often documented and intervention solutions are scant. Here we examined performance of elderly adults (61-82 years old) with symmetrical or asymmetrical age-related hearing-loss, while they localized sounds with their head fixed or free to move. Using motion-tracking in combination with free-field sound delivery in visual virtual reality, we tested participants in two auditory spatial tasks: front-back discrimination and 3D sound localization in front space. Front-back discrimination was easier for participants with symmetrical compared to asymmetrical hearing-loss, yet both groups reduced their front-back errors when head-movements were allowed. In 3D sound localization, free head-movements reduced errors in the horizontal dimension and in a composite measure that computed errors in 3D space. Errors in 3D space improved for participants with asymmetrical hearing-impairment when the head was free to move. These preliminary findings extend to aging adults with hearing-loss the literature on the advantage of head-movements on sound localization, and suggest that the disparity of auditory cues at the two ears can modulate this benefit. These results point to the possibility of taking advantage of self-regulation strategies and active behavior when promoting spatial hearing skills.
Collapse
Affiliation(s)
- Elena Gessa
- Center for Mind/Brian Sciences - CIMeC, University of Trento, Rovereto, Italy
| | - Elena Giovanelli
- Center for Mind/Brian Sciences - CIMeC, University of Trento, Rovereto, Italy
| | | | - Grégoire Verdelet
- Integrative, Multisensory, Perception, Action and Cognition Team-IMPACT, Centre de Recherche en Neuroscience de Lyon, University Lyon 1, Lyon, France
- Neuro-immersion, Centre de Recherche en Neuroscience de Lyon, Lyon, France
| | - Alessandro Farnè
- Center for Mind/Brian Sciences - CIMeC, University of Trento, Rovereto, Italy
- Integrative, Multisensory, Perception, Action and Cognition Team-IMPACT, Centre de Recherche en Neuroscience de Lyon, University Lyon 1, Lyon, France
- Neuro-immersion, Centre de Recherche en Neuroscience de Lyon, Lyon, France
| | | | - Francesco Pavani
- Center for Mind/Brian Sciences - CIMeC, University of Trento, Rovereto, Italy
- Integrative, Multisensory, Perception, Action and Cognition Team-IMPACT, Centre de Recherche en Neuroscience de Lyon, University Lyon 1, Lyon, France
| | - Chiara Valzolgher
- Center for Mind/Brian Sciences - CIMeC, University of Trento, Rovereto, Italy
- Integrative, Multisensory, Perception, Action and Cognition Team-IMPACT, Centre de Recherche en Neuroscience de Lyon, University Lyon 1, Lyon, France
| |
Collapse
|
7
|
Zheng Y, Swanson J, Koehnke J, Guan J. Sound Localization of Listeners With Normal Hearing, Impaired Hearing, Hearing Aids, Bone-Anchored Hearing Instruments, and Cochlear Implants: A Review. Am J Audiol 2022; 31:819-834. [PMID: 35917460 DOI: 10.1044/2022_aja-22-00006] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022] Open
Abstract
PURPOSE This review article reviews the contemporary studies of localization ability for different populations in different listening environments and provides possible future research directions. CONCLUSIONS The ability to accurately localize a sound source relying on three cues (interaural time difference, interaural level difference, and spectral cues) is important for communication, learning, and safety. Confounding effects including noise and reverberation, which exist in common listening environments, mask or alter localization cues and negatively affect localization performance. Hearing loss, a common public health issue, also affects localization accuracy. Although hearing devices have been developed to provide excellent audibility of speech signals, less attention has been paid to preserving and replicating crucial localization cues. Unique challenges are faced by users of various hearing devices, including hearing aids, bone-anchored hearing instruments, and cochlear implants. Hearing aids have failed to consistently improve localization performance and, in some cases, significantly impair sound localization. Bone-conduction hearing instruments show little to no benefit for sound localization performance in most cases, although some improvement is seen in binaural users. Although cochlear implants provide great hearing benefit to individuals with severe-to-profound sensorineural hearing loss, cochlear implant users have significant difficulty localizing sound, even with two implants. However, technologies in each of these areas are advancing to reduce interference with desired sound signals and preserve localization cues to help users achieve better hearing and sound localization in real-life environments.
Collapse
Affiliation(s)
- Yunfang Zheng
- Department of Communication Sciences and Disorders, Central Michigan University, Mount Pleasant, MI
| | - Jacob Swanson
- Department of Communication Sciences and Disorders, Central Michigan University, Mount Pleasant, MI
| | - Janet Koehnke
- Department of Communication Sciences and Disorders, Montclair State University, Bloomfield, NJ
| | - Jianwei Guan
- Department of Communication Sciences and Disorders, Central Michigan University, Mount Pleasant, MI
| |
Collapse
|
8
|
|
9
|
Yost WA, Pastore MT, Dorman MF. Sound source localization is a multisystem process. ACOUSTICAL SCIENCE AND TECHNOLOGY 2020; 41:113-120. [PMID: 34305431 PMCID: PMC8297655 DOI: 10.1250/ast.41.113] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
A review of data published or presented by the authors from two populations of subjects (normal hearing listeners and patients fit with cochlear implants, CIs) involving research on sound source localization when listeners move is provided. The overall theme of the review is that sound source localization requires an integration of auditory-spatial and head-position cues and is, therefore, a multisystem process. Research with normal hearing listeners includes that related to the Wallach Azimuth Illusion, and additional aspects of sound source localization perception when listeners and sound sources rotate. Research with CI patients involves investigations of sound source localization performance by patients fit with a single CI, bilateral CIs, a CI and a hearing aid (bimodal patients), and single-sided deaf patients with one normal functioning ear and the other ear fit with a CI. Past research involving CI patients who were stationary and more recent data based on CI patients' use of head rotation to localize sound sources is summarized.
Collapse
Affiliation(s)
- William A. Yost
- Spatial Hearing Laboratory, Speech and Hearing Science, Arizona State University, PO Box 870102, Tempe, Arizona, 85287, USA
| | - M. Torben Pastore
- Spatial Hearing Laboratory, Speech and Hearing Science, Arizona State University, PO Box 870102, Tempe, Arizona, 85287, USA
| | - Michael F. Dorman
- Cochlear Implant Laboratory, Speech and Hearing Science, Arizona State University, PO Box 870102, Tempe, Arizona, 85287, USA
| |
Collapse
|
10
|
Yost WA, Pastore MT, Pulling KR. Sound-source localization as a multisystem process: The Wallach azimuth illusion. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:382. [PMID: 31370595 PMCID: PMC6656578 DOI: 10.1121/1.5116003] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/09/2018] [Revised: 06/13/2019] [Accepted: 06/17/2019] [Indexed: 06/10/2023]
Abstract
Wallach [J. Exp. Psychol. 27, 339-368 (1940)] described a "2-1" rotation scenario in which a sound source rotates on an azimuth circle around a rotating listener at twice the listener's rate of rotation. In this scenario, listeners often perceive an illusionary stationary sound source, even though the actual sound source is rotating. This Wallach Azimuth Illusion (WAI) was studied to explore Wallach's description of sound-source localization as a required interaction of binaural and head-position cues (i.e., sound-source localization is a multisystem process). The WAI requires front-back reversed sound-source localization. To extend and consolidate the current understanding of the WAI, listeners and sound sources were rotated over large distances and long time periods, which had not been done before. The data demonstrate a strong correlation between measures of the predicted WAI locations and front-back reversals (FBRs). When sounds are unlikely to elicit FBRs, sound sources are perceived veridically as rotating, but the results are listener dependent. Listeners' eyes were always open and there was little evidence under these conditions that changes in vestibular function affected the occurrence of the WAI. The results show that the WAI is a robust phenomenon that should be useful for further exploration of sound-source localization as a multisystem process.
Collapse
Affiliation(s)
- William A Yost
- Spatial Hearing Laboratory, College of Health Solutions, Arizona State University, P.O. Box 870102, Tempe, Arizona 85287, USA
| | - M Torben Pastore
- Spatial Hearing Laboratory, College of Health Solutions, Arizona State University, P.O. Box 870102, Tempe, Arizona 85287, USA
| | - Kathryn R Pulling
- Spatial Hearing Laboratory, College of Health Solutions, Arizona State University, P.O. Box 870102, Tempe, Arizona 85287, USA
| |
Collapse
|
11
|
Pastore MT, Natale S, Yost W, Dorman MF. Head Movements Allow Listeners Bilaterally Implanted With Cochlear Implants to Resolve Front-Back Confusions. Ear Hear 2019; 39:1224-1231. [PMID: 29664750 PMCID: PMC6191386 DOI: 10.1097/aud.0000000000000581] [Citation(s) in RCA: 20] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVES We report on the ability of patients fit with bilateral cochlear implants (CIs) to distinguish the front-back location of sound sources both with and without head movements. At issue was (i) whether CI patients are more prone to front-back confusions than normal hearing listeners for wideband, high-frequency stimuli; and (ii) if CI patients can utilize dynamic binaural difference cues, in tandem with their own head rotation, to resolve these front-back confusions. Front-back confusions offer a binary metric to gain insight into CI patients' ability to localize sound sources under dynamic conditions not generally measured in laboratory settings where both the sound source and patient are static. DESIGN Three-second duration Gaussian noise samples were bandpass filtered to 2 to 8 kHz and presented from one of six loudspeaker locations located 60° apart, surrounding the listener. Perceived sound source localization for seven listeners bilaterally implanted with CIs, was tested under conditions where the patient faced forward and did not move their head and under conditions where they were encouraged to moderately rotate their head. The same conditions were repeated for 5 of the patients with one implant turned off (the implant at the better ear remained on). A control group of normal hearing listeners was also tested for a baseline of comparison. RESULTS All seven CI patients demonstrated a high rate of front-back confusions when their head was stationary (41.9%). The proportion of front-back confusions was reduced to 6.7% when these patients were allowed to rotate their head within a range of approximately ± 30°. When only one implant was turned on, listeners' localization acuity suffered greatly. In these conditions, head movement or the lack thereof made little difference to listeners' performance. CONCLUSIONS Bilateral implantation can offer CI listeners the ability to track dynamic auditory spatial difference cues and compare these changes to changes in their own head position, resulting in a reduced rate of front-back confusions. This suggests that, for these patients, estimates of auditory acuity based solely on static laboratory settings may underestimate their real-world localization abilities.
Collapse
|
12
|
Ganguly A, Kucuk A, Panahi I. Real-time Smartphone application for improving spatial awareness of Hearing Assistive Devices. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2018; 2018:433-436. [PMID: 30440427 DOI: 10.1109/embc.2018.8512318] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
In this paper, we present an improved version of a Speech source Iocalization method for Direction of Arrival (DOA) estimation using only two microphones. We also present a real-time Android application on a latest smartphone to help improve the spatial awareness of hearing impaired users. Unlike earlier methods, the proposed method is computationally more efficient and fully adaptive to dynamically changing background noise. We compare the performance of proposed method with similar earlier methods and demonstrate significantly lower DOA estimation errors as well as lower computation times. People who find it difficult to localize speech sources during group conversations or social activities can use the `easy-to-use' Android application. The proposed implementation does not need any additional hardware or external microphone attachments, and can run on any dual-microphone device, such as a smartphone or tablet.
Collapse
|
13
|
Kollmeier B, Kiessling J. Functionality of hearing aids: state-of-the-art and future model-based solutions. Int J Audiol 2016; 57:S3-S28. [DOI: 10.1080/14992027.2016.1256504] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
Affiliation(s)
- Birger Kollmeier
- Medizinische Physik, Universität Oldenburg and Cluster of Excellence Hearing4all, Hörzentrum Oldenburg, HörTech gGmbH and Fraunhofer IDMT/HSA, Oldenburg, Germany and
| | - Jürgen Kiessling
- Funktionsbereich Audiologie, Justus-Liebig-Universität Gießen, Giessen, Germany
| |
Collapse
|