1
|
Bayülgen TÖ, Yaralı M. The Relationship Between Speech Discrimination in Noise and Spectral Resolution, Temporal Resolution, and Melodic Contour Identification Abilities in Cochlear Implant Users. Am J Audiol 2025:1-12. [PMID: 40408186 DOI: 10.1044/2025_aja-24-00150] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/25/2025] Open
Abstract
PURPOSE Speech-in-noise performance of cochlear implant (CI) users varies considerably, and understanding speech in a complex auditory environment remains challenging. It is still unclear which auditory skill is causing this difficulty. This study aimed to evaluate spectral resolution, temporal resolution, and melodic contour identification (MCI) skills to determine which of these skills is most closely related to speech understanding in noise and to investigate whether these three skills differ among CI users with varying performances in speech-in-noise tasks. METHOD The three-syllable word identification test (in quiet), the Turkish Matrix Test (in quiet, at fixed +8 dB SNR, and adaptively in noise) were administered to evaluate speech perception to 24 adult CI users (18-55 years old). To evaluate auditory abilities, tests involving spectral-temporally modulated ripple, gap detection, and MCI were administered. CI recipients were separated into two groups (good and low performers) based on their adaptive matrix speech reception threshold (SRT) in noise. RESULTS The abilities of temporal and spectral resolution were correlated with speech understanding at +8 dB SNR. However, the effect of spectral resolution on understanding at +8 dB SN was not significant. The skill most associated with the Matrix SRT in noise was spectral resolution. The groups differed significantly in both the speech tests and spectral resolution skills. CONCLUSIONS Varying spectral resolution skills among CI users may affect their speech perception performance in noise. The effects of spectral and temporal resolution on speech perception in noise varied across noise conditions. Thus, there may be an increased requirement for spectral resolution in more adverse listening environments.
Collapse
Affiliation(s)
- Türkan Özlem Bayülgen
- Department of Audiology, School of Health Sciences, Ankara Medipol University, Turkey
| | - Mehmet Yaralı
- Department of Audiology, Faculty of Healthy Sciences, Hacettepe University, Ankara, Turkey
| |
Collapse
|
2
|
Algazlan AJ, Aljazeeri I, Yousef M, Abdelsamad Y, Almuhawas F, Alzahrani F, Alsanosi A. Electrophysiological and Behavioral Programming Parameters in Patients with Facial Nerve Stimulation Post-Cochlear Implantation. J Int Adv Otol 2025; 21:1-6. [PMID: 40207349 PMCID: PMC12001502 DOI: 10.5152/iao.2025.241704] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/20/2024] [Accepted: 11/25/2024] [Indexed: 04/11/2025] Open
Abstract
Background The aim of this study is to compare patients who develop facial nerve stimulation (FNS) after cochlear implantation (CI) and are managed with a triphasic stimulation pulse pattern (TPP) to those who do not develop FNS regarding the behavioral mapping parameters including the most comfortable loudness level (MCL) charge and amplitude, and the threshold level (THR), as well as the electrophysiological mapping parameters including phase duration (PD) and impedance level. Methods A retrospective chart review of the patients who developed FNS at any point after device activation and were managed with TPP was carried out. Electrophysiological and behavioral mapping parameters were retrieved from the programming software database at 3 time points: the time of implantation, the time of shift to TPP, and the last programming session. A control group with no FNS was matched randomly to evaluate any difference in the mapping parameters that could be attributed to FNS. Results Sixteen ears with FNS were found to be eligible for inclusion in this study. These cases were matched to 16 ears in the control group. The programming was changed from biphasic pulse pattern (BPP) to TPP (time point -1) after a period of 22.37 ± 14.62 months. Resolution of FNS was achieved in 14 ears (87.5%) by using TPP alone. Conclusion The TPP mapping strategy, in addition to decreased phase duration, showed successful results in managing facial nerve stimulation while allowing an increase in the hearing level in the form of increased MCL amplitude.
Collapse
Affiliation(s)
- Alhassan J Algazlan
- King Abdullah Ear Specialist Center, King Saud University Medical City, Riyadh, Saudi Arabia
- Department of Otolaryngology-Head and Neck Surgery, King Saud University College of Medicine, Riyadh, Saudi Arabia
| | - Isra Aljazeeri
- King Abdullah Ear Specialist Center, King Saud University Medical City, Riyadh, Saudi Arabia
- Department of Otolaryngology-Head and Neck Surgery, King Saud University College of Medicine, Riyadh, Saudi Arabia
- Otolaryngology and Ophthalmology Specialized Aljaber Hospital, Ministry of Health, Ahsa, Saudi Arabia
| | - Medhat Yousef
- King Abdullah Ear Specialist Center, King Saud University Medical City, Riyadh, Saudi Arabia
- Department of Otolaryngology-Head and Neck Surgery, King Saud University College of Medicine, Riyadh, Saudi Arabia
- Department of ENT, Audio Vestibular Medicine Unit, Menoufia University, Menoufia, Egypt
| | | | - Fida Almuhawas
- King Abdullah Ear Specialist Center, King Saud University Medical City, Riyadh, Saudi Arabia
- Department of Otolaryngology-Head and Neck Surgery, King Saud University College of Medicine, Riyadh, Saudi Arabia
| | - Farid Alzahrani
- King Abdullah Ear Specialist Center, King Saud University Medical City, Riyadh, Saudi Arabia
- Department of Otolaryngology-Head and Neck Surgery, King Saud University College of Medicine, Riyadh, Saudi Arabia
| | - Abdulrahman Alsanosi
- King Abdullah Ear Specialist Center, King Saud University Medical City, Riyadh, Saudi Arabia
- Department of Otolaryngology-Head and Neck Surgery, King Saud University College of Medicine, Riyadh, Saudi Arabia
| |
Collapse
|
3
|
Zhang Y, Johannesen PT, Molaee-Ardekani B, Wijetillake A, Attili Chiea R, Hasan PY, Segovia-Martínez M, Lopez-Poveda EA. Comparison of Performance for Cochlear-Implant Listeners Using Audio Processing Strategies Based on Short-Time Fast Fourier Transform or Spectral Feature Extraction. Ear Hear 2025; 46:163-183. [PMID: 39680489 PMCID: PMC11637581 DOI: 10.1097/aud.0000000000001565] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2023] [Accepted: 06/27/2024] [Indexed: 09/19/2024]
Abstract
OBJECTIVES We compared sound quality and performance for a conventional cochlear-implant (CI) audio processing strategy based on short-time fast-Fourier transform (Crystalis) and an experimental strategy based on spectral feature extraction (SFE). In the latter, the more salient spectral features (acoustic events) were extracted and mapped into the CI stimulation electrodes. We hypothesized that (1) SFE would be superior to Crystalis because it can encode acoustic spectral features without the constraints imposed by the short-time fast-Fourier transform bin width, and (2) the potential benefit of SFE would be greater for CI users who have less neural cross-channel interactions. DESIGN To examine the first hypothesis, 6 users of Oticon Medical Digisonic SP CIs were tested in a double-blind design with the SFE and Crystalis strategies on various aspects: word recognition in quiet, speech-in-noise reception threshold (SRT), consonant discrimination in quiet, listening effort, melody contour identification (MCI), and subjective sound quality. Word recognition and SRTs were measured on the first and last day of testing (4 to 5 days apart) to assess potential learning and/or acclimatization effects. Other tests were run once between the first and last testing day. Listening effort was assessed by measuring pupil dilation. MCI involved identifying a five-tone contour among five possible contours. Sound quality was assessed subjectively using the multiple stimulus with hidden reference and anchor (MUSHRA) paradigm for sentences, music, and ambient sounds. To examine the second hypothesis, cross-channel interaction was assessed behaviorally using forward masking. RESULTS Word recognition was similar for the two strategies on the first day of testing and improved for both strategies on the last day of testing, with Crystalis improving significantly more. SRTs were worse with SFE than Crystalis on the first day of testing but became comparable on the last day of testing. Consonant discrimination scores were higher for Crystalis than for the SFE strategy. MCI scores and listening effort were not substantially different across strategies. Subjective sound quality scores were lower for the SFE than for the Crystalis strategy. The difference in performance with SFE and Crystalis was greater for CI users with higher channel interaction. CONCLUSIONS CI-user performance was similar with the SFE and Crystalis strategies. Longer acclimatization times may be required to reveal the full potential of the SFE strategy.
Collapse
Affiliation(s)
- Yue Zhang
- Department of Research and Technology, Oticon Medical, Vallauris, France
| | - Peter T. Johannesen
- Laboratorio de Audición Computacional y Piscoacústica, Instituto de Neurociencias de Castilla y León, Universidad de Salamanca, Salamanca, Spain
- Grupo de Audiología, Instituto de Investigación Biomédica de Salamanca, Universidad de Salamanca, Salamanca, Spain
| | | | - Aswin Wijetillake
- Department of Research and Technology, Oticon Medical, Smørum, Denmark
| | | | - Pierre-Yves Hasan
- Department of Research and Technology, Oticon Medical, Smørum, Denmark
| | | | - Enrique A. Lopez-Poveda
- Laboratorio de Audición Computacional y Piscoacústica, Instituto de Neurociencias de Castilla y León, Universidad de Salamanca, Salamanca, Spain
- Grupo de Audiología, Instituto de Investigación Biomédica de Salamanca, Universidad de Salamanca, Salamanca, Spain
- Departamento de Cirugía, Facultad de Medicina, Universidad de Salamanca, Salamanca, Spain
| |
Collapse
|
4
|
Yüksel M, Çiprut A. Reduced Channel Interaction Improves Timbre Recognition Under Vocoder Simulation of Cochlear Implant Processing. Otol Neurotol 2024; 45:e297-e306. [PMID: 38437807 DOI: 10.1097/mao.0000000000004151] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/06/2024]
Abstract
OBJECTIVE This study aimed to investigate the influence of the number of channels and channel interaction on timbre perception in cochlear implant (CI) processing. By utilizing vocoder simulations of CI processing, the effects of different numbers of channels and channel interaction were examined to assess their impact on timbre perception, an essential aspect of music and auditory performance. STUDY DESIGN, SETTING, AND PATIENTS Fourteen CI recipients, with at least 1 year of CI device use, and two groups (N = 16 and N = 19) of normal hearing (NH) participants completed a timbre recognition (TR) task. NH participants were divided into two groups, with each group being tested on different aspects of the study. The first group underwent testing with varying numbers of channels (8, 12, 16, and 20) to determine an ideal number that closely reflected the TR performance of CI recipients. Subsequently, the second group of NH participants participated in the assessment of channel interaction, utilizing the identified ideal number of 20 channels, with three conditions: low interaction (54 dB/octave), medium interaction (24 dB/octave), and high interaction (12 dB/octave). Statistical analyses, including repeated-measures analysis of variance and pairwise comparisons, were conducted to examine the effects. RESULTS The number of channels did not demonstrate a statistically significant effect on TR in NH participants ( p > 0.05). However, it was observed that the condition with 20 channels closely resembled the TR performance of CI recipients. In contrast, channel interaction exhibited a significant effect ( p < 0.001) on TR. Both the low interaction (54 dB/octave) and high interaction (12 dB/octave) conditions differed significantly from the actual CI recipients' performance. CONCLUSION Timbre perception, a complex ability reliant on highly detailed spectral resolution, was not significantly influenced by the number of channels. However, channel interaction emerged as a significant factor affecting timbre perception. The differences observed under different channel interaction conditions suggest potential mechanisms, including reduced spectro-temporal resolution and degraded spectral cues. These findings highlight the importance of considering channel interaction and optimizing CI processing strategies to enhance music perception and overall auditory performance for CI recipients.
Collapse
Affiliation(s)
- Mustafa Yüksel
- Department of Audiology, Ankara Medipol University Faculty of Health Sciences, Ankara
| | - Ayça Çiprut
- Department of Audiology, Marmara University Faculty of Medicine, Istanbul, Turkey
| |
Collapse
|
5
|
Cychosz M, Winn MB, Goupell MJ. How to vocode: Using channel vocoders for cochlear-implant research. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 155:2407-2437. [PMID: 38568143 PMCID: PMC10994674 DOI: 10.1121/10.0025274] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/19/2023] [Revised: 02/14/2024] [Accepted: 02/23/2024] [Indexed: 04/05/2024]
Abstract
The channel vocoder has become a useful tool to understand the impact of specific forms of auditory degradation-particularly the spectral and temporal degradation that reflect cochlear-implant processing. Vocoders have many parameters that allow researchers to answer questions about cochlear-implant processing in ways that overcome some logistical complications of controlling for factors in individual cochlear implant users. However, there is such a large variety in the implementation of vocoders that the term "vocoder" is not specific enough to describe the signal processing used in these experiments. Misunderstanding vocoder parameters can result in experimental confounds or unexpected stimulus distortions. This paper highlights the signal processing parameters that should be specified when describing vocoder construction. The paper also provides guidance on how to determine vocoder parameters within perception experiments, given the experimenter's goals and research questions, to avoid common signal processing mistakes. Throughout, we will assume that experimenters are interested in vocoders with the specific goal of better understanding cochlear implants.
Collapse
Affiliation(s)
- Margaret Cychosz
- Department of Linguistics, University of California, Los Angeles, Los Angeles, California 90095, USA
| | - Matthew B Winn
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, Minnesota 55455, USA
| | - Matthew J Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park, College Park, Maryland 20742, USA
| |
Collapse
|
6
|
Ashihara T, Furukawa S, Kashino M. Estimating Pitch Information From Simulated Cochlear Implant Signals With Deep Neural Networks. Trends Hear 2024; 28:23312165241298606. [PMID: 39569552 DOI: 10.1177/23312165241298606] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2024] Open
Abstract
Cochlear implant (CI) users, even with substantial speech comprehension, generally have poor sensitivity to pitch information (or fundamental frequency, F0). This insensitivity is often attributed to limited spectral and temporal resolution in the CI signals. However, the pitch sensitivity markedly varies among individuals, and some users exhibit fairly good sensitivity. This indicates that the CI signal contains sufficient information about F0, and users' sensitivity is predominantly limited by other physiological conditions such as neuroplasticity or neural health. We estimated the upper limit of F0 information that a CI signal can convey by decoding F0 from simulated CI signals (multi-channel pulsatile signals) with a deep neural network model (referred to as the CI model). We varied the number of electrode channels and the pulse rate, which should respectively affect spectral and temporal resolutions of stimulus representations. The F0-estimation performance generally improved with increasing number of channels and pulse rate. For the sounds presented under quiet conditions, the model performance was at best comparable to that of a control waveform model, which received raw-waveform inputs. Under conditions in which background noise was imposed, the performance of the CI model generally degraded by a greater degree than that of the waveform model. The pulse rate had a particularly large effect on predicted performance. These observations indicate that the CI signal contains some information for predicting F0, which is particularly sufficient for targets under quiet conditions. The temporal resolution (represented as pulse rate) plays a critical role in pitch representation under noisy conditions.
Collapse
Affiliation(s)
- Takanori Ashihara
- NTT Human Informatics Laboratories, NTT Corporation, Kanagawa, Japan
- Department of Information and Communications Engineering, Tokyo Institute of Technology, Yokohama-shi, Kanagawa-ken, Japan
| | - Shigeto Furukawa
- Graduate School of Public Health, Shizuoka Graduate University of Public Health, Shizuoka, Japan
- Speech-Language-Hearing Center, Shizuoka General Hospital, Shizuoka, Japan
- NTT Communication Science Laboratories, NTT Corporation, Kanagawa, Japan
| | - Makio Kashino
- NTT Communication Science Laboratories, NTT Corporation, Kanagawa, Japan
| |
Collapse
|
7
|
Quimby AE, Wei K, Adewole D, Eliades S, Cullen DK, Brant JA. Signal processing and stimulation potential within the ascending auditory pathway: a review. Front Neurosci 2023; 17:1277627. [PMID: 38027521 PMCID: PMC10658786 DOI: 10.3389/fnins.2023.1277627] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2023] [Accepted: 10/12/2023] [Indexed: 12/01/2023] Open
Abstract
The human auditory system encodes sound with a high degree of temporal and spectral resolution. When hearing fails, existing neuroprosthetics such as cochlear implants may partially restore hearing through stimulation of auditory neurons at the level of the cochlea, though not without limitations inherent to electrical stimulation. Novel approaches to hearing restoration, such as optogenetics, offer the potential of improved performance. We review signal processing in the ascending auditory pathway and the current state of conventional and emerging neural stimulation strategies at various levels of the auditory system.
Collapse
Affiliation(s)
- Alexandra E. Quimby
- Department of Otolaryngology and Communication Sciences, SUNY Upstate Medical University, Syracuse, NY, United States
| | - Kimberly Wei
- Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, United States
| | - Dayo Adewole
- Corporal Michael J. Crescenz Veterans Affairs Medical Center, Philadelphia, PA, United States
| | - Steven Eliades
- Department of Head and Neck Surgery and Communication Sciences, Duke University, Durham, NC, United States
| | - D. Kacy Cullen
- Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, United States
- Corporal Michael J. Crescenz Veterans Affairs Medical Center, Philadelphia, PA, United States
| | - Jason A. Brant
- Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, United States
- Department of Otorhinolaryngology – Head and Neck Surgery, University of Pennsylvania, Philadelphia, PA, United States
| |
Collapse
|
8
|
Cychosz M, Xu K, Fu QJ. Effects of spectral smearing on speech understanding and masking release in simulated bilateral cochlear implants. PLoS One 2023; 18:e0287728. [PMID: 37917727 PMCID: PMC10621938 DOI: 10.1371/journal.pone.0287728] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2023] [Accepted: 06/11/2023] [Indexed: 11/04/2023] Open
Abstract
Differences in spectro-temporal degradation may explain some variability in cochlear implant users' speech outcomes. The present study employs vocoder simulations on listeners with typical hearing to evaluate how differences in degree of channel interaction across ears affects spatial speech recognition. Speech recognition thresholds and spatial release from masking were measured in 16 normal-hearing subjects listening to simulated bilateral cochlear implants. 16-channel sine-vocoded speech simulated limited, broad, or mixed channel interaction, in dichotic and diotic target-masker conditions, across ears. Thresholds were highest with broad channel interaction in both ears but improved when interaction decreased in one ear and again in both ears. Masking release was apparent across conditions. Results from this simulation study on listeners with typical hearing show that channel interaction may impact speech recognition more than masking release, and may have implications for the effects of channel interaction on cochlear implant users' speech recognition outcomes.
Collapse
Affiliation(s)
- Margaret Cychosz
- Department of Linguistics, University of California, Los Angeles, Los Angeles, CA, United States of America
| | - Kevin Xu
- Department of Head and Neck Surgery, David Geffen School of Medicine, University of California, Los Angeles, Los Angeles, CA, United States of America
| | - Qian-Jie Fu
- Department of Head and Neck Surgery, David Geffen School of Medicine, University of California, Los Angeles, Los Angeles, CA, United States of America
| |
Collapse
|
9
|
Herrmann DP, Kalkman RK, Frijns JHM, Bahmer A. Intra-cochlear differences in the spread of excitation between biphasic and triphasic pulse stimulation in cochlear implants: A modeling and experimental study. Hear Res 2023; 432:108752. [PMID: 37019060 DOI: 10.1016/j.heares.2023.108752] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/06/2022] [Revised: 03/21/2023] [Accepted: 03/27/2023] [Indexed: 03/30/2023]
Abstract
Triphasic pulse stimulation can prevent unpleasant facial nerve stimulation in cochlear implant users. Using electromyographic measurements on facial nerve effector muscles, previous studies have shown that biphasic and triphasic pulse stimulations produce different input-output functions. However, little is known about the intracochlear effects of triphasic stimulation and how these may contribute to the amelioration of facial nerve stimulation. The present study used a computational model of implanted human cochleae to investigate the effect of pulse shape on the intracochlear spread of excitation. Biphasic and triphasic pulse stimulations were simulated from three different cochlear implant electrode contact positions. To validate the model results, experimental spread of excitation measurements were conducted with biphasic and triphasic pulse stimulation from three different electrode contact positions in 13 cochlear implant users. The model results depict differences between biphasic and triphasic pulse stimulations depending on the position of the stimulating electrode contact. While biphasic and triphasic pulse stimulations from a medial or basal electrode contact caused similar extents of neural excitation, differences between the pulse shapes were observed when the stimulating contact was located in the cochlear apex. In contrast, the experimental results showed no difference between the biphasic and triphasic initiated spread of excitation for any of the tested contact positions. The model was also used to study responses of neurons without peripheral processes to mimic the effect of neural degeneration. For all three contact positions, simulated degeneration shifted the neural responses towards the apex. Biphasic pulse stimulation showed a stronger response with neural degeneration compared to without degeneration, while triphasic pulse stimulation showed no difference. As previous measurements have demonstrated an ameliorative effect of triphasic pulse stimulation on facial nerve stimulation from medial electrode contact positions, the results imply that a complementary effect located at the facial nerve level must be responsible for reducing facial nerve stimulation.
Collapse
Affiliation(s)
- David P Herrmann
- Department of Otorhinolaryngology, Plastic, Aesthetic and Reconstructive Head and Neck Surgery and the Comprehensive Hearing Center, University Hospital Würzburg, Josef-Schneider-Str. 11, Würzburg 97080, Germany.
| | - Randy K Kalkman
- Department of Otorhinolaryngology and Head & Neck Surgery, Leiden University Medical Centre, PO Box 9600, RC Leiden 2300, the Netherlands
| | - Johan H M Frijns
- Department of Otorhinolaryngology and Head & Neck Surgery, Leiden University Medical Centre, PO Box 9600, RC Leiden 2300, the Netherlands; Leiden Institute for Brain and Cognition, PO Box 9600, RC Leiden 2300, the Netherlands
| | - Andreas Bahmer
- Department of Otorhinolaryngology, Plastic, Aesthetic and Reconstructive Head and Neck Surgery and the Comprehensive Hearing Center, University Hospital Würzburg, Josef-Schneider-Str. 11, Würzburg 97080, Germany
| |
Collapse
|
10
|
Harding EE, Gaudrain E, Hrycyk IJ, Harris RL, Tillmann B, Maat B, Free RH, Başkent D. Musical Emotion Categorization with Vocoders of Varying Temporal and Spectral Content. Trends Hear 2023; 27:23312165221141142. [PMID: 36628512 PMCID: PMC9837297 DOI: 10.1177/23312165221141142] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/12/2023] Open
Abstract
While previous research investigating music emotion perception of cochlear implant (CI) users observed that temporal cues informing tempo largely convey emotional arousal (relaxing/stimulating), it remains unclear how other properties of the temporal content may contribute to the transmission of arousal features. Moreover, while detailed spectral information related to pitch and harmony in music - often not well perceived by CI users- reportedly conveys emotional valence (positive, negative), it remains unclear how the quality of spectral content contributes to valence perception. Therefore, the current study used vocoders to vary temporal and spectral content of music and tested music emotion categorization (joy, fear, serenity, sadness) in 23 normal-hearing participants. Vocoders were varied with two carriers (sinewave or noise; primarily modulating temporal information), and two filter orders (low or high; primarily modulating spectral information). Results indicated that emotion categorization was above-chance in vocoded excerpts but poorer than in a non-vocoded control condition. Among vocoded conditions, better temporal content (sinewave carriers) improved emotion categorization with a large effect while better spectral content (high filter order) improved it with a small effect. Arousal features were comparably transmitted in non-vocoded and vocoded conditions, indicating that lower temporal content successfully conveyed emotional arousal. Valence feature transmission steeply declined in vocoded conditions, revealing that valence perception was difficult for both lower and higher spectral content. The reliance on arousal information for emotion categorization of vocoded music suggests that efforts to refine temporal cues in the CI user signal may immediately benefit their music emotion perception.
Collapse
Affiliation(s)
- Eleanor E. Harding
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen,
The Netherlands,Graduate School of Medical Sciences, Research School of Behavioural
and Cognitive Neurosciences, University of Groningen, Groningen,
The Netherlands,Prins Claus Conservatoire, Hanze University of Applied Sciences, Groningen, The Netherlands,Eleanor E. Harding, Department of Otorhinolarynology, University Medical Center Groningen, Hanzeplein 1 9713 GZ, Groningen, The Netherlands.
| | - Etienne Gaudrain
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen,
The Netherlands,Lyon Neuroscience Research Center, CNRS UMR5292, Inserm U1028, Université Lyon 1, Université de Saint-Etienne, Lyon, France
| | - Imke J. Hrycyk
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen,
The Netherlands,Graduate School of Medical Sciences, Research School of Behavioural
and Cognitive Neurosciences, University of Groningen, Groningen,
The Netherlands
| | - Robert L. Harris
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen,
The Netherlands,Prins Claus Conservatoire, Hanze University of Applied Sciences, Groningen, The Netherlands
| | - Barbara Tillmann
- Lyon Neuroscience Research Center, CNRS UMR5292, Inserm U1028, Université Lyon 1, Université de Saint-Etienne, Lyon, France
| | - Bert Maat
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen,
The Netherlands,Graduate School of Medical Sciences, Research School of Behavioural
and Cognitive Neurosciences, University of Groningen, Groningen,
The Netherlands,Cochlear Implant Center Northern Netherlands, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
| | - Rolien H. Free
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen,
The Netherlands,Graduate School of Medical Sciences, Research School of Behavioural
and Cognitive Neurosciences, University of Groningen, Groningen,
The Netherlands,Cochlear Implant Center Northern Netherlands, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
| | - Deniz Başkent
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen,
The Netherlands,Graduate School of Medical Sciences, Research School of Behavioural
and Cognitive Neurosciences, University of Groningen, Groningen,
The Netherlands
| |
Collapse
|
11
|
Bissmeyer SRS, Goldsworthy RL. Combining Place and Rate of Stimulation Improves Frequency Discrimination in Cochlear Implant Users. Hear Res 2022; 424:108583. [PMID: 35930901 PMCID: PMC10849775 DOI: 10.1016/j.heares.2022.108583] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/25/2022] [Revised: 06/20/2022] [Accepted: 07/21/2022] [Indexed: 11/04/2022]
Abstract
In the auditory system, frequency is represented as tonotopic and temporal response properties of the auditory nerve. While these response properties are inextricably linked in normal hearing, cochlear implants can separately excite tonotopic location and temporal synchrony using different electrodes and stimulation rates, respectively. This separation allows for the investigation of the contributions of tonotopic and temporal cues for frequency discrimination. The present study examines frequency discrimination in adult cochlear implant users as conveyed by electrode position and stimulation rate, separately and combined. The working hypothesis is that frequency discrimination is better provided by place and rate cues combined compared to either cue alone. This hypothesis was tested in two experiments. In the first experiment, frequency discrimination needed for melodic contour identification was measured for frequencies near 100, 200, and 400 Hz using frequency allocation modeled after clinical processors. In the second experiment, frequency discrimination for pitch ranking was measured for frequencies between 100 and 1600 Hz using an experimental frequency allocation designed to provide better access to place cues. The results of both experiments indicate that frequency discrimination is better with place and rate cues combined than with either cue alone. These results clarify how signal processing for cochlear implants could better encode frequency into place and rate of electrical stimulation. Further, the results provide insight into the contributions of place and rate cues for pitch.
Collapse
Affiliation(s)
- Susan R S Bissmeyer
- Department of Biomedical Engineering, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States; Auditory Research Center, Health Research Association, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, 1640 Marengo Street Suite 326, Los Angeles, CA 90033, United States.
| | - Raymond L Goldsworthy
- Department of Biomedical Engineering, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States; Auditory Research Center, Health Research Association, Caruso Department of Otolaryngology, Keck School of Medicine, University of Southern California, 1640 Marengo Street Suite 326, Los Angeles, CA 90033, United States
| |
Collapse
|
12
|
Leterme G, Guigou C, Guenser G, Bigand E, Bozorg Grayeli A. Effect of Sound Coding Strategies on Music Perception with a Cochlear Implant. J Clin Med 2022; 11:jcm11154425. [PMID: 35956042 PMCID: PMC9369156 DOI: 10.3390/jcm11154425] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2022] [Revised: 07/15/2022] [Accepted: 07/26/2022] [Indexed: 11/21/2022] Open
Abstract
The goal of this study was to evaluate the music perception of cochlear implantees with two different sound processing strategies. Methods: Twenty-one patients with unilateral or bilateral cochlear implants (Oticon Medical®) were included. A music trial evaluated emotions (sad versus happy based on tempo and/or minor versus major modes) with three tests of increasing difficulty. This was followed by a test evaluating the perception of musical dissonances (marked out of 10). A novel sound processing strategy reducing spectral distortions (CrystalisXDP, Oticon Medical) was compared to the standard strategy (main peak interleaved sampling). Each strategy was used one week before the music trial. Results: Total music score was higher with CrystalisXDP than with the standard strategy. Nine patients (21%) categorized music above the random level (>5) on test 3 only based on mode with either of the strategies. In this group, CrystalisXDP improved the performances. For dissonance detection, 17 patients (40%) scored above random level with either of the strategies. In this group, CrystalisXDP did not improve the performances. Conclusions: CrystalisXDP, which enhances spectral cues, seemed to improve the categorization of happy versus sad music. Spectral cues could participate in musical emotions in cochlear implantees and improve the quality of musical perception.
Collapse
Affiliation(s)
- Gaëlle Leterme
- Otolaryngology, Head and Neck Surgery Department, Dijon University Hospital, 21000 Dijon, France; (G.L.); (G.G.); (A.B.G.)
- ImVia Research Laboratory, Bourgogne-Franche-Comté University, 21000 Dijon, France
| | - Caroline Guigou
- Otolaryngology, Head and Neck Surgery Department, Dijon University Hospital, 21000 Dijon, France; (G.L.); (G.G.); (A.B.G.)
- ImVia Research Laboratory, Bourgogne-Franche-Comté University, 21000 Dijon, France
- Correspondence: ; Tel.: +33-615718531
| | - Geoffrey Guenser
- Otolaryngology, Head and Neck Surgery Department, Dijon University Hospital, 21000 Dijon, France; (G.L.); (G.G.); (A.B.G.)
| | - Emmanuel Bigand
- LEAD Research Laboratory, CNRS UMR 5022, Bourgogne-Franche-Comté University, 21000 Dijon, France;
| | - Alexis Bozorg Grayeli
- Otolaryngology, Head and Neck Surgery Department, Dijon University Hospital, 21000 Dijon, France; (G.L.); (G.G.); (A.B.G.)
- ImVia Research Laboratory, Bourgogne-Franche-Comté University, 21000 Dijon, France
| |
Collapse
|
13
|
Bissmeyer SRS, Ortiz JR, Gan H, Goldsworthy RL. Computer-based musical interval training program for Cochlear implant users and listeners with no known hearing loss. Front Neurosci 2022; 16:903924. [PMID: 35968373 PMCID: PMC9363605 DOI: 10.3389/fnins.2022.903924] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2022] [Accepted: 07/11/2022] [Indexed: 11/15/2022] Open
Abstract
A musical interval is the difference in pitch between two sounds. The way that musical intervals are used in melodies relative to the tonal center of a key can strongly affect the emotion conveyed by the melody. The present study examines musical interval identification in people with no known hearing loss and in cochlear implant users. Pitch resolution varies widely among cochlear implant users with average resolution an order of magnitude worse than in normal hearing. The present study considers the effect of training on musical interval identification and tests for correlations between low-level psychophysics and higher-level musical abilities. The overarching hypothesis is that cochlear implant users are limited in their ability to identify musical intervals both by low-level access to frequency cues for pitch as well as higher-level mapping of the novel encoding of pitch that implants provide. Participants completed a 2-week, online interval identification training. The benchmark tests considered before and after interval identification training were pure tone detection thresholds, pure tone frequency discrimination, fundamental frequency discrimination, tonal and rhythm comparisons, and interval identification. The results indicate strong correlations between measures of pitch resolution with interval identification; however, only a small effect of training on interval identification was observed for the cochlear implant users. Discussion focuses on improving access to pitch cues for cochlear implant users and on improving auditory training for musical intervals.
Collapse
Affiliation(s)
- Susan Rebekah Subrahmanyam Bissmeyer
- Caruso Department of Otolaryngology, Auditory Research Center, Keck School of Medicine, University of Southern California, Los Angeles, CA, United States
- Department of Biomedical Engineering, Viterbi School of Engineering, University of Southern California, Los Angeles, CA, United States
- *Correspondence: Susan Rebekah Subrahmanyam Bissmeyer,
| | - Jacqueline Rose Ortiz
- Caruso Department of Otolaryngology, Auditory Research Center, Keck School of Medicine, University of Southern California, Los Angeles, CA, United States
| | - Helena Gan
- Caruso Department of Otolaryngology, Auditory Research Center, Keck School of Medicine, University of Southern California, Los Angeles, CA, United States
| | - Raymond Lee Goldsworthy
- Caruso Department of Otolaryngology, Auditory Research Center, Keck School of Medicine, University of Southern California, Los Angeles, CA, United States
| |
Collapse
|
14
|
Tamati TN, Sevich VA, Clausing EM, Moberly AC. Lexical Effects on the Perceived Clarity of Noise-Vocoded Speech in Younger and Older Listeners. Front Psychol 2022; 13:837644. [PMID: 35432072 PMCID: PMC9010567 DOI: 10.3389/fpsyg.2022.837644] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2021] [Accepted: 02/16/2022] [Indexed: 11/13/2022] Open
Abstract
When listening to degraded speech, such as speech delivered by a cochlear implant (CI), listeners make use of top-down linguistic knowledge to facilitate speech recognition. Lexical knowledge supports speech recognition and enhances the perceived clarity of speech. Yet, the extent to which lexical knowledge can be used to effectively compensate for degraded input may depend on the degree of degradation and the listener's age. The current study investigated lexical effects in the compensation for speech that was degraded via noise-vocoding in younger and older listeners. In an online experiment, younger and older normal-hearing (NH) listeners rated the clarity of noise-vocoded sentences on a scale from 1 ("very unclear") to 7 ("completely clear"). Lexical information was provided by matching text primes and the lexical content of the target utterance. Half of the sentences were preceded by a matching text prime, while half were preceded by a non-matching prime. Each sentence also consisted of three key words of high or low lexical frequency and neighborhood density. Sentences were processed to simulate CI hearing, using an eight-channel noise vocoder with varying filter slopes. Results showed that lexical information impacted the perceived clarity of noise-vocoded speech. Noise-vocoded speech was perceived as clearer when preceded by a matching prime, and when sentences included key words with high lexical frequency and low neighborhood density. However, the strength of the lexical effects depended on the level of degradation. Matching text primes had a greater impact for speech with poorer spectral resolution, but lexical content had a smaller impact for speech with poorer spectral resolution. Finally, lexical information appeared to benefit both younger and older listeners. Findings demonstrate that lexical knowledge can be employed by younger and older listeners in cognitive compensation during the processing of noise-vocoded speech. However, lexical content may not be as reliable when the signal is highly degraded. Clinical implications are that for adult CI users, lexical knowledge might be used to compensate for the degraded speech signal, regardless of age, but some CI users may be hindered by a relatively poor signal.
Collapse
Affiliation(s)
- Terrin N. Tamati
- Department of Otolaryngology – Head and Neck Surgery, The Ohio State University Wexner Medical Center, Columbus, OH, United States
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, Netherlands
| | - Victoria A. Sevich
- Department of Speech and Hearing Science, The Ohio State University, Columbus, OH, United States
| | - Emily M. Clausing
- Department of Otolaryngology – Head and Neck Surgery, The Ohio State University Wexner Medical Center, Columbus, OH, United States
| | - Aaron C. Moberly
- Department of Otolaryngology – Head and Neck Surgery, The Ohio State University Wexner Medical Center, Columbus, OH, United States
| |
Collapse
|
15
|
Arjmandi M, Houston D, Wang Y, Dilley L. Estimating the reduced benefit of infant-directed speech in cochlear implant-related speech processing. Neurosci Res 2021; 171:49-61. [PMID: 33484749 PMCID: PMC8289972 DOI: 10.1016/j.neures.2021.01.007] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2020] [Revised: 12/19/2020] [Accepted: 01/17/2021] [Indexed: 11/27/2022]
Abstract
Caregivers modify their speech when talking to infants, a specific type of speech known as infant-directed speech (IDS). This speaking style facilitates language learning compared to adult-directed speech (ADS) in infants with normal hearing (NH). While infants with NH and those with cochlear implants (CIs) prefer listening to IDS over ADS, it is yet unknown how CI processing may affect the acoustic distinctiveness between ADS and IDS, as well as the degree of intelligibility of these. This study analyzed speech of seven female adult talkers to model the effects of simulated CI processing on (1) acoustic distinctiveness between ADS and IDS, (2) estimates of intelligibility of caregivers' speech in ADS and IDS, and (3) individual differences in caregivers' ADS-to-IDS modification and estimated speech intelligibility. Results suggest that CI processing is substantially detrimental to the acoustic distinctiveness between ADS and IDS, as well as to the intelligibility benefit derived from ADS-to-IDS modifications. Moreover, the observed variability across individual talkers in acoustic implementation of ADS-to-IDS modification and the estimated speech intelligibility was significantly reduced due to CI processing. The findings are discussed in the context of the link between IDS and language learning in infants with CIs.
Collapse
Affiliation(s)
- Meisam Arjmandi
- Department of Communicative Sciences and Disorders, Michigan State University, 1026 Red Cedar Road, East Lansing, MI 48824, USA.
| | - Derek Houston
- Department of Otolaryngology - Head and Neck Surgery, The Ohio State University, 915 Olentangy River Road, Columbus, OH 43212, USA
| | - Yuanyuan Wang
- Department of Otolaryngology - Head and Neck Surgery, The Ohio State University, 915 Olentangy River Road, Columbus, OH 43212, USA
| | - Laura Dilley
- Department of Communicative Sciences and Disorders, Michigan State University, 1026 Red Cedar Road, East Lansing, MI 48824, USA
| |
Collapse
|
16
|
Xu K, Willis S, Gopen Q, Fu QJ. Effects of Spectral Resolution and Frequency Mismatch on Speech Understanding and Spatial Release From Masking in Simulated Bilateral Cochlear Implants. Ear Hear 2021; 41:1362-1371. [PMID: 32132377 DOI: 10.1097/aud.0000000000000865] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVES Due to interaural frequency mismatch, bilateral cochlear-implant (CI) users may be less able to take advantage of binaural cues that normal-hearing (NH) listeners use for spatial hearing, such as interaural time differences and interaural level differences. As such, bilateral CI users have difficulty segregating competing speech even when the target and competing talkers are spatially separated. The goal of this study was to evaluate the effects of spectral resolution, tonotopic mismatch (the frequency mismatch between the acoustic center frequency assigned to CI electrode within an implanted ear relative to the expected spiral ganglion characteristic frequency), and interaural mismatch (differences in the degree of tonotopic mismatch in each ear) on speech understanding and spatial release from masking (SRM) in the presence of competing talkers in NH subjects listening to bilateral vocoder simulations. DESIGN During testing, both target and masker speech were presented in five-word sentences that had the same syntax but were not necessarily meaningful. The sentences were composed of five categories in fixed order (Name, Verb, Number, Color, and Clothes), each of which had 10 items, such that multiple sentences could be generated by randomly selecting a word from each category. Speech reception thresholds (SRTs) for the target sentence presented in competing speech maskers were measured. The target speech was delivered to both ears and the two speech maskers were delivered to (1) both ears (diotic masker), or (2) different ears (dichotic masker: one delivered to the left ear and the other delivered to the right ear). Stimuli included the unprocessed speech and four 16-channel sine-vocoder simulations with different interaural mismatch (0, 1, and 2 mm). SRM was calculated as the difference between the diotic and dichotic listening conditions. RESULTS With unprocessed speech, SRTs were 0.3 and -18.0 dB for the diotic and dichotic maskers, respectively. For the spectrally degraded speech with mild tonotopic mismatch and no interaural mismatch, SRTs were 5.6 and -2.0 dB for the diotic and dichotic maskers, respectively. When the tonotopic mismatch increased in both ears, SRTs worsened to 8.9 and 2.4 dB for the diotic and dichotic maskers, respectively. When the two ears had different tonotopic mismatch (e.g., there was interaural mismatch), the performance drop in SRTs was much larger for the dichotic than for the diotic masker. The largest SRM was observed with unprocessed speech (18.3 dB). With the CI simulations, SRM was significantly reduced to 7.6 dB even with mild tonotopic mismatch but no interaural mismatch; SRM was further reduced with increasing interaural mismatch. CONCLUSIONS The results demonstrate that frequency resolution, tonotopic mismatch, and interaural mismatch have differential effects on speech understanding and SRM in simulation of bilateral CIs. Minimizing interaural mismatch may be critical to optimize binaural benefits and improve CI performance for competing speech, a typical listening environment. SRM (the difference in SRTs between diotic and dichotic maskers) may be a useful clinical tool to assess interaural frequency mismatch in bilateral CI users and to evaluate the benefits of optimization methods that minimize interaural mismatch.
Collapse
Affiliation(s)
- Kevin Xu
- Department of Head and Neck Surgery, David Geffen School of Medicine, University of California, Los Angeles, Los Angeles, California, USA
| | | | | | | |
Collapse
|
17
|
Söderqvist S, Lamminmäki S, Aarnisalo A, Hirvonen T, Sinkkonen ST, Sivonen V. Intraoperative transimpedance and spread of excitation profile correlations with a lateral-wall cochlear implant electrode array. Hear Res 2021; 405:108235. [PMID: 33901994 DOI: 10.1016/j.heares.2021.108235] [Citation(s) in RCA: 17] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/20/2020] [Revised: 03/21/2021] [Accepted: 03/25/2021] [Indexed: 01/18/2023]
Abstract
A limiting factor of cochlear implant technology is the spread of electrode-generated intracochlear electrical field (EF) leading to spread of neural excitation (SOE). In this study, we investigated the relation of the spread of the intracochlear EF, assessed via transimpedance matrix (TIM), and SOE. A total of 43 consecutive patients (ages 0.7-82 years; 31.0 ± 25.7 years, mean ± SD) implanted with a Cochlear Nucleus CI522 or CI622 cochlear implant with Slim Straight electrode array (altogether 51 ears) were included in the study. Cochlear nerve was visualized for all patients in preoperative imaging and there were no cochlear anomalies in the study sample. The stimulated electrodes were in the basal, middle, and apical parts of the electrode array (electrode numbers 6, 11, and 19, respectively). The stimulation level was 210 CL on average for the TIM measurement and always 230 CL for the SOE measurement. Approximately 90% of the individual TIM and SOE profiles correlated with each other (p < .05; r = 0.61-0.99). Also, the widths of the TIM and SOE peaks, computed at 50% of the maximum height, exhibited a weak correlation (r = 0.39, p = .007). The 50% widths of TIM and SOE were the same only in the apical part of the electrode array; in the basal part SOE was wider than TIM, and in the middle part TIM was wider than SOE (p < .01 and p = .048, respectively). Within each measurement, TIM 50% widths were different between all three parts of the electrode array, while for SOE, only the basal electrode differed from the middle electrode. Finally, the size of the cochlea and the 50% widths of TIM and SOE had the strongest correlation in the middle part of the electrode array (r = -0.63, and -0.37, respectively). Our results suggest that there is a correlation between the spread of intracochlear EF and neural SOE at least in the apical part of the electrode array used in this study, and that larger cochleae are associated with more focused TIM and SOE.
Collapse
Affiliation(s)
- Samuel Söderqvist
- Department of Otorhinolaryngology - Head and Neck Surgery, Head and Neck Center, Helsinki University Hospital and University of Helsinki, Kasarmikatu 11-13, 00130 Helsinki, Finland.
| | - Satu Lamminmäki
- Department of Otorhinolaryngology - Head and Neck Surgery, Head and Neck Center, Helsinki University Hospital and University of Helsinki, Kasarmikatu 11-13, 00130 Helsinki, Finland
| | - Antti Aarnisalo
- Department of Otorhinolaryngology - Head and Neck Surgery, Head and Neck Center, Helsinki University Hospital and University of Helsinki, Kasarmikatu 11-13, 00130 Helsinki, Finland
| | - Timo Hirvonen
- Department of Otorhinolaryngology - Head and Neck Surgery, Head and Neck Center, Helsinki University Hospital and University of Helsinki, Kasarmikatu 11-13, 00130 Helsinki, Finland
| | - Saku T Sinkkonen
- Department of Otorhinolaryngology - Head and Neck Surgery, Head and Neck Center, Helsinki University Hospital and University of Helsinki, Kasarmikatu 11-13, 00130 Helsinki, Finland
| | - Ville Sivonen
- Department of Otorhinolaryngology - Head and Neck Surgery, Head and Neck Center, Helsinki University Hospital and University of Helsinki, Kasarmikatu 11-13, 00130 Helsinki, Finland
| |
Collapse
|
18
|
Weighting of Prosodic and Lexical-Semantic Cues for Emotion Identification in Spectrally Degraded Speech and With Cochlear Implants. Ear Hear 2021; 42:1727-1740. [PMID: 34294630 PMCID: PMC8545870 DOI: 10.1097/aud.0000000000001057] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Normally-hearing (NH) listeners rely more on prosodic cues than on lexical-semantic cues for emotion perception in speech. In everyday spoken communication, the ability to decipher conflicting information between prosodic and lexical-semantic cues to emotion can be important: for example, in identifying sarcasm or irony. Speech degradation in cochlear implants (CIs) can be sufficiently overcome to identify lexical-semantic cues, but the distortion of voice pitch cues makes it particularly challenging to hear prosody with CIs. The purpose of this study was to examine changes in relative reliance on prosodic and lexical-semantic cues in NH adults listening to spectrally degraded speech and adult CI users. We hypothesized that, compared with NH counterparts, CI users would show increased reliance on lexical-semantic cues and reduced reliance on prosodic cues for emotion perception. We predicted that NH listeners would show a similar pattern when listening to CI-simulated versions of emotional speech. DESIGN Sixteen NH adults and 8 postlingually deafened adult CI users participated in the study. Sentences were created to convey five lexical-semantic emotions (angry, happy, neutral, sad, and scared), with five sentences expressing each category of emotion. Each of these 25 sentences was then recorded with the 5 (angry, happy, neutral, sad, and scared) prosodic emotions by 2 adult female talkers. The resulting stimulus set included 125 recordings (25 Sentences × 5 Prosodic Emotions) per talker, of which 25 were congruent (consistent lexical-semantic and prosodic cues to emotion) and the remaining 100 were incongruent (conflicting lexical-semantic and prosodic cues to emotion). The recordings were processed to have 3 levels of spectral degradation: full-spectrum, CI-simulated (noise-vocoded) to have 8 channels and 16 channels of spectral information, respectively. Twenty-five recordings (one sentence per lexical-semantic emotion recorded in all five prosodies) were used for a practice run in the full-spectrum condition. The remaining 100 recordings were used as test stimuli. For each talker and condition of spectral degradation, listeners indicated the emotion associated with each recording in a single-interval, five-alternative forced-choice task. The responses were scored as proportion correct, where "correct" responses corresponded to the lexical-semantic emotion. CI users heard only the full-spectrum condition. RESULTS The results showed a significant interaction between hearing status (NH, CI) and congruency in identifying the lexical-semantic emotion associated with the stimuli. This interaction was as predicted, that is, CI users showed increased reliance on lexical-semantic cues in the incongruent conditions, while NH listeners showed increased reliance on the prosodic cues in the incongruent conditions. As predicted, NH listeners showed increased reliance on lexical-semantic cues to emotion when the stimuli were spectrally degraded. CONCLUSIONS The present study confirmed previous findings of prosodic dominance for emotion perception by NH listeners in the full-spectrum condition. Further, novel findings with CI patients and NH listeners in the CI-simulated conditions showed reduced reliance on prosodic cues and increased reliance on lexical-semantic cues to emotion. These results have implications for CI listeners' ability to perceive conflicts between prosodic and lexical-semantic cues, with repercussions for their identification of sarcasm and humor. Understanding instances of sarcasm or humor can impact a person's ability to develop relationships, follow conversation, understand vocal emotion and intended message of a speaker, following jokes, and everyday communication in general.
Collapse
|
19
|
Shekar RCMC, Hansen JHL. An evaluation framework for research platforms to advance cochlear implant/hearing aid technology: A case study with CCi-MOBILE. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:229. [PMID: 33514127 PMCID: PMC7803384 DOI: 10.1121/10.0002989] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/12/2020] [Revised: 12/06/2020] [Accepted: 12/09/2020] [Indexed: 06/12/2023]
Abstract
Cochlear implants (CIs) and hearing aids (HAs) are advanced assistive hearing devices that perform sound processing to achieve acoustic to acoustic/electrical stimulation, thus enabling the prospects for hearing restoration and rehabilitation. Since commercial CIs/HAs are typically constrained by manufacturer design/production constraints, it is necessary for researchers to use research platforms (RPs) to advance algorithms and conduct investigational studies with CI/HA subjects. While previous CI/HA research platforms exist, no study has explored establishing a formal evaluation protocol for the operational safety and reliability of RPs. This study proposes a two-phase analysis and evaluation paradigm for RPs. In the acoustic phase 1 step, a signal processing acoustic space is explored in order to present a sampled set of audio input content to explore the safety of the resulting output electric/acoustic stimulation. In the parameter phase 2 step, the configurable space for realizable electrical stimulation pulses is determined, and overall stimulation reliability and safety are evaluated. The proposed protocol is applied and demonstrated using Costakis Cochlear Implant Mobile. Assessment protocol observations, results, and additional best practices for subsampling of the acoustic and parameter test spaces are discussed. The proposed analysis-evaluation protocol establishes a viable framework for assessing RP operational safety and reliability. Guidelines for adapting the proposed protocol to address variability in RP configuration due to experimental factors such as custom algorithms, stimulation techniques, and/or individualization are also considered.
Collapse
Affiliation(s)
- Ram C M C Shekar
- Cochlear Implant Processing Laboratory, Center for Robust Speech Systems (CRSS-CILab), Department of Electrical and Computer Engineering, Jonsson School of Engineering and Computer Science, University of Texas at Dallas, Richardson, Texas 75080, USA
| | - John H L Hansen
- Cochlear Implant Processing Laboratory, Center for Robust Speech Systems (CRSS-CILab), Department of Electrical and Computer Engineering, Jonsson School of Engineering and Computer Science, University of Texas at Dallas, Richardson, Texas 75080, USA
| |
Collapse
|
20
|
Performance of cochlear implant recipients fitted with triphasic pulse patterns. Eur Arch Otorhinolaryngol 2020; 278:3211-3216. [PMID: 32979117 DOI: 10.1007/s00405-020-06382-0] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2020] [Accepted: 09/15/2020] [Indexed: 10/23/2022]
Abstract
PURPOSE To assess the effect of triphasic pulse pattern stimulation strategy on the audiological performance of cochlear implant recipients with unintended facial nerve stimulation (FNS), and to compare the audiological and speech outcomes before and after switching to triphasic stimulation. METHODS A retrospective study of patients who have changed their fitting maps from biphasic to triphasic pulse pattern stimulation because of FNS after cochlear implantation (CI). All identified patients with FNS after CI from 2017 to 2019 were included in this study. The medical records of 11 patients (16 ears) were queried for demographic and radiological data, pure tone audiometry, speech reception thresholds, speech discrimination score at 65 dB, maximum comfortable levels, thresholds, and dynamic range. Then, these parameters were compared in the two conditions, biphasic and triphasic. RESULTS Using triphasic pulse stimulation only or combined with switch-off of few channels, complete resolution of FNS was achieved. Triphasic pulse pattern stimulation was associated with better speech discrimination scores (75.25 ± 26.13%) compared to the biphasic pulse (58.25 ± 26.13%). This triphasic strategy also showed higher maximum comfortable levels (36.62 ± 1.63 qu) than biphasic strategy (31.58 ± 2.5 qu). Moreover, the dynamic range was wider using triphasic pulse strategy. In general, the triphasic pulse pattern resulted in successful suppression of facial nerve stimulation with suitable maximum comfortable levels and better speech discrimination. CONCLUSION Triphasic pulse pattern stimulation is an appropriate tool in controlling FNS following cochlear implantation with wider dynamic range. We recommend that all patients with facial nerve stimulation after CI surgery be switched to a triphasic pulse program prior to considering further surgery.
Collapse
|
21
|
Goehring T, Arenberg JG, Carlyon RP. Using Spectral Blurring to Assess Effects of Channel Interaction on Speech-in-Noise Perception with Cochlear Implants. J Assoc Res Otolaryngol 2020; 21:353-371. [PMID: 32519088 PMCID: PMC7445227 DOI: 10.1007/s10162-020-00758-z] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2019] [Accepted: 05/21/2020] [Indexed: 01/07/2023] Open
Abstract
Cochlear implant (CI) listeners struggle to understand speech in background noise. Interactions between electrode channels due to current spread increase the masking of speech by noise and lead to difficulties with speech perception. Strategies that reduce channel interaction therefore have the potential to improve speech-in-noise perception by CI listeners, but previous results have been mixed. We investigated the effects of channel interaction on speech-in-noise perception and its association with spectro-temporal acuity in a listening study with 12 experienced CI users. Instead of attempting to reduce channel interaction, we introduced spectral blurring to simulate some of the effects of channel interaction by adjusting the overlap between electrode channels at the input level of the analysis filters or at the output by using several simultaneously stimulated electrodes per channel. We measured speech reception thresholds in noise as a function of the amount of blurring applied to either all 15 electrode channels or to 5 evenly spaced channels. Performance remained roughly constant as the amount of blurring applied to all channels increased up to some knee point, above which it deteriorated. This knee point differed across listeners in a way that correlated with performance on a non-speech spectro-temporal task, and is proposed here as an individual measure of channel interaction. Surprisingly, even extreme amounts of blurring applied to 5 channels did not affect performance. The effects on speech perception in noise were similar for blurring at the input and at the output of the CI. The results are in line with the assumption that experienced CI users can make use of a limited number of effective channels of information and tolerate some deviations from their everyday settings when identifying speech in the presence of a masker. Furthermore, these findings may explain the mixed results by strategies that optimized or deactivated a small number of electrodes evenly distributed along the array by showing that blurring or deactivating one-third of the electrodes did not harm speech-in-noise performance.
Collapse
Affiliation(s)
- Tobias Goehring
- Cambridge Hearing Group, Medical Research Council Cognition and Brain Sciences Unit, University of Cambridge, 15 Chaucer Road, Cambridge, CB2 7EF, UK.
| | - Julie G Arenberg
- Massachusetts Eye and Ear, Harvard Medical School, 243 Charles St, Boston, MA, 02114, USA
| | - Robert P Carlyon
- Cambridge Hearing Group, Medical Research Council Cognition and Brain Sciences Unit, University of Cambridge, 15 Chaucer Road, Cambridge, CB2 7EF, UK
| |
Collapse
|
22
|
Karoui C, James C, Barone P, Bakhos D, Marx M, Macherey O. Searching for the Sound of a Cochlear Implant: Evaluation of Different Vocoder Parameters by Cochlear Implant Users With Single-Sided Deafness. Trends Hear 2020; 23:2331216519866029. [PMID: 31533581 PMCID: PMC6753516 DOI: 10.1177/2331216519866029] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
Cochlear implantation in subjects with single-sided deafness (SSD) offers a unique opportunity to directly compare the percepts evoked by a cochlear implant (CI) with those evoked acoustically. Here, nine SSD-CI users performed a forced-choice task evaluating the similarity of speech processed by their CI with speech processed by several vocoders presented to their healthy ear. In each trial, subjects heard two intervals: their CI followed by a certain vocoder in Interval 1 and their CI followed by a different vocoder in Interval 2. The vocoders differed either (i) in carrier type-(sinusoidal [SINE], bandfiltered noise [NOISE], and pulse-spreading harmonic complex) or (ii) in frequency mismatch between the analysis and synthesis frequency ranges-(no mismatch, and two frequency-mismatched conditions of 2 and 4 equivalent rectangular bandwidths [ERBs]). Subjects had to state in which of the two intervals the CI and vocoder sounds were more similar. Despite a large intersubject variability, the PSHC vocoder was judged significantly more similar to the CI than SINE or NOISE vocoders. Furthermore, the No-mismatch and 2-ERB mismatch vocoders were judged significantly more similar to the CI than the 4-ERB mismatch vocoder. The mismatch data were also interpreted by comparing spiral ganglion characteristic frequencies with electrode contact positions determined from postoperative computed tomography scans. Only one subject demonstrated a pattern of preference consistent with adaptation to the CI sound processor frequency-to-electrode allocation table and two subjects showed possible partial adaptation. Those subjects with adaptation patterns presented overall small and consistent frequency mismatches across their electrode arrays.
Collapse
Affiliation(s)
- Chadlia Karoui
- Centre de Recherche Cerveau et Cognition, Toulouse, France.,Cochlear France SAS, Toulouse, France
| | - Chris James
- Cochlear France SAS, Toulouse, France.,Department of Otology-Neurotology and Skull Base Surgery, Purpan University Hospital, Toulouse, France
| | - Pascal Barone
- Centre de Recherche Cerveau et Cognition, Toulouse, France
| | - David Bakhos
- Université François-Rabelais de Tours, CHRU de Tours, France.,Ear Nose and Throat department, CHUR de Tours, Tours, France
| | - Mathieu Marx
- Centre de Recherche Cerveau et Cognition, Toulouse, France.,Department of Otology-Neurotology and Skull Base Surgery, Purpan University Hospital, Toulouse, France
| | - Olivier Macherey
- Aix Marseille University, CNRS, Centrale Marseille, LMA, Marseille, France
| |
Collapse
|
23
|
Winn MB. Accommodation of gender-related phonetic differences by listeners with cochlear implants and in a variety of vocoder simulations. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:174. [PMID: 32006986 PMCID: PMC7341679 DOI: 10.1121/10.0000566] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/30/2019] [Revised: 12/06/2019] [Accepted: 12/13/2019] [Indexed: 06/01/2023]
Abstract
Speech perception requires accommodation of a wide range of acoustic variability across talkers. A classic example is the perception of "sh" and "s" fricative sounds, which are categorized according to spectral details of the consonant itself, and also by the context of the voice producing it. Because women's and men's voices occupy different frequency ranges, a listener is required to make a corresponding adjustment of acoustic-phonetic category space for these phonemes when hearing different talkers. This pattern is commonplace in everyday speech communication, and yet might not be captured in accuracy scores for whole words, especially when word lists are spoken by a single talker. Phonetic accommodation for fricatives "s" and "sh" was measured in 20 cochlear implant (CI) users and in a variety of vocoder simulations, including those with noise carriers with and without peak picking, simulated spread of excitation, and pulsatile carriers. CI listeners showed strong phonetic accommodation as a group. Each vocoder produced phonetic accommodation except the 8-channel noise vocoder, despite its historically good match with CI users in word intelligibility. Phonetic accommodation is largely independent of linguistic factors and thus might offer information complementary to speech intelligibility tests which are partially affected by language processing.
Collapse
Affiliation(s)
- Matthew B Winn
- Department of Speech & Hearing Sciences, University of Minnesota, 164 Pillsbury Drive Southeast, Minneapolis, Minnesota 55455, USA
| |
Collapse
|
24
|
Luo X, Hayes L. Vibrotactile Stimulation Based on the Fundamental Frequency Can Improve Melodic Contour Identification of Normal-Hearing Listeners With a 4-Channel Cochlear Implant Simulation. Front Neurosci 2019; 13:1145. [PMID: 31736688 PMCID: PMC6828932 DOI: 10.3389/fnins.2019.01145] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/13/2019] [Accepted: 10/10/2019] [Indexed: 11/13/2022] Open
Abstract
Cochlear implant (CI) users' poor speech recognition in noise and music perception may be both due to their limited access to pitch cues such as the fundamental frequency (F0). Recent studies showed that similar to residual low-frequency acoustic hearing, vibrotactile presentation of the F0 significantly improved speech recognition in noise of CI users. The present study tested whether F0-based vibrotactile stimulation can improve melodic contour identification (MCI) of normal-hearing listeners with acoustically simulated CI processing. Each melodic contour consisted of five musical notes with one of nine contour patterns (rising, falling, or flat in each half of the contour). The F0 of the middle note was 220 or 880 Hz, and the frequency intervals between adjacent notes were 1, 3, or 5 semitones. The F0 of each note was extracted in real time and transposed to a vibration frequency centered around 110 Hz at the right forearm top. MCI was tested in five experimental conditions (with a 4- or 8-channel CI simulation alone, vibrotactile stimulation alone, and 4- or 8-channel CI simulation plus vibrotactile stimulation), each after the same amount of brief training was provided. Results showed that discrimination of vibrotactile stimuli significantly improved from chance to near perfect as the vibration frequency interval increased from 0.25 to 3 semitones. The MCI performance with vibrotactile stimulation alone was similar to that with the 4-channel CI simulation alone, but was significantly worse than that with the 8-channel CI simulation alone. Significant improvement in MCI performance with the addition of vibrotactile stimulation was only found with the 4-channel CI simulation when the middle F0 was 880 Hz and when the frequency intervals were 3 or 5 semitones. The improvement in MCI performance with than without vibrotactile stimulation was significantly correlated with the baseline MCI performance with 4-channel CI simulation alone or with the MCI performance difference between vibrotactile stimulation and 8-channel CI simulation. Therefore, when the simulated or real CI performance is relatively poor, vibrotactile stimulation based on the F0 may improve MCI with acoustic CI simulations and perhaps in real CI users as well.
Collapse
Affiliation(s)
- Xin Luo
- College of Health Solutions, Arizona State University, Tempe, AZ, United States
| | - Lauren Hayes
- School of Arts, Media and Engineering, Arizona State University, Tempe, AZ, United States
| |
Collapse
|
25
|
O'Neill ER, Kreft HA, Oxenham AJ. Cognitive factors contribute to speech perception in cochlear-implant users and age-matched normal-hearing listeners under vocoded conditions. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:195. [PMID: 31370651 PMCID: PMC6637026 DOI: 10.1121/1.5116009] [Citation(s) in RCA: 37] [Impact Index Per Article: 6.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/12/2023]
Abstract
This study examined the contribution of perceptual and cognitive factors to speech-perception abilities in cochlear-implant (CI) users. Thirty CI users were tested on word intelligibility in sentences with and without semantic context, presented in quiet and in noise. Performance was compared with measures of spectral-ripple detection and discrimination, thought to reflect peripheral processing, as well as with cognitive measures of working memory and non-verbal intelligence. Thirty age-matched and thirty younger normal-hearing (NH) adults also participated, listening via tone-excited vocoders, adjusted to produce mean performance for speech in noise comparable to that of the CI group. Results suggest that CI users may rely more heavily on semantic context than younger or older NH listeners, and that non-auditory working memory explains significant variance in the CI and age-matched NH groups. Between-subject variability in spectral-ripple detection thresholds was similar across groups, despite the spectral resolution for all NH listeners being limited by the same vocoder, whereas speech perception scores were more variable between CI users than between NH listeners. The results highlight the potential importance of central factors in explaining individual differences in CI users and question the extent to which standard measures of spectral resolution in CIs reflect purely peripheral processing.
Collapse
Affiliation(s)
- Erin R O'Neill
- Department of Psychology, University of Minnesota, Elliott Hall, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| | - Heather A Kreft
- Department of Psychology, University of Minnesota, Elliott Hall, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, Elliott Hall, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| |
Collapse
|
26
|
Speech Perception with Spectrally Non-overlapping Maskers as Measure of Spectral Resolution in Cochlear Implant Users. J Assoc Res Otolaryngol 2018; 20:151-167. [PMID: 30456730 DOI: 10.1007/s10162-018-00702-2] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/24/2018] [Accepted: 10/07/2018] [Indexed: 10/27/2022] Open
Abstract
Poor spectral resolution contributes to the difficulties experienced by cochlear implant (CI) users when listening to speech in noise. However, correlations between measures of spectral resolution and speech perception in noise have not always been found to be robust. It may be that the relationship between spectral resolution and speech perception in noise becomes clearer in conditions where the speech and noise are not spectrally matched, so that improved spectral resolution can assist in separating the speech from the masker. To test this prediction, speech intelligibility was measured with noise or tone maskers that were presented either in the same spectral channels as the speech or in interleaved spectral channels. Spectral resolution was estimated via a spectral ripple discrimination task. Results from vocoder simulations in normal-hearing listeners showed increasing differences in speech intelligibility between spectrally overlapped and interleaved maskers as well as improved spectral ripple discrimination with increasing spectral resolution. However, no clear differences were observed in CI users between performance with spectrally interleaved and overlapped maskers, or between tone and noise maskers. The results suggest that spectral resolution in current CIs is too poor to take advantage of the spectral separation produced by spectrally interleaved speech and maskers. Overall, the spectrally interleaved and tonal maskers produce a much larger difference in performance between normal-hearing listeners and CI users than do traditional speech-in-noise measures, and thus provide a more sensitive test of speech perception abilities for current and future implantable devices.
Collapse
|
27
|
Nie Y, Galvin JJ, Morikawa M, André V, Wheeler H, Fu QJ. Music and Speech Perception in Children Using Sung Speech. Trends Hear 2018; 22:2331216518766810. [PMID: 29609496 PMCID: PMC5888806 DOI: 10.1177/2331216518766810] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
This study examined music and speech perception in normal-hearing children with some or no musical training. Thirty children (mean age = 11.3 years), 15 with and 15 without formal music training participated in the study. Music perception was measured using a melodic contour identification (MCI) task; stimuli were a piano sample or sung speech with a fixed timbre (same word for each note) or a mixed timbre (different words for each note). Speech perception was measured in quiet and in steady noise using a matrix-styled sentence recognition task; stimuli were naturally intonated speech or sung speech with a fixed pitch (same note for each word) or a mixed pitch (different notes for each word). Significant musician advantages were observed for MCI and speech in noise but not for speech in quiet. MCI performance was significantly poorer with the mixed timbre stimuli. Speech performance in noise was significantly poorer with the fixed or mixed pitch stimuli than with spoken speech. Across all subjects, age at testing and MCI performance were significantly correlated with speech performance in noise. MCI and speech performance in quiet was significantly poorer for children than for adults from a related study using the same stimuli and tasks; speech performance in noise was significantly poorer for young than for older children. Long-term music training appeared to benefit melodic pitch perception and speech understanding in noise in these pediatric listeners.
Collapse
Affiliation(s)
- Yingjiu Nie
- 1 Department of Communication Sciences and Disorders, 3745 James Madison University , Harrisonburg, VA, USA
| | | | - Michael Morikawa
- 1 Department of Communication Sciences and Disorders, 3745 James Madison University , Harrisonburg, VA, USA
| | - Victoria André
- 1 Department of Communication Sciences and Disorders, 3745 James Madison University , Harrisonburg, VA, USA
| | - Harley Wheeler
- 1 Department of Communication Sciences and Disorders, 3745 James Madison University , Harrisonburg, VA, USA
| | - Qian-Jie Fu
- 3 Department of Head and Neck Surgery, University of California-Los Angeles, CA, USA
| |
Collapse
|
28
|
Preventing Facial Nerve Stimulation by Triphasic Pulse Stimulation in Cochlear Implant Users: Intraoperative Recordings. Otol Neurotol 2017; 38:e438-e444. [DOI: 10.1097/mao.0000000000001603] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
29
|
Grange JA, Culling JF, Harris NSL, Bergfeld S. Cochlear implant simulator with independent representation of the full spiral ganglion. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 142:EL484. [PMID: 29195445 DOI: 10.1121/1.5009602] [Citation(s) in RCA: 28] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/23/2023]
Abstract
In cochlear implant simulation with vocoders, narrow-band carriers deliver the envelopes from each analysis band to the cochlear positions of the simulated electrodes. However, this approach does not faithfully represent the continuous nature of the spiral ganglion. The proposed "SPIRAL" vocoder simulates current spread by mixing all envelopes across many tonal carriers. SPIRAL demonstrated that the classic finding of reduced speech-intelligibility benefit with additional electrodes could be due to current spread. SPIRAL produced lower speech reception thresholds than an equivalent noise vocoder. These thresholds are stable for between 20 and 160 carriers.
Collapse
Affiliation(s)
- Jacques A Grange
- School of Psychology, Cardiff University, CF103AT, Cardiff, United Kingdom , ,
| | - John F Culling
- School of Psychology, Cardiff University, CF103AT, Cardiff, United Kingdom , ,
| | - Naomi S L Harris
- School of Psychology, Cardiff University, CF103AT, Cardiff, United Kingdom , ,
| | - Sven Bergfeld
- Department of Cognitive Neuroscience, Bielefeld University, 33615 Bielefeld, Germany
| |
Collapse
|
30
|
Abstract
Auditory perception is our main gateway to communication with others via speech and music, and it also plays an important role in alerting and orienting us to new events. This review provides an overview of selected topics pertaining to the perception and neural coding of sound, starting with the first stage of filtering in the cochlea and its profound impact on perception. The next topic, pitch, has been debated for millennia, but recent technical and theoretical developments continue to provide us with new insights. Cochlear filtering and pitch both play key roles in our ability to parse the auditory scene, enabling us to attend to one auditory object or stream while ignoring others. An improved understanding of the basic mechanisms of auditory perception will aid us in the quest to tackle the increasingly important problem of hearing loss in our aging population.
Collapse
Affiliation(s)
- Andrew J Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455;
| |
Collapse
|
31
|
Vocoder Simulations Explain Complex Pitch Perception Limitations Experienced by Cochlear Implant Users. J Assoc Res Otolaryngol 2017; 18:789-802. [PMID: 28733803 DOI: 10.1007/s10162-017-0632-x] [Citation(s) in RCA: 36] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2017] [Accepted: 06/26/2017] [Indexed: 01/06/2023] Open
Abstract
Pitch plays a crucial role in speech and music, but is highly degraded for people with cochlear implants, leading to severe communication challenges in noisy environments. Pitch is determined primarily by the first few spectrally resolved harmonics of a tone. In implants, access to this pitch is limited by poor spectral resolution, due to the limited number of channels and interactions between adjacent channels. Here we used noise-vocoder simulations to explore how many channels, and how little channel interaction, are required to elicit pitch. Results suggest that two to four times the number of channels are needed, along with interactions reduced by an order of magnitude, than available in current devices. These new constraints not only provide insights into the basic mechanisms of pitch coding in normal hearing but also suggest that spectrally based complex pitch is unlikely to be generated in implant users without significant changes in the method or site of stimulation.
Collapse
|
32
|
van de Velde DJ, Schiller NO, van Heuven VJ, Levelt CC, van Ginkel J, Beers M, Briaire JJ, Frijns JHM. The perception of emotion and focus prosody with varying acoustic cues in cochlear implant simulations with varying filter slopes. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 141:3349. [PMID: 28599540 PMCID: PMC5436976 DOI: 10.1121/1.4982198] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/27/2016] [Revised: 03/15/2017] [Accepted: 04/11/2017] [Indexed: 06/07/2023]
Abstract
This study aimed to find the optimal filter slope for cochlear implant simulations (vocoding) by testing the effect of a wide range of slopes on the discrimination of emotional and linguistic (focus) prosody, with varying availability of F0 and duration cues. Forty normally hearing participants judged if (non-)vocoded sentences were pronounced with happy or sad emotion, or with adjectival or nominal focus. Sentences were recorded as natural stimuli and manipulated to contain only emotion- or focus-relevant segmental duration or F0 information or both, and then noise-vocoded with 5, 20, 80, 120, and 160 dB/octave filter slopes. Performance increased with steeper slopes, but only up to 120 dB/octave, with bigger effects for emotion than for focus perception. For emotion, results with both cues most closely resembled results with F0, while for focus results with both cues most closely resembled those with duration, showing emotion perception relies primarily on F0, and focus perception on duration. This suggests that filter slopes affect focus perception less than emotion perception because for emotion, F0 is both more informative and more affected. The performance increase until extreme filter slope values suggests that much performance improvement in prosody perception is still to be gained for CI users.
Collapse
Affiliation(s)
- Daan J van de Velde
- Leiden University Centre for Linguistics, Leiden University, Van Wijkplaats 3, 2311 BX, Leiden, the Netherlands
| | - Niels O Schiller
- Leiden University Centre for Linguistics, Leiden University, Van Wijkplaats 3, 2311 BX, Leiden, the Netherlands
| | - Vincent J van Heuven
- Department of Applied Linguistics, Pannon Egyetem, 10 Egyetem Utca, 8200 Veszprém, Hungary
| | - Claartje C Levelt
- Leiden University Centre for Linguistics, Leiden University, Van Wijkplaats 3, 2311 BX, Leiden, the Netherlands
| | - Joost van Ginkel
- Leiden University Centre for Child and Family Studies, Wassenaarseweg 52, 2333 AK, Leiden, the Netherlands
| | - Mieke Beers
- Leiden University Medical Center, Ears, Nose, and Throat Department, Postbus 9600, 2300 RC, Leiden, the Netherlands
| | - Jeroen J Briaire
- Leiden University Medical Center, Ears, Nose, and Throat Department, Postbus 9600, 2300 RC, Leiden, the Netherlands
| | - Johan H M Frijns
- Leiden University Medical Center, Ears, Nose, and Throat Department, Postbus 9600, 2300 RC, Leiden, the Netherlands
| |
Collapse
|
33
|
Abstract
Combined use of a hearing aid (HA) and cochlear implant (CI) has been shown to improve CI users’ speech and music performance. However, different hearing devices, test stimuli, and listening tasks may interact and obscure bimodal benefits. In this study, speech and music perception were measured in bimodal listeners for CI-only, HA-only, and CI + HA conditions, using the Sung Speech Corpus, a database of monosyllabic words produced at different fundamental frequencies. Sentence recognition was measured using sung speech in which pitch was held constant or varied across words, as well as for spoken speech. Melodic contour identification (MCI) was measured using sung speech in which the words were held constant or varied across notes. Results showed that sentence recognition was poorer with sung speech relative to spoken, with little difference between sung speech with a constant or variable pitch; mean performance was better with CI-only relative to HA-only, and best with CI + HA. MCI performance was better with constant words versus variable words; mean performance was better with HA-only than with CI-only and was best with CI + HA. Relative to CI-only, a strong bimodal benefit was observed for speech and music perception. Relative to the better ear, bimodal benefits remained strong for sentence recognition but were marginal for MCI. While variations in pitch and timbre may negatively affect CI users’ speech and music perception, bimodal listening may partially compensate for these deficits.
Collapse
Affiliation(s)
- Joseph D Crew
- University of Southern California, Los Angeles, CA, USA
| | | | - Qian-Jie Fu
- University of California-Los Angeles, CA, USA
| |
Collapse
|
34
|
Langner F, Jürgens T. Forward-Masked Frequency Selectivity Improvements in Simulated and Actual Cochlear Implant Users Using a Preprocessing Algorithm. Trends Hear 2016; 20:20/0/2331216516659632. [PMID: 27604785 PMCID: PMC5017570 DOI: 10.1177/2331216516659632] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Frequency selectivity can be quantified using masking paradigms, such as psychophysical tuning curves (PTCs). Normal-hearing (NH) listeners show sharp PTCs that are level- and frequency-dependent, whereas frequency selectivity is strongly reduced in cochlear implant (CI) users. This study aims at (a) assessing individual shapes of PTCs in CI users, (b) comparing these shapes to those of simulated CI listeners (NH listeners hearing through a CI simulation), and (c) increasing the sharpness of PTCs using a biologically inspired dynamic compression algorithm, BioAid, which has been shown to sharpen the PTC shape in hearing-impaired listeners. A three-alternative-forced-choice forward-masking technique was used to assess PTCs in 8 CI users (with their own speech processor) and 11 NH listeners (with and without listening through a vocoder to simulate electric hearing). CI users showed flat PTCs with large interindividual variability in shape, whereas simulated CI listeners had PTCs of the same average flatness, but more homogeneous shapes across listeners. The algorithm BioAid was used to process the stimuli before entering the CI users’ speech processor or the vocoder simulation. This algorithm was able to partially restore frequency selectivity in both groups, particularly in seven out of eight CI users, meaning significantly sharper PTCs than in the unprocessed condition. The results indicate that algorithms can improve the large-scale sharpness of frequency selectivity in some CI users. This finding may be useful for the design of sound coding strategies particularly for situations in which high frequency selectivity is desired, such as for music perception.
Collapse
Affiliation(s)
- Florian Langner
- Medizinische Physik, Cluster of Excellence "Hearing4all," Carl von Ossietzky University, Oldenburg, Germany Forschungszentrum Neurosensorik, Carl von Ossietzky University, Oldenburg, Germany Department of Otolaryngology, Medical University Hannover, Hannover, Germany
| | - Tim Jürgens
- Medizinische Physik, Cluster of Excellence "Hearing4all," Carl von Ossietzky University, Oldenburg, Germany Forschungszentrum Neurosensorik, Carl von Ossietzky University, Oldenburg, Germany
| |
Collapse
|
35
|
Guevara N, Bozorg-Grayeli A, Bebear JP, Ardoint M, Saaï S, Gnansia D, Hoen M, Romanet P, Lavieille JP. The Voice Track multiband single-channel modified Wiener-filter noise reduction system for cochlear implants: patients' outcomes and subjective appraisal. Int J Audiol 2016; 55:431-8. [DOI: 10.3109/14992027.2016.1172267] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
Affiliation(s)
- Nicolas Guevara
- CHU de Nice, University Head and Neck Institute, Nice, France,
| | - Alexis Bozorg-Grayeli
- Department of Head and Neck Surgery, CHU - University Hospital of Dijon, Dijon, France,
| | - Jean-Pierre Bebear
- ENT Department, Pellegrin Hospital, University of Bordeaux, Bordeaux, France,
| | - Marine Ardoint
- Clinical and Scientific Research Department, Oticon Medical - Neurelec, Vallauris, France, and
| | - Sonia Saaï
- Clinical and Scientific Research Department, Oticon Medical - Neurelec, Vallauris, France, and
| | - Dan Gnansia
- Clinical and Scientific Research Department, Oticon Medical - Neurelec, Vallauris, France, and
| | - Michel Hoen
- Clinical and Scientific Research Department, Oticon Medical - Neurelec, Vallauris, France, and
| | - Philippe Romanet
- Department of Head and Neck Surgery, CHU - University Hospital of Dijon, Dijon, France,
| | | |
Collapse
|
36
|
DeVries L, Scheperle R, Bierer JA. Assessing the Electrode-Neuron Interface with the Electrically Evoked Compound Action Potential, Electrode Position, and Behavioral Thresholds. J Assoc Res Otolaryngol 2016; 17:237-52. [PMID: 26926152 DOI: 10.1007/s10162-016-0557-9] [Citation(s) in RCA: 106] [Impact Index Per Article: 11.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2015] [Accepted: 02/04/2016] [Indexed: 11/26/2022] Open
Abstract
Variability in speech perception scores among cochlear implant listeners may largely reflect the variable efficacy of implant electrodes to convey stimulus information to the auditory nerve. In the present study, three metrics were applied to assess the quality of the electrode-neuron interface of individual cochlear implant channels: the electrically evoked compound action potential (ECAP), the estimation of electrode position using computerized tomography (CT), and behavioral thresholds using focused stimulation. The primary motivation of this approach is to evaluate the ECAP as a site-specific measure of the electrode-neuron interface in the context of two peripheral factors that likely contribute to degraded perception: large electrode-to-modiolus distance and reduced neural density. Ten unilaterally implanted adults with Advanced Bionics HiRes90k devices participated. ECAPs were elicited with monopolar stimulation within a forward-masking paradigm to construct channel interaction functions (CIF), behavioral thresholds were obtained with quadrupolar (sQP) stimulation, and data from imaging provided estimates of electrode-to-modiolus distance and scalar location (scala tympani (ST), intermediate, or scala vestibuli (SV)) for each electrode. The width of the ECAP CIF was positively correlated with electrode-to-modiolus distance; both of these measures were also influenced by scalar position. The ECAP peak amplitude was negatively correlated with behavioral thresholds. Moreover, subjects with low behavioral thresholds and large ECAP amplitudes, averaged across electrodes, tended to have higher speech perception scores. These results suggest a potential clinical role for the ECAP in the objective assessment of individual cochlear implant channels, with the potential to improve speech perception outcomes.
Collapse
Affiliation(s)
- Lindsay DeVries
- Department of Speech and Hearing Sciences, University of Washington, 4131 15th Ave NE, Seattle, WA, 98105, USA.
| | - Rachel Scheperle
- Department of Communication Sciences and Disorders, Wendell Johnson Speech and Hearing Center, University of Iowa, Iowa City, Iowa, 52242, USA
| | - Julie Arenberg Bierer
- Department of Speech and Hearing Sciences, University of Washington, 4131 15th Ave NE, Seattle, WA, 98105, USA
| |
Collapse
|
37
|
Shannon RV. Auditory implant research at the House Ear Institute 1989-2013. Hear Res 2015; 322:57-66. [PMID: 25449009 PMCID: PMC4380593 DOI: 10.1016/j.heares.2014.11.003] [Citation(s) in RCA: 21] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/11/2014] [Revised: 11/04/2014] [Accepted: 11/07/2014] [Indexed: 11/29/2022]
Abstract
The House Ear Institute (HEI) had a long and distinguished history of auditory implant innovation and development. Early clinical innovations include being one of the first cochlear implant (CI) centers, being the first center to implant a child with a cochlear implant in the US, developing the auditory brainstem implant, and developing multiple surgical approaches and tools for Otology. This paper reviews the second stage of auditory implant research at House - in-depth basic research on perceptual capabilities and signal processing for both cochlear implants and auditory brainstem implants. Psychophysical studies characterized the loudness and temporal perceptual properties of electrical stimulation as a function of electrical parameters. Speech studies with the noise-band vocoder showed that only four bands of tonotopically arrayed information were sufficient for speech recognition, and that most implant users were receiving the equivalent of 8-10 bands of information. The noise-band vocoder allowed us to evaluate the effects of the manipulation of the number of bands, the alignment of the bands with the original tonotopic map, and distortions in the tonotopic mapping, including holes in the neural representation. Stimulation pulse rate was shown to have only a small effect on speech recognition. Electric fields were manipulated in position and sharpness, showing the potential benefit of improved tonotopic selectivity. Auditory training shows great promise for improving speech recognition for all patients. And the Auditory Brainstem Implant was developed and improved and its application expanded to new populations. Overall, the last 25 years of research at HEI helped increase the basic scientific understanding of electrical stimulation of hearing and contributed to the improved outcomes for patients with the CI and ABI devices. This article is part of a Special Issue entitled .
Collapse
Affiliation(s)
- Robert V Shannon
- Department of Otolaryngology, University of Southern California, Keck School of Medicine of USC, 806 W. Adams Blvd, Los Angeles, CA 90007-2505, USA.
| |
Collapse
|
38
|
Won JH, Jones GL, Moon IJ, Rubinstein JT. Spectral and temporal analysis of simulated dead regions in cochlear implants. J Assoc Res Otolaryngol 2015; 16:285-307. [PMID: 25740402 DOI: 10.1007/s10162-014-0502-8] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2014] [Accepted: 12/23/2014] [Indexed: 11/29/2022] Open
Abstract
A cochlear implant (CI) electrode in a "cochlear dead region" will excite neighboring neural populations. In previous research that simulated such dead regions, stimulus information in the simulated dead region was either added to the immediately adjacent frequency regions or dropped entirely. There was little difference in speech perception ability between the two conditions. This may imply that there may be little benefit of ensuring that stimulus information on an electrode in a suspected cochlear dead region is transmitted. Alternatively, performance may be enhanced by a broader frequency redistribution, rather than adding stimuli from the dead region to the edges. In the current experiments, cochlear dead regions were introduced by excluding selected CI electrodes or vocoder noise-bands. Participants were assessed for speech understanding as well as spectral and temporal sensitivities as a function of the size of simulated dead regions. In one set of tests, the normal input frequency range of the sound processor was distributed among the active electrodes in bands with approximately logarithmic spacing ("redistributed" maps); in the remaining tests, information in simulated dead regions was dropped ("dropped" maps). Word recognition and Schroeder-phase discrimination performance, which require both spectral and temporal sensitivities, decreased as the size of simulated dead regions increased, but the redistributed and dropped remappings showed similar performance in these two tasks. Psychoacoustic experiments showed that the near match in word scores may reflect a tradeoff between spectral and temporal sensitivity: spectral-ripple discrimination was substantially degraded in the redistributed condition relative to the dropped condition while performance in a temporal modulation detection task degraded in the dropped condition but remained constant in the redistributed condition.
Collapse
Affiliation(s)
- Jong Ho Won
- Virginia Merrill Bloedel Hearing Research Center, Department of Otolaryngology-Head and Neck Surgery, University of Washington, Seattle, WA, 98195, USA
| | | | | | | |
Collapse
|
39
|
Mesnildrey Q, Macherey O. Simulating the dual-peak excitation pattern produced by bipolar stimulation of a cochlear implant: effects on speech intelligibility. Hear Res 2014; 319:32-47. [PMID: 25449010 DOI: 10.1016/j.heares.2014.11.001] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/19/2014] [Revised: 10/28/2014] [Accepted: 11/05/2014] [Indexed: 10/24/2022]
Abstract
Several electrophysiological and psychophysical studies have shown that the spatial excitation pattern produced by bipolar stimulation of a cochlear implant (CI) can have a dual-peak shape. The perceptual effects of this dual-peak shape were investigated using noise-vocoded CI simulations in which synthesis filters were designed to simulate the spread of neural activity produced by various electrode configurations, as predicted by a simple cochlear model. Experiments 1 and 2 tested speech recognition in the presence of a concurrent speech masker for various sets of single-peak and dual-peak synthesis filters and different numbers of channels. Similarly as results obtained in real CIs, both monopolar (MP, single-peak) and bipolar (BP + 1, dual-peak) simulations showed a plateau of performance above 8 channels. The benefit of increasing the number of channels was also lower for BP + 1 than for MP. This shows that channel interactions in BP + 1 become especially deleterious for speech intelligibility when a simulated electrode acts both as an active and as a return electrode for different channels because envelope information from two different analysis bands are being conveyed to the same spectral location. Experiment 3 shows that these channel interactions are even stronger in wide BP configuration (BP + 5), likely because the interfering speech envelopes are less correlated than in narrow BP + 1. Although the exact effects of dual- or multi-peak excitation in real CIs remain to be determined, this series of experiments suggest that multipolar stimulation strategies, such as bipolar or tripolar, should be controlled to avoid neural excitation in the vicinity of the return electrodes.
Collapse
Affiliation(s)
- Quentin Mesnildrey
- LMA-CNRS, UPR 7051, Aix-Marseille Univ., Centrale Marseille, 31 Chemin Joseph Aiguier, 13402 Marseille Cedex 20, France.
| | - Olivier Macherey
- LMA-CNRS, UPR 7051, Aix-Marseille Univ., Centrale Marseille, 31 Chemin Joseph Aiguier, 13402 Marseille Cedex 20, France
| |
Collapse
|
40
|
Jones H, Kan A, Litovsky RY. Comparing sound localization deficits in bilateral cochlear-implant users and vocoder simulations with normal-hearing listeners. Trends Hear 2014; 18:18/0/2331216514554574. [PMID: 25385244 PMCID: PMC4271768 DOI: 10.1177/2331216514554574] [Citation(s) in RCA: 28] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
Bilateral cochlear-implant (BiCI) users are less accurate at localizing free-field (FF) sound sources than normal-hearing (NH) listeners. This performance gap is not well understood but is likely due to a combination of compromises in acoustic signal representation by the two independent speech processors and neural degradation of auditory pathways associated with a patient's hearing loss. To exclusively investigate the effect of CI speech encoding on horizontal-plane sound localization, the present study measured sound localization performance in NH subjects listening to vocoder processed and nonvocoded virtual acoustic space (VAS) stimuli. Various aspects of BiCI stimulation such as independently functioning devices, variable across-ear channel selection, and pulsatile stimulation were simulated using uncorrelated noise (Nu), correlated noise (N0), or Gaussian-enveloped tone (GET) carriers during vocoder processing. Additionally, FF sound localization in BiCI users was measured in the same testing environment for comparison. Distinct response patterns across azimuthal locations were evident for both listener groups and were analyzed using a multilevel regression analysis. Simulated implant speech encoding, regardless of carrier, was detrimental to NH localization and the GET vocoder best simulated BiCI FF performance in NH listeners. Overall, the detrimental effect of vocoder processing on NH performance suggests that sound localization deficits may persist even for BiCI patients who have minimal neural degradation associated with their hearing loss and indicates that CI speech encoding plays a significant role in the sound localization deficits experienced by BiCI users.
Collapse
Affiliation(s)
- Heath Jones
- Waisman Center, University of Wisconsin-Madison, WI, USA
| | - Alan Kan
- Waisman Center, University of Wisconsin-Madison, WI, USA
| | | |
Collapse
|
41
|
Oxenham AJ, Kreft HA. Speech perception in tones and noise via cochlear implants reveals influence of spectral resolution on temporal processing. Trends Hear 2014; 18:18/0/2331216514553783. [PMID: 25315376 PMCID: PMC4227666 DOI: 10.1177/2331216514553783] [Citation(s) in RCA: 76] [Impact Index Per Article: 6.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Under normal conditions, human speech is remarkably robust to degradation by noise and other distortions. However, people with hearing loss, including those with cochlear implants, often experience great difficulty in understanding speech in noisy environments. Recent work with normal-hearing listeners has shown that the amplitude fluctuations inherent in noise contribute strongly to the masking of speech. In contrast, this study shows that speech perception via a cochlear implant is unaffected by the inherent temporal fluctuations of noise. This qualitative difference between acoustic and electric auditory perception does not seem to be due to differences in underlying temporal acuity but can instead be explained by the poorer spectral resolution of cochlear implants, relative to the normally functioning ear, which leads to an effective smoothing of the inherent temporal-envelope fluctuations of noise. The outcome suggests an unexpected trade-off between the detrimental effects of poorer spectral resolution and the beneficial effects of a smoother noise temporal envelope. This trade-off provides an explanation for the long-standing puzzle of why strong correlations between speech understanding and spectral resolution have remained elusive. The results also provide a potential explanation for why cochlear-implant users and hearing-impaired listeners exhibit reduced or absent masking release when large and relatively slow temporal fluctuations are introduced in noise maskers. The multitone maskers used here may provide an effective new diagnostic tool for assessing functional hearing loss and reduced spectral resolution.
Collapse
Affiliation(s)
- Andrew J Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, MN, USA Department of Otolaryngology, University of Minnesota, Minneapolis, MN, USA
| | - Heather A Kreft
- Department of Otolaryngology, University of Minnesota, Minneapolis, MN, USA
| |
Collapse
|
42
|
Oxenham AJ, Kreft HA. Speech perception in tones and noise via cochlear implants reveals influence of spectral resolution on temporal processing. Trends Hear 2014; 18:18/0/2331216514553783. [PMID: 25315376 DOI: 10.1177/233121651455378] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/26/2023] Open
Abstract
Under normal conditions, human speech is remarkably robust to degradation by noise and other distortions. However, people with hearing loss, including those with cochlear implants, often experience great difficulty in understanding speech in noisy environments. Recent work with normal-hearing listeners has shown that the amplitude fluctuations inherent in noise contribute strongly to the masking of speech. In contrast, this study shows that speech perception via a cochlear implant is unaffected by the inherent temporal fluctuations of noise. This qualitative difference between acoustic and electric auditory perception does not seem to be due to differences in underlying temporal acuity but can instead be explained by the poorer spectral resolution of cochlear implants, relative to the normally functioning ear, which leads to an effective smoothing of the inherent temporal-envelope fluctuations of noise. The outcome suggests an unexpected trade-off between the detrimental effects of poorer spectral resolution and the beneficial effects of a smoother noise temporal envelope. This trade-off provides an explanation for the long-standing puzzle of why strong correlations between speech understanding and spectral resolution have remained elusive. The results also provide a potential explanation for why cochlear-implant users and hearing-impaired listeners exhibit reduced or absent masking release when large and relatively slow temporal fluctuations are introduced in noise maskers. The multitone maskers used here may provide an effective new diagnostic tool for assessing functional hearing loss and reduced spectral resolution.
Collapse
Affiliation(s)
- Andrew J Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, MN, USA Department of Otolaryngology, University of Minnesota, Minneapolis, MN, USA
| | - Heather A Kreft
- Department of Otolaryngology, University of Minnesota, Minneapolis, MN, USA
| |
Collapse
|
43
|
Abstract
OBJECTIVES A finite element model of the human coiled cochlea was used to model the voltage distribution due to stimulation by the individual electrodes of a cochlear implant. The scalar position of the electrode array was also varied in order to investigate its effect on the voltage distribution. Multi-electrode current focusing methods were then investigated, with the aim of increasing spatial selectivity. METHODS Simultaneous current focusing is initially achieved, as in previous publications, by calculating the input currents to the 22 electrodes that best separates the voltages at these electrode positions. The benefits of this electrode focusing strategy do not, however, entirely carry over to the predicted voltage distributions at the position of the spiral ganglion cells, where excitation is believed to occur. A novel focusing strategy is then simulated, which compensates for the impedances between the currents at the electrode sites and the voltage distribution directly at the position of the spiral ganglion cells. RESULTS The new strategy produces much better focusing at the sites of the spiral ganglion cells, as expected, but at the cost of increased current requirements. Regularization was introduced in order to reduce current requirements, which also reduced the sensitivity of the solution to uncertainties in the impedance matrix, so that improved focusing was achieved with similar current requirements to that of electrode focusing. DISCUSSION Although such focusing strategies cannot be achieved in practice at the moment, since the responses from the electrodes to the neural sites cannot be determined with currently available recording methods, these results do support the feasibility of a more effective focusing strategy, which may provide improved spectral resolution leading to improved perception of sound.
Collapse
|
44
|
van de Velde DJ, Dritsakis G, Frijns JHM, van Heuven VJ, Schiller NO. The effect of spectral smearing on the identification of pureF0intonation contours in vocoder simulations of cochlear implants. Cochlear Implants Int 2014; 16:77-87. [DOI: 10.1179/1754762814y.0000000086] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/31/2022]
|
45
|
Fuller CD, Galvin JJ, Maat B, Free RH, Başkent D. The musician effect: does it persist under degraded pitch conditions of cochlear implant simulations? Front Neurosci 2014; 8:179. [PMID: 25071428 PMCID: PMC4075350 DOI: 10.3389/fnins.2014.00179] [Citation(s) in RCA: 61] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/05/2014] [Accepted: 06/08/2014] [Indexed: 12/05/2022] Open
Abstract
Cochlear implants (CIs) are auditory prostheses that restore hearing via electrical stimulation of the auditory nerve. Compared to normal acoustic hearing, sounds transmitted through the CI are spectro-temporally degraded, causing difficulties in challenging listening tasks such as speech intelligibility in noise and perception of music. In normal hearing (NH), musicians have been shown to better perform than non-musicians in auditory processing and perception, especially for challenging listening tasks. This “musician effect” was attributed to better processing of pitch cues, as well as better overall auditory cognitive functioning in musicians. Does the musician effect persist when pitch cues are degraded, as it would be in signals transmitted through a CI? To answer this question, NH musicians and non-musicians were tested while listening to unprocessed signals or to signals processed by an acoustic CI simulation. The task increasingly depended on pitch perception: (1) speech intelligibility (words and sentences) in quiet or in noise, (2) vocal emotion identification, and (3) melodic contour identification (MCI). For speech perception, there was no musician effect with the unprocessed stimuli, and a small musician effect only for word identification in one noise condition, in the CI simulation. For emotion identification, there was a small musician effect for both. For MCI, there was a large musician effect for both. Overall, the effect was stronger as the importance of pitch in the listening task increased. This suggests that the musician effect may be more rooted in pitch perception, rather than in a global advantage in cognitive processing (in which musicians would have performed better in all tasks). The results further suggest that musical training before (and possibly after) implantation might offer some advantage in pitch processing that could partially benefit speech perception, and more strongly emotion and music perception.
Collapse
Affiliation(s)
- Christina D Fuller
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen Groningen, Netherlands ; Research School of Behavioral and Cognitive Neurosciences, Graduate School of Medical Sciences, University of Groningen Groningen, Netherlands
| | - John J Galvin
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen Groningen, Netherlands ; Research School of Behavioral and Cognitive Neurosciences, Graduate School of Medical Sciences, University of Groningen Groningen, Netherlands ; Division of Communication and Auditory Neuroscience, House Research Institute Los Angeles, CA, USA ; Department of Head and Neck Surgery, David Geffen School of Medicine, UCLA Los Angeles, CA, USA
| | - Bert Maat
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen Groningen, Netherlands ; Research School of Behavioral and Cognitive Neurosciences, Graduate School of Medical Sciences, University of Groningen Groningen, Netherlands
| | - Rolien H Free
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen Groningen, Netherlands ; Research School of Behavioral and Cognitive Neurosciences, Graduate School of Medical Sciences, University of Groningen Groningen, Netherlands
| | - Deniz Başkent
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen Groningen, Netherlands ; Research School of Behavioral and Cognitive Neurosciences, Graduate School of Medical Sciences, University of Groningen Groningen, Netherlands
| |
Collapse
|
46
|
Limb CJ, Roy AT. Technological, biological, and acoustical constraints to music perception in cochlear implant users. Hear Res 2014; 308:13-26. [DOI: 10.1016/j.heares.2013.04.009] [Citation(s) in RCA: 164] [Impact Index Per Article: 14.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/17/2013] [Revised: 04/04/2013] [Accepted: 04/22/2013] [Indexed: 11/30/2022]
|
47
|
|