1
|
Karosas DM, Gonzales L, Wang Y, Bergevin C, Carney LH, Henry KS. Otoacoustic emissions but not behavioral measurements predict cochlear-nerve frequency tuning in an avian vocal-communication specialist. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2025:2024.08.29.610326. [PMID: 39257830 PMCID: PMC11383700 DOI: 10.1101/2024.08.29.610326] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/12/2024]
Abstract
Frequency analysis by the cochlea forms a key foundation for all subsequent auditory processing. Stimulus-frequency otoacoustic emissions (SFOAEs) are a potentially powerful alternative to traditional behavioral experiments for estimating cochlear tuning without invasive testing, as is necessary in humans. Which methods accurately predict cochlear tuning remains controversial due to only a single animal study comparing SFOAE-based, behavioral, and cochlear frequency tuning in the same species. The budgerigar ( Melopsittacus undulatus ) is a parakeet species with human-like behavioral sensitivity to many sounds and the capacity to mimic speech. Intriguingly, previous studies of critical bands, psychophysical tuning curves, and critical ratios in budgerigars show that behavioral tuning sharpness increases dramatically with increasing frequency from 1-3.5 kHz, doubling once per octave with peak tuning sharpness from 3.5-4 kHz. The pattern contrasts with slower monotonic growth of behavioral tuning sharpness with increasing frequency in other animals, including most avian species, suggesting a possible auditory specialization in budgerigars. We measured SFOAE-based and cochlear-afferent tuning in budgerigars, for comparison to previously reported behavioral results. SFOAE-based and cochlear-afferent tuning sharpness both increased monotonically and relatively slowly for higher frequencies, in contrast to the behavioral pattern. SFOAE-based tuning in budgerigars accurately predicted cochlear frequency tuning, and both measures aligned with typical patterns of cochlear tuning in other species. Divergent behavioral tuning in budgerigars is unlikely attributable to the periphery and could reflect specializations for central processing of masked signals. Our findings highlight the value of SFOAEs for estimating cochlear tuning and caution against direct inference of peripheral tuning from behavioral critical bands, psychophysical tuning curves, and critical ratios.
Collapse
|
2
|
Guest DR, Cameron DA, Schwarz DM, Leong UC, Richards VM, Carney LH. Profile analysis in listeners with normal and elevated audiometric thresholds: Behavioral and modeling results. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 156:4303-4325. [PMID: 39740047 DOI: 10.1121/10.0034635] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/26/2024] [Accepted: 11/28/2024] [Indexed: 01/02/2025]
Abstract
Profile-analysis experiments measure the ability to discriminate complex sounds based on patterns, or profiles, in their amplitude spectra. Studies of profile analysis have focused on normal-hearing listeners and target frequencies near 1 kHz. To provide more insight into underlying mechanisms, we studied profile analysis over a large target frequency range (0.5-4 kHz) and in listeners with both normal and elevated audiometric thresholds. We found that profile analysis degrades at high frequencies and that the effect of spacing between nearby frequency components differs with frequency. Consistent with prior reports, elevated audiometric thresholds were not associated with impaired performance when stimuli consisted of few distantly spaced frequency components. However, elevated audiometric thresholds were associated with elevated profile-analysis thresholds for stimuli composed of many closely spaced frequency components. Behavioral thresholds from listeners with and without hearing loss were predicted by decoding firing rates from simulated auditory-nerve fibers or simulated modulation-sensitive inferior-colliculus neurons. Although responses from both model stages informed some aspects of the behavioral data, only population decoding of inferior-colliculus responses accounted for the worsening of profile-analysis thresholds at high target frequencies. Collectively, these results suggest that profile analysis involves multiple non-peripheral factors, including multichannel comparisons and midbrain tuning to amplitude modulation.
Collapse
Affiliation(s)
- Daniel R Guest
- Department of Biomedical Engineering, University of Rochester, Rochester, New York 14620, USA
| | - David A Cameron
- Department of Biomedical Engineering, University of Rochester, Rochester, New York 14620, USA
| | - Douglas M Schwarz
- Department of Biomedical Engineering, University of Rochester, Rochester, New York 14620, USA
- Department of Neuroscience, University of Rochester, Rochester, New York 14620, USA
| | - U-Cheng Leong
- Department of Otolaryngology, University of Rochester, Rochester, New York 14620, USA
| | - Virginia M Richards
- Department of Cognitive Sciences, University of California, Irvine, California 92697, USA
| | - Laurel H Carney
- Department of Biomedical Engineering, University of Rochester, Rochester, New York 14620, USA
- Department of Neuroscience, University of Rochester, Rochester, New York 14620, USA
| |
Collapse
|
3
|
Mackey CA, Hauser S, Schoenhaut AM, Temghare N, Ramachandran R. Hierarchical differences in the encoding of amplitude modulation in the subcortical auditory system of awake nonhuman primates. J Neurophysiol 2024; 132:1098-1114. [PMID: 39140590 PMCID: PMC11427057 DOI: 10.1152/jn.00329.2024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2024] [Revised: 07/31/2024] [Accepted: 08/12/2024] [Indexed: 08/15/2024] Open
Abstract
Sinusoidal amplitude modulation (SAM) is a key feature of complex sounds. Although psychophysical studies have characterized SAM perception, and neurophysiological studies in anesthetized animals report a transformation from the cochlear nucleus' (CN; brainstem) temporal code to the inferior colliculus' (IC; midbrain's) rate code, none have used awake animals or nonhuman primates to compare CN and IC's coding strategies to modulation-frequency perception. To address this, we recorded single-unit responses and compared derived neurometric measures in the CN and IC to psychometric measures of modulation frequency (MF) discrimination in macaques. IC and CN neurons often exhibited tuned responses to SAM in rate and spike-timing measures of modulation coding. Neurometric thresholds spanned a large range (2-200 Hz ΔMF). The lowest 40% of IC thresholds were less than or equal to psychometric thresholds, regardless of which code was used, whereas CN thresholds were greater than psychometric thresholds. Discrimination at 10-20 Hz could be explained by indiscriminately pooling 30 units in either structure, whereas discrimination at higher MFs was best explained by more selective pooling. This suggests that pooled CN activity was sufficient for AM discrimination. Psychometric and neurometric thresholds decreased as stimulus duration increased, but IC and CN thresholds were higher and more variable than behavior at short durations. This slower subcortical temporal integration compared with behavior was consistent with a drift diffusion model that reproduced individual differences in performance and can constrain future neurophysiological studies of temporal integration. These measures provide an account of AM perception at the neurophysiological, computational, and behavioral levels.NEW & NOTEWORTHY In everyday environments, the brain is tasked with extracting information from sound envelopes, which involves both sensory encoding and perceptual decision-making. Different neural codes for envelope representation have been characterized in midbrain and cortex, but studies of brainstem nuclei such as the cochlear nucleus (CN) have usually been conducted under anesthesia in nonprimate species. Here, we found that subcortical activity in awake monkeys and a biologically plausible perceptual decision-making model accounted for sound envelope discrimination behavior.
Collapse
Affiliation(s)
- Chase A Mackey
- Neuroscience Graduate Program, Vanderbilt University, Nashville, Tennessee, United States
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, Tennessee, United States
| | - Samantha Hauser
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, Tennessee, United States
| | - Adriana M Schoenhaut
- Neuroscience Graduate Program, Vanderbilt University, Nashville, Tennessee, United States
| | - Namrata Temghare
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, Tennessee, United States
| | - Ramnarayan Ramachandran
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, Tennessee, United States
| |
Collapse
|
4
|
Guest DR, Rajappa N, Oxenham AJ. Limitations in human auditory spectral analysis at high frequencies. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 156:326-340. [PMID: 38990035 PMCID: PMC11240212 DOI: 10.1121/10.0026475] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/18/2023] [Revised: 06/04/2024] [Accepted: 06/07/2024] [Indexed: 07/12/2024]
Abstract
Humans are adept at identifying spectral patterns, such as vowels, in different rooms, at different sound levels, or produced by different talkers. How this feat is achieved remains poorly understood. Two psychoacoustic analogs of spectral pattern recognition are spectral profile analysis and spectrotemporal ripple direction discrimination. This study tested whether pattern-recognition abilities observed previously at low frequencies are also observed at extended high frequencies. At low frequencies (center frequency ∼500 Hz), listeners were able to achieve accurate profile-analysis thresholds, consistent with prior literature. However, at extended high frequencies (center frequency ∼10 kHz), listeners' profile-analysis thresholds were either unmeasurable or could not be distinguished from performance based on overall loudness cues. A similar pattern of results was observed with spectral ripple discrimination, where performance was again considerably better at low than at high frequencies. Collectively, these results suggest a severe deficit in listeners' ability to analyze patterns of intensity across frequency in the extended high-frequency region that cannot be accounted for by cochlear frequency selectivity. One interpretation is that the auditory system is not optimized to analyze such fine-grained across-frequency profiles at extended high frequencies, as they are not typically informative for everyday sounds.
Collapse
Affiliation(s)
- Daniel R Guest
- Department of Biomedical Engineering, University of Rochester, Rochester, New York 14642, USA
| | - Neha Rajappa
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455, USA
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455, USA
| |
Collapse
|
5
|
Carney LH. Neural Fluctuation Contrast as a Code for Complex Sounds: The Role and Control of Peripheral Nonlinearities. Hear Res 2024; 443:108966. [PMID: 38310710 PMCID: PMC10923127 DOI: 10.1016/j.heares.2024.108966] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/30/2023] [Revised: 01/14/2024] [Accepted: 01/26/2024] [Indexed: 02/06/2024]
Abstract
The nonlinearities of the inner ear are often considered to be obstacles that the central nervous system has to overcome to decode neural responses to sounds. This review describes how peripheral nonlinearities, such as saturation of the inner-hair-cell response and of the IHC-auditory-nerve synapse, are instead beneficial to the neural encoding of complex sounds such as speech. These nonlinearities set up contrast in the depth of neural-fluctuations in auditory-nerve responses along the tonotopic axis, referred to here as neural fluctuation contrast (NFC). Physiological support for the NFC coding hypothesis is reviewed, and predictions of several psychophysical phenomena, including masked detection and speech intelligibility, are presented. Lastly, a framework based on the NFC code for understanding how the medial olivocochlear (MOC) efferent system contributes to the coding of complex sounds is presented. By modulating cochlear gain control in response to both sound energy and fluctuations in neural responses, the MOC system is hypothesized to function not as a simple feedback gain-control device, but rather as a mechanism for enhancing NFC along the tonotopic axis, enabling robust encoding of complex sounds across a wide range of sound levels and in the presence of background noise. Effects of sensorineural hearing loss on the NFC code and on the MOC feedback system are presented and discussed.
Collapse
Affiliation(s)
- Laurel H Carney
- Depts. of Biomedical Engineering, Neuroscience, and Electrical & Computer Engineering University of Rochester, Rochester, NY, USA.
| |
Collapse
|
6
|
Alamri Y, Jennings SG. Computational modeling of the human compound action potential. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 153:2376. [PMID: 37092943 PMCID: PMC10119875 DOI: 10.1121/10.0017863] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/23/2022] [Revised: 03/21/2023] [Accepted: 04/04/2023] [Indexed: 05/03/2023]
Abstract
The auditory nerve (AN) compound action potential (CAP) is an important tool for assessing auditory disorders and monitoring the health of the auditory periphery during surgical procedures. The CAP has been mathematically conceptualized as the convolution of a unit response (UR) waveform with the firing rate of a population of AN fibers. Here, an approach for predicting experimentally recorded CAPs in humans is proposed, which involves the use of human-based computational models to simulate AN activity. CAPs elicited by clicks, chirps, and amplitude-modulated carriers were simulated and compared with empirically recorded CAPs from human subjects. In addition, narrowband CAPs derived from noise-masked clicks and tone bursts were simulated. Many morphological, temporal, and spectral aspects of human CAPs were captured by the simulations for all stimuli tested. These findings support the use of model simulations of the human CAP to refine existing human-based models of the auditory periphery, aid in the design and analysis of auditory experiments, and predict the effects of hearing loss, synaptopathy, and other auditory disorders on the human CAP.
Collapse
Affiliation(s)
- Yousef Alamri
- Department of Biomedical Engineering, The University of Utah, 390 South, 1530 East, BEHS 1201, Salt Lake City, Utah 84112, USA
| | - Skyler G Jennings
- Department of Communication Sciences and Disorders, The University of Utah, 390 South, 1530 East, BEHS 1201, Salt Lake City, Utah 84112, USA
| |
Collapse
|
7
|
Zaar J, Carney LH. Predicting speech intelligibility in hearing-impaired listeners using a physiologically inspired auditory model. Hear Res 2022; 426:108553. [PMID: 35750575 PMCID: PMC10560534 DOI: 10.1016/j.heares.2022.108553] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/14/2022] [Revised: 05/17/2022] [Accepted: 06/08/2022] [Indexed: 11/18/2022]
Abstract
This study presents a major update and full evaluation of a speech intelligibility (SI) prediction model previously introduced by Scheidiger, Carney, Dau, and Zaar [(2018), Acta Acust. United Ac. 104, 914-917]. The model predicts SI in speech-in-noise conditions via comparison of the noisy speech and the noise-alone reference. The two signals are processed through a physiologically inspired nonlinear model of the auditory periphery, for a range of characteristic frequencies (CFs), followed by a modulation analysis in the range of the fundamental frequency of speech. The decision metric of the model is the mean of a series of short-term, across-CF correlations between population responses to noisy speech and noise alone, with a sensitivity-limitation process imposed. The decision metric is assumed to be inversely related to SI and is converted to a percent-correct score using a single data-based fitting function. The model performance was evaluated in conditions of stationary, fluctuating, and speech-like interferers using sentence-based speech-reception thresholds (SRTs) previously obtained in 5 normal-hearing (NH) and 13 hearing-impaired (HI) listeners. For the NH listener group, the model accurately predicted SRTs across the different acoustic conditions (apart from a slight overestimation of the masking release observed for fluctuating maskers), as well as plausible effects in response to changes in presentation level. For HI listeners, the model was adjusted to account for the individual audiograms using standard assumptions concerning the amount of HI attributed to inner-hair-cell (IHC) and outer-hair-cell (OHC) impairment. HI model results accounted remarkably well for elevated individual SRTs and reduced masking release. Furthermore, plausible predictions of worsened SI were obtained when the relative contribution of IHC impairment to HI was increased. Overall, the present model provides a useful tool to accurately predict speech-in-noise outcomes in NH and HI listeners, and may yield important insights into auditory processes that are crucial for speech understanding.
Collapse
Affiliation(s)
- Johannes Zaar
- Eriksholm Research Centre, DK-3070 Snekkersten, Denmark; Hearing Systems Section, Department of Health Technology, Technical University of Denmark, DK-2800 Kgs. Lyngby, Denmark.
| | - Laurel H Carney
- Departments of Biomedical Engineering and Neuroscience, University of Rochester, Rochester, NY, 14642, USA
| |
Collapse
|
8
|
Osses Vecchi A, Varnet L, Carney LH, Dau T, Bruce IC, Verhulst S, Majdak P. A comparative study of eight human auditory models of monaural processing. ACTA ACUSTICA. EUROPEAN ACOUSTICS ASSOCIATION 2022; 6:17. [PMID: 36325461 PMCID: PMC9625898 DOI: 10.1051/aacus/2022008] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/19/2023]
Abstract
A number of auditory models have been developed using diverging approaches, either physiological or perceptual, but they share comparable stages of signal processing, as they are inspired by the same constitutive parts of the auditory system. We compare eight monaural models that are openly accessible in the Auditory Modelling Toolbox. We discuss the considerations required to make the model outputs comparable to each other, as well as the results for the following model processing stages or their equivalents: Outer and middle ear, cochlear filter bank, inner hair cell, auditory nerve synapse, cochlear nucleus, and inferior colliculus. The discussion includes a list of recommendations for future applications of auditory models.
Collapse
Affiliation(s)
- Alejandro Osses Vecchi
- Laboratoire des systèmes perceptifs, Département d’études cognitives, École Normale Supérieure, PSL University, CNRS, 75005 Paris, France
| | - Léo Varnet
- Laboratoire des systèmes perceptifs, Département d’études cognitives, École Normale Supérieure, PSL University, CNRS, 75005 Paris, France
| | - Laurel H. Carney
- Departments of Biomedical Engineering and Neuroscience, University of Rochester, Rochester, NY 14642, USA
| | - Torsten Dau
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, DK-2800 Kgs. Lyngby, Denmark
| | - Ian C. Bruce
- Department of Electrical and Computer Engineering, McMaster University, Hamilton, ON L8S 4K1, Canada
| | - Sarah Verhulst
- Hearing Technology group, WAVES, Department of Information Technology, Ghent University, 9000 Ghent, Belgium
| | - Piotr Majdak
- Acoustics Research Institute, Austrian Academy of Sciences, 1040 Vienna, Austria
| |
Collapse
|
9
|
Temporal integration contributes to the masking release by amplitude modulation. Hear Res 2022; 420:108514. [DOI: 10.1016/j.heares.2022.108514] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/24/2021] [Revised: 03/29/2022] [Accepted: 05/04/2022] [Indexed: 11/18/2022]
|
10
|
Abstract
Hearing in noise is a core problem in audition, and a challenge for hearing-impaired listeners, yet the underlying mechanisms are poorly understood. We explored whether harmonic frequency relations, a signature property of many communication sounds, aid hearing in noise for normal hearing listeners. We measured detection thresholds in noise for tones and speech synthesized to have harmonic or inharmonic spectra. Harmonic signals were consistently easier to detect than otherwise identical inharmonic signals. Harmonicity also improved discrimination of sounds in noise. The largest benefits were observed for two-note up-down "pitch" discrimination and melodic contour discrimination, both of which could be performed equally well with harmonic and inharmonic tones in quiet, but which showed large harmonic advantages in noise. The results show that harmonicity facilitates hearing in noise, plausibly by providing a noise-robust pitch cue that aids detection and discrimination.
Collapse
|
11
|
Guest DR, Oxenham AJ. Human discrimination and modeling of high-frequency complex tones shed light on the neural codes for pitch. PLoS Comput Biol 2022; 18:e1009889. [PMID: 35239639 PMCID: PMC8923464 DOI: 10.1371/journal.pcbi.1009889] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2021] [Revised: 03/15/2022] [Accepted: 02/02/2022] [Indexed: 11/24/2022] Open
Abstract
Accurate pitch perception of harmonic complex tones is widely believed to rely on temporal fine structure information conveyed by the precise phase-locked responses of auditory-nerve fibers. However, accurate pitch perception remains possible even when spectrally resolved harmonics are presented at frequencies beyond the putative limits of neural phase locking, and it is unclear whether residual temporal information, or a coarser rate-place code, underlies this ability. We addressed this question by measuring human pitch discrimination at low and high frequencies for harmonic complex tones, presented either in isolation or in the presence of concurrent complex-tone maskers. We found that concurrent complex-tone maskers impaired performance at both low and high frequencies, although the impairment introduced by adding maskers at high frequencies relative to low frequencies differed between the tested masker types. We then combined simulated auditory-nerve responses to our stimuli with ideal-observer analysis to quantify the extent to which performance was limited by peripheral factors. We found that the worsening of both frequency discrimination and F0 discrimination at high frequencies could be well accounted for (in relative terms) by optimal decoding of all available information at the level of the auditory nerve. A Python package is provided to reproduce these results, and to simulate responses to acoustic stimuli from the three previously published models of the human auditory nerve used in our analyses.
Collapse
Affiliation(s)
- Daniel R. Guest
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota, United States of America
| | - Andrew J. Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota, United States of America
| |
Collapse
|
12
|
Altoè A, Charaziak KK, Dewey JB, Moleti A, Sisto R, Oghalai JS, Shera CA. The Elusive Cochlear Filter: Wave Origin of Cochlear Cross-Frequency Masking. J Assoc Res Otolaryngol 2021; 22:623-640. [PMID: 34677710 DOI: 10.1007/s10162-021-00814-2] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2020] [Accepted: 08/23/2021] [Indexed: 11/28/2022] Open
Abstract
The mammalian cochlea achieves its remarkable sensitivity, frequency selectivity, and dynamic range by spatially segregating the different frequency components of sound via nonlinear processes that remain only partially understood. As a consequence of the wave-based nature of cochlear processing, the different frequency components of complex sounds interact spatially and nonlinearly, mutually suppressing one another as they propagate. Because understanding nonlinear wave interactions and their effects on hearing appears to require mathematically complex or computationally intensive models, theories of hearing that do not deal specifically with cochlear mechanics have often neglected the spatial nature of suppression phenomena. Here we describe a simple framework consisting of a nonlinear traveling-wave model whose spatial response properties can be estimated from basilar-membrane (BM) transfer functions. Without invoking jazzy details of organ-of-Corti mechanics, the model accounts well for the peculiar frequency-dependence of suppression found in two-tone suppression experiments. In particular, our analysis shows that near the peak of the traveling wave, the amplitude of the BM response depends primarily on the nonlinear properties of the traveling wave in more basal (high-frequency) regions. The proposed framework provides perhaps the simplest representation of cochlear signal processing that accounts for the spatially distributed effects of nonlinear wave propagation. Shifting the perspective from local filters to non-local, spatially distributed processes not only elucidates the character of cochlear signal processing, but also has important consequences for interpreting psychophysical experiments.
Collapse
Affiliation(s)
- Alessandro Altoè
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA.
| | - Karolina K Charaziak
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA
| | - James B Dewey
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA
| | - Arturo Moleti
- Department of Physics, University of Roma Tor Vergata, Rome, Italy
| | - Renata Sisto
- DIMEILA, INAIL, Monte Porzio Catone, Rome, Italy
| | - John S Oghalai
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA
| | - Christopher A Shera
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA.,Department of Physics and Astronomy, University of Southern California, Los Angeles, CA, USA
| |
Collapse
|
13
|
Fan L, Henry KS, Carney LH. Responses to diotic tone-in-noise stimuli in the inferior colliculus: stimulus envelope and neural fluctuation cues. Hear Res 2021; 409:108328. [PMID: 34391193 PMCID: PMC8419138 DOI: 10.1016/j.heares.2021.108328] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/11/2021] [Revised: 07/16/2021] [Accepted: 07/27/2021] [Indexed: 11/30/2022]
Abstract
Human detection thresholds in tone-in-noise (TIN) paradigms cannot be explained by the prevalent power-spectrum model when stimulus energy is made less reliable, e.g., in roving-level or equal-energy paradigms. Envelope-related cues provide an alternative that is more robust across level. The TIN stimulus envelope is encoded by slow fluctuations in auditory-nerve (AN) responses - a temporal representation affected by inner-hair-cell (IHC) saturation and cochlear compression. Here, envelope-related fluctuations in AN responses were hypothesized to be reflected in responses of neurons in the inferior colliculus (IC), which have average discharge rates that are sensitive to amplitude-modulation (AM) depth and frequency. Responses to tones masked by narrowband gaussian noise (GN) and low-noise noise (LNN) were recorded in the IC of awake rabbits. Fluctuation amplitudes in the stimulus envelope and in model AN responses decrease for GN maskers and increase for LNN upon addition of tones near threshold. Response rates of IC neurons that are excited by AM were expected to be positively correlated with fluctuation amplitudes, whereas rates of neurons suppressed by AM were expected to be negatively correlated. Of neurons with measurable TIN-detection thresholds, most had the predicted changes in rate with increasing tone level for both GN and LNN maskers. Changes in rate with tone level were correlated with envelope sensitivity measured with two methods, including the maximum slopes of modulation transfer functions. IC rate-based thresholds were broadly consistent with published human and rabbit behavioral data. These results highlight the importance of midbrain sensitivity to envelope cues, as represented in peripheral neural fluctuations, for detection of signals in noise.
Collapse
Affiliation(s)
- Langchen Fan
- Department of Biomedical Engineering, University of Rochester, New York, United States.
| | - Kenneth S Henry
- Department of Biomedical Engineering, University of Rochester, New York, United States; Department of Neuroscience, University of Rochester, New York, United States; Department of Otolaryngology, University of Rochester, New York, United States
| | - Laurel H Carney
- Department of Biomedical Engineering, University of Rochester, New York, United States; Department of Neuroscience, University of Rochester, New York, United States
| |
Collapse
|
14
|
Jennings SG. The role of the medial olivocochlear reflex in psychophysical masking and intensity resolution in humans: a review. J Neurophysiol 2021; 125:2279-2308. [PMID: 33909513 PMCID: PMC8285664 DOI: 10.1152/jn.00672.2020] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2020] [Revised: 03/16/2021] [Accepted: 04/02/2021] [Indexed: 02/01/2023] Open
Abstract
This review addresses the putative role of the medial olivocochlear (MOC) reflex in psychophysical masking and intensity resolution in humans. A framework for interpreting psychophysical results in terms of the expected influence of the MOC reflex is introduced. This framework is used to review the effects of a precursor or contralateral acoustic stimulation on 1) simultaneous masking of brief tones, 2) behavioral estimates of cochlear gain and frequency resolution in forward masking, 3) the buildup and decay of forward masking, and 4) measures of intensity resolution. Support, or lack thereof, for a role of the MOC reflex in psychophysical perception is discussed in terms of studies on estimates of MOC strength from otoacoustic emissions and the effects of resection of the olivocochlear bundle in patients with vestibular neurectomy. Novel, innovative approaches are needed to resolve the dissatisfying conclusion that current results are unable to definitively confirm or refute the role of the MOC reflex in masking and intensity resolution.
Collapse
Affiliation(s)
- Skyler G Jennings
- Department of Communication Sciences and Disorders, The University of Utah, Salt Lake City, Utah
| |
Collapse
|
15
|
Osses Vecchi A, Kohlrausch A. Perceptual similarity between piano notes: Simulations with a template-based perception model. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:3534. [PMID: 34241098 DOI: 10.1121/10.0004818] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/31/2020] [Accepted: 04/11/2021] [Indexed: 06/13/2023]
Abstract
In this paper, the auditory model developed by Dau, Kollmeier, and Kohlrausch [(1997). J. Acoust. Soc. Am. 102, 2892-2905] was used to simulate the perceptual similarity between complex sounds. As complex sounds, a set of piano recordings was used, whose perceptual similarity has recently been measured by Osses, Kohlrausch, and Chaigne [(2019). J. Acoust. Soc. Am. 146, 1024-1035] using a three-alternative forced-choice discrimination task in noise. To simulate this discrimination task, the auditory model required a new back-end stage, the central processor, which is preceded by several processing stages that are to a greater or lesser extent inspired by physiological aspects of the normal-hearing system. Therefore, a comprehensive review of the model parameters as used in the literature is given, indicating the fixed set of parameter values that is used in all simulations. Due to the perceptual relevance of the piano note onsets, this review includes an in-depth description of the auditory adaptation stage, the adaptation loops. A moderate to high correlation was found between the simulation results and existing experimental data.
Collapse
Affiliation(s)
- Alejandro Osses Vecchi
- Human-Technology Interaction Group, Department of Industrial Engineering and Innovation Sciences, Eindhoven University of Technology, 5600MB Eindhoven, The Netherlands
| | - Armin Kohlrausch
- Human-Technology Interaction Group, Department of Industrial Engineering and Innovation Sciences, Eindhoven University of Technology, 5600MB Eindhoven, The Netherlands
| |
Collapse
|
16
|
de Cheveigné A. Harmonic Cancellation-A Fundamental of Auditory Scene Analysis. Trends Hear 2021; 25:23312165211041422. [PMID: 34698574 PMCID: PMC8552394 DOI: 10.1177/23312165211041422] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2021] [Revised: 07/23/2021] [Accepted: 07/09/2021] [Indexed: 11/16/2022] Open
Abstract
This paper reviews the hypothesis of harmonic cancellation according to which an interfering sound is suppressed or canceled on the basis of its harmonicity (or periodicity in the time domain) for the purpose of Auditory Scene Analysis. It defines the concept, discusses theoretical arguments in its favor, and reviews experimental results that support it, or not. If correct, the hypothesis may draw on time-domain processing of temporally accurate neural representations within the brainstem, as required also by the classic equalization-cancellation model of binaural unmasking. The hypothesis predicts that a target sound corrupted by interference will be easier to hear if the interference is harmonic than inharmonic, all else being equal. This prediction is borne out in a number of behavioral studies, but not all. The paper reviews those results, with the aim to understand the inconsistencies and come up with a reliable conclusion for, or against, the hypothesis of harmonic cancellation within the auditory system.
Collapse
Affiliation(s)
- Alain de Cheveigné
- Laboratoire des systèmes perceptifs, CNRS, Paris, France
- Département d’études cognitives, École normale supérieure, PSL
University, Paris, France
- UCL Ear Institute, London, UK
| |
Collapse
|