1
|
Irino T, Yokota K, Patterson RD. Improving Auditory Filter Estimation by Incorporating Absolute Threshold and a Level-dependent Internal Noise. Trends Hear 2023; 27:23312165231209750. [PMID: 37905400 PMCID: PMC10619342 DOI: 10.1177/23312165231209750] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/24/2022] [Accepted: 10/07/2023] [Indexed: 11/02/2023] Open
Abstract
Auditory filter (AF) shape has traditionally been estimated with a combination of a notched-noise (NN) masking experiment and a power spectrum model (PSM) of masking. However, there are several challenges that remain in both the simultaneous and forward masking paradigms. We hypothesized that AF shape estimation would be improved if absolute threshold (AT) and a level-dependent internal noise were explicitly represented in the PSM. To document the interaction between NN threshold and AT in normal hearing (NH) listeners, a large set of NN thresholds was measured at four center frequencies (500, 1000, 2000, and 4000 Hz) with the emphasis on low-level maskers. The proposed PSM, consisting of the compressive gammachirp (cGC) filter and three nonfilter parameters, allowed AF estimation over a wide range of frequencies and levels with fewer coefficients and less error than previous models. The results also provided new insights into the nonfilter parameters. The detector signal-to-noise ratio (K ) was found to be constant across signal frequencies, suggesting that no frequency dependence hypothesis is required in the postfiltering process. The ANSI standard "Hearing Level-0dB" function, i.e., AT of NH listeners, could be applied to the frequency distribution of the noise floor for the best AF estimation. The introduction of a level-dependent internal noise could mitigate the nonlinear effects that occur in the simultaneous NN masking paradigm. The new PSM improves the applicability of the model, particularly when the sound pressure level of the NN threshold is close to AT.
Collapse
Affiliation(s)
- Toshio Irino
- Faculty of Systems Engineering, Wakayama University, Japan
| | - Kenji Yokota
- Faculty of Systems Engineering, Wakayama University, Japan
| | - Roy D. Patterson
- Department of Physiology, Development and Neuroscience, University
of Cambridge, UK
| |
Collapse
|
2
|
Leschke J, Orellana GR, Shera CA, Oxenham AJ. Auditory filter shapes derived from forward and simultaneous masking at low frequencies: Implications for human cochlear tuning. Hear Res 2022; 420:108500. [PMID: 35405591 PMCID: PMC9167757 DOI: 10.1016/j.heares.2022.108500] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/17/2021] [Revised: 03/08/2022] [Accepted: 03/28/2022] [Indexed: 01/04/2023]
Abstract
Behavioral forward-masking thresholds with a spectrally notched-noise masker and a fixed low-level probe tone have been shown to provide accurate estimates of cochlear tuning. Estimates using simultaneous masking are similar but generally broader, presumably due to nonlinear cochlear suppression effects. So far, estimates with forward masking have been limited to frequencies of 1 kHz and above. This study used spectrally notched noise under forward and simultaneous masking to estimate frequency selectivity between 200 and 1000 Hz for young adult listeners with normal hearing. Estimates of filter tuning at 1000 Hz were in agreement with previous studies. Estimated tuning broadened below 1000 Hz, with the filter quality factor based on the equivalent rectangular bandwidth (QERB) decreasing more rapidly with decreasing frequency than predicted by previous equations, in line with earlier predictions based on otoacoustic-emission latencies. Estimates from simultaneous masking remained broader than those from forward masking by approximately the same ratio. The new data provide a way to compare human cochlear tuning estimates with auditory-nerve tuning curves from other species across most of the auditory frequency range.
Collapse
Affiliation(s)
| | | | - Christopher A. Shera
- Auditory Research Center, Caruso Department of Otolaryngology, University of Southern California, Los Angeles, CA 90033
| | - Andrew J. Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, MN 55455
| |
Collapse
|
3
|
Perez-Flores MC, Verschooten E, Lee JH, Kim HJ, Joris PX, Yamoah EN. Intrinsic mechanical sensitivity of mammalian auditory neurons as a contributor to sound-driven neural activity. eLife 2022; 11:74948. [PMID: 35266451 PMCID: PMC8942473 DOI: 10.7554/elife.74948] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2021] [Accepted: 03/09/2022] [Indexed: 11/18/2022] Open
Abstract
Mechanosensation – by which mechanical stimuli are converted into a neuronal signal – is the basis for the sensory systems of hearing, balance, and touch. Mechanosensation is unmatched in speed and its diverse range of sensitivities, reaching its highest temporal limits with the sense of hearing; however, hair cells (HCs) and the auditory nerve (AN) serve as obligatory bottlenecks for sounds to engage the brain. Like other sensory neurons, auditory neurons use the canonical pathway for neurotransmission and millisecond-duration action potentials (APs). How the auditory system utilizes the relatively slow transmission mechanisms to achieve ultrafast speed, and high audio-frequency hearing remains an enigma. Here, we address this paradox and report that the mouse, and chinchilla, AN are mechanically sensitive, and minute mechanical displacement profoundly affects its response properties. Sound-mimicking sinusoidal mechanical and electrical current stimuli affect phase-locked responses. In a phase-dependent manner, the two stimuli can also evoke suppressive responses. We propose that mechanical sensitivity interacts with synaptic responses to shape responses in the AN, including frequency tuning and temporal phase locking. Combining neurotransmission and mechanical sensation to control spike patterns gives the mammalian AN a secondary receptor role, an emerging theme in primary neuronal functions.
Collapse
Affiliation(s)
| | - Eric Verschooten
- Laboratory of Auditory Neurophysiology, University of Leuven, Leuven, Belgium
| | | | | | - Philip X Joris
- Laboratory of Auditory Neurophysiology, University of Leuven, Leuven, Belgium
| | | |
Collapse
|
4
|
Altoè A, Charaziak KK, Dewey JB, Moleti A, Sisto R, Oghalai JS, Shera CA. The Elusive Cochlear Filter: Wave Origin of Cochlear Cross-Frequency Masking. J Assoc Res Otolaryngol 2021; 22:623-640. [PMID: 34677710 DOI: 10.1007/s10162-021-00814-2] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2020] [Accepted: 08/23/2021] [Indexed: 11/28/2022] Open
Abstract
The mammalian cochlea achieves its remarkable sensitivity, frequency selectivity, and dynamic range by spatially segregating the different frequency components of sound via nonlinear processes that remain only partially understood. As a consequence of the wave-based nature of cochlear processing, the different frequency components of complex sounds interact spatially and nonlinearly, mutually suppressing one another as they propagate. Because understanding nonlinear wave interactions and their effects on hearing appears to require mathematically complex or computationally intensive models, theories of hearing that do not deal specifically with cochlear mechanics have often neglected the spatial nature of suppression phenomena. Here we describe a simple framework consisting of a nonlinear traveling-wave model whose spatial response properties can be estimated from basilar-membrane (BM) transfer functions. Without invoking jazzy details of organ-of-Corti mechanics, the model accounts well for the peculiar frequency-dependence of suppression found in two-tone suppression experiments. In particular, our analysis shows that near the peak of the traveling wave, the amplitude of the BM response depends primarily on the nonlinear properties of the traveling wave in more basal (high-frequency) regions. The proposed framework provides perhaps the simplest representation of cochlear signal processing that accounts for the spatially distributed effects of nonlinear wave propagation. Shifting the perspective from local filters to non-local, spatially distributed processes not only elucidates the character of cochlear signal processing, but also has important consequences for interpreting psychophysical experiments.
Collapse
Affiliation(s)
- Alessandro Altoè
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA.
| | - Karolina K Charaziak
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA
| | - James B Dewey
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA
| | - Arturo Moleti
- Department of Physics, University of Roma Tor Vergata, Rome, Italy
| | - Renata Sisto
- DIMEILA, INAIL, Monte Porzio Catone, Rome, Italy
| | - John S Oghalai
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA
| | - Christopher A Shera
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA.,Department of Physics and Astronomy, University of Southern California, Los Angeles, CA, USA
| |
Collapse
|
5
|
Temporal Correlates to Monaural Edge Pitch in the Distribution of Interspike Interval Statistics in the Auditory Nerve. eNeuro 2021; 8:ENEURO.0292-21.2021. [PMID: 34281977 PMCID: PMC8387151 DOI: 10.1523/eneuro.0292-21.2021] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2021] [Accepted: 07/07/2021] [Indexed: 12/02/2022] Open
Abstract
Pitch is a perceptual attribute enabling perception of melody. There is no consensus regarding the fundamental nature of pitch and its underlying neural code. A stimulus which has received much interest in psychophysical and computational studies is noise with a sharp spectral edge. High-pass (HP) or low-pass (LP) noise gives rise to a pitch near the edge frequency (monaural edge pitch; MEP). The simplicity of this stimulus, combined with its spectral and autocorrelation properties, make it an interesting stimulus to examine spectral versus temporal cues that could underly its pitch. We recorded responses of single auditory nerve (AN) fibers in chinchilla to MEP-stimuli varying in edge frequency. Temporal cues were examined with shuffled autocorrelogram (SAC) analysis. Correspondence between the population’s dominant interspike interval and reported pitch estimates was poor. A fuller analysis of the population interspike interval distribution, which incorporates not only the dominant but all intervals, results in good matches with behavioral results, but not for the entire range of edge frequencies that generates pitch. Finally, we also examined temporal structure over a slower time scale, intermediate between average firing rate and interspike intervals, by studying the SAC envelope. We found that, in response to a given MEP stimulus, this feature also systematically varies with edge frequency, across fibers with different characteristic frequency (CF). Because neural mechanisms to extract envelope cues are well established, and because this cue is not limited by coding of stimulus fine-structure, this newly identified slower temporal cue is a more plausible basis for pitch than cues based on fine-structure.
Collapse
|
6
|
Vencovský V, Vetešník A, Gummer AW. Nonlinear reflection as a cause of the short-latency component in stimulus-frequency otoacoustic emissions simulated by the methods of compression and suppression. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:3992. [PMID: 32611132 DOI: 10.1121/10.0001394] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/19/2020] [Accepted: 05/22/2020] [Indexed: 06/11/2023]
Abstract
Stimulus-frequency otoacoustic emissions (SFOAEs) are generated by coherent reflection of forward traveling waves by perturbations along the basilar membrane. The strongest wavelets are backscattered near the place where the traveling wave reaches its maximal amplitude (tonotopic place). Therefore, the SFOAE group delay might be expected to be twice the group delay estimated in the cochlear filters. However, experimental data have yielded steady-state SFOAE components with near-zero latency. A cochlear model is used to show that short-latency SFOAE components can be generated due to nonlinear reflection of the compressor or suppressor tones used in SFOAE measurements. The simulations indicate that suppressors produce more pronounced short-latency components than compressors. The existence of nonlinear reflection components due to suppressors can also explain why SFOAEs can still be detected when suppressors are presented more than half an octave above the probe-tone frequency. Simulations of the SFOAE suppression tuning curves showed that phase changes in the SFOAE residual as the suppressor frequency increases are mostly determined by phase changes of the nonlinear reflection component.
Collapse
Affiliation(s)
- Václav Vencovský
- Department of Radioelectronics, Czech Technical University in Prague, Technická 2, 166 27 Prague, Czech Republic
| | - Aleš Vetešník
- Department of Nuclear Chemistry, Czech Technical University in Prague, Břehová 7, 115 19 Prague, Czech Republic
| | - Anthony W Gummer
- Department of Otolaryngology, Section of Physiological Acoustics and Communication, Eberhard-Karls-University Tübingen, Elfriede-Aulhorn-Strasse 5, 72076 Tübingen, Germany
| |
Collapse
|
7
|
Charaziak KK, Dong W, Altoè A, Shera CA. Asymmetry and Microstructure of Temporal-Suppression Patterns in Basilar-Membrane Responses to Clicks: Relation to Tonal Suppression and Traveling-Wave Dispersion. J Assoc Res Otolaryngol 2020; 21:151-170. [PMID: 32166602 DOI: 10.1007/s10162-020-00747-2] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2019] [Accepted: 02/13/2020] [Indexed: 10/24/2022] Open
Abstract
The cochlea's wave-based signal processing allows it to efficiently decompose a complex acoustic waveform into frequency components. Because cochlear responses are nonlinear, the waves arising from one frequency component of a complex sound can be altered by the presence of others that overlap with it in time and space (e.g., two-tone suppression). Here, we investigate the suppression of basilar-membrane (BM) velocity responses to a transient signal (a test click) by another click or tone. We show that the BM response to the click can be reduced when the stimulus is shortly preceded or followed by another (suppressor) click. More surprisingly, the data reveal two curious dependencies on the interclick interval, Δt. First, the temporal suppression curve (amount of suppression vs. Δt) manifests a pronounced and nearly periodic microstructure. Second, temporal suppression is generally strongest not when the two clicks are presented simultaneously (Δt = 0), but when the suppressor click precedes the test click by a time interval corresponding to one to two periods of the best frequency (BF) at the measurement location. By systematically varying the phase of the suppressor click, we demonstrate that the suppression microstructure arises from alternating constructive and destructive interference between the BM responses to the two clicks. And by comparing temporal and tonal suppression in the same animals, we test the hypothesis that the asymmetry of the temporal-suppression curve around Δt = 0 stems from cochlear dispersion and the well-known asymmetry of tonal suppression around the BF. Just as for two-tone suppression, BM responses to clicks are most suppressed by tones at frequencies just above the BF of the measurement location. On average, the frequency place of maximal suppressibility of the click response predicted from temporal-suppression data agrees with the frequency at which tonal suppression peaks, consistent with our hypothesis.
Collapse
Affiliation(s)
- Karolina K Charaziak
- Caruso Department of Otolaryngology, University of Southern California, Los Angeles, CA, USA.
| | - Wei Dong
- Research Service, VA Loma Linda Healthcare System, Loma Linda, CA, USA.,Department of Otolaryngology-Head & Neck Surgery, Loma Linda University Health, Loma Linda, USA
| | - Alessandro Altoè
- Caruso Department of Otolaryngology, University of Southern California, Los Angeles, CA, USA
| | - Christopher A Shera
- Caruso Department of Otolaryngology, University of Southern California, Los Angeles, CA, USA.,Department of Physics and Astronomy, University of Southern California, Los Angeles, CA, USA
| |
Collapse
|
8
|
Rasetshwane DM, Bosen EC, Kopun JG, Neely ST. Comparison of distortion-product otoacoustic emission and stimulus-frequency otoacoustic emission two-tone suppression in humans. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:4481. [PMID: 31893726 PMCID: PMC6930138 DOI: 10.1121/1.5139660] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/15/2019] [Revised: 11/19/2019] [Accepted: 11/20/2019] [Indexed: 06/10/2023]
Abstract
Distortion-product otoacoustic emission (DPOAE) and stimulus-frequency otoacoustic emission (SFOAE) are two types of acoustic signals emitted by the inner ear in response to tonal stimuli. The levels of both emission types may be reduced by the inclusion of additional (suppressor) tones with the stimulus. Comparison of two-tone suppression properties across emission type addresses a clinically relevant question of whether these two types of emission provide similar information about cochlear status. The purpose of this study was to compare DPOAE suppression to SFOAE suppression from the same ear in a group of participants with normal hearing. Probe frequency was approximately 1000 Hz, and the suppressor frequency varied from -1.5 to 0.5 octaves relative to the probe frequency. DPOAE and SFOAE suppression were compared in terms of (1) suppression growth rate (SGR), (2) superimposed suppression tuning curves (STCs), and (3) STC-derived metrics, such as high-frequency slope, cochlear amplifier gain, and QERB (ERB, equivalent rectangular bandwidth). Below the probe frequency, the SGR was slightly greater than one for SFOAEs and slightly less than two for DPOAEs. There were no differences in STC metrics across emission types. These observations may provide useful constraints on physiology-based models of otoacoustic emission suppression.
Collapse
Affiliation(s)
- Daniel M Rasetshwane
- Center for Hearing Research, Boys Town National Research Hospital, 555 North 30th Street, Omaha, Nebraska 68131, USA
| | - Emily C Bosen
- Center for Hearing Research, Boys Town National Research Hospital, 555 North 30th Street, Omaha, Nebraska 68131, USA
| | - Judy G Kopun
- Center for Hearing Research, Boys Town National Research Hospital, 555 North 30th Street, Omaha, Nebraska 68131, USA
| | - Stephen T Neely
- Center for Hearing Research, Boys Town National Research Hospital, 555 North 30th Street, Omaha, Nebraska 68131, USA
| |
Collapse
|
9
|
Naert G, Pasdelou MP, Le Prell CG. Use of the guinea pig in studies on the development and prevention of acquired sensorineural hearing loss, with an emphasis on noise. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:3743. [PMID: 31795705 PMCID: PMC7195866 DOI: 10.1121/1.5132711] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/10/2019] [Revised: 07/30/2019] [Accepted: 08/12/2019] [Indexed: 05/10/2023]
Abstract
Guinea pigs have been used in diverse studies to better understand acquired hearing loss induced by noise and ototoxic drugs. The guinea pig has its best hearing at slightly higher frequencies relative to humans, but its hearing is more similar to humans than the rat or mouse. Like other rodents, it is more vulnerable to noise injury than the human or nonhuman primate models. There is a wealth of information on auditory function and vulnerability of the inner ear to diverse insults in the guinea pig. With respect to the assessment of potential otoprotective agents, guinea pigs are also docile animals that are relatively easy to dose via systemic injections or gavage. Of interest, the cochlea and the round window are easily accessible, notably for direct cochlear therapy, as in the chinchilla, making the guinea pig a most relevant and suitable model for hearing. This article reviews the use of the guinea pig in basic auditory research, provides detailed discussion of its use in studies on noise injury and other injuries leading to acquired sensorineural hearing loss, and lists some therapeutics assessed in these laboratory animal models to prevent acquired sensorineural hearing loss.
Collapse
Affiliation(s)
| | | | - Colleen G Le Prell
- School of Behavioral and Brain Sciences, University of Texas at Dallas, Dallas, Texas 75080, USA
| |
Collapse
|
10
|
Exploring the Role of Medial Olivocochlear Efferents on the Detection of Amplitude Modulation for Tones Presented in Noise. J Assoc Res Otolaryngol 2019; 20:395-413. [PMID: 31140010 PMCID: PMC6646499 DOI: 10.1007/s10162-019-00722-6] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/21/2018] [Accepted: 04/15/2019] [Indexed: 10/26/2022] Open
Abstract
The medial olivocochlear reflex has been hypothesized to improve the detection and discrimination of dynamic signals in noisy backgrounds. This hypothesis was tested here by comparing behavioral outcomes with otoacoustic emissions. The effects of a precursor on amplitude-modulation (AM) detection were measured for a 1- and 6-kHz carrier at levels of 40, 60, and 80 dB SPL in a two-octave-wide noise masker with a level designed to produce poor, but above-chance, performance. Three types of precursor were used: a two-octave noise band, an inharmonic complex tone, and a pure tone. Precursors had the same overall level as the simultaneous noise masker that immediately followed the precursor. The noise precursor produced a large improvement in AM detection for both carrier frequencies and at all three levels. The complex tone produced a similarly large improvement in AM detection at the highest level but had a smaller effect for the two lower carrier levels. The tonal precursor did not significantly affect AM detection in noise. Comparisons of behavioral thresholds and medial olivocochlear efferent effects on stimulus frequency otoacoustic emissions measured with similar stimuli did not support the hypothesis that efferent-based reduction of cochlear responses contributes to the precursor effects on AM detection.
Collapse
|
11
|
Probing hair cell's mechano-transduction using two-tone suppression measurements. Sci Rep 2019; 9:4626. [PMID: 30874606 PMCID: PMC6420497 DOI: 10.1038/s41598-019-41112-5] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2018] [Accepted: 03/01/2019] [Indexed: 11/27/2022] Open
Abstract
When two sound tones are delivered to the cochlea simultaneously, they interact with each other in a suppressive way, a phenomenon referred to as two-tone suppression (2TS). This nonlinear response is ascribed to the saturation of the outer hair cell’s mechano-transduction. Thus, 2TS can be used as a non-invasive probe to investigate the fundamental properties of cochlear mechano-transduction. We developed a nonlinear cochlear model in the time domain to interpret 2TS data. The multi-scale model incorporates cochlear fluid dynamics, organ of Corti (OoC) mechanics and outer hair cell electrophysiology. The model simulations of 2TS show that the threshold amplitudes and rates of low-side suppression are dependent on mechano-transduction properties. By comparing model responses to existing 2TS measurement data, we estimate intrinsic characteristics of mechano-transduction such as sensitivity and adaptation. For mechano-transduction sensitivity at the basal location (characteristic frequency of 17 kHz) at 0.06 nm−1, the simulation results agree with 2TS measurements of basilar membrane responses. This estimate is an order of magnitude higher than the values observed in experiments on isolated outer hair cells. The model also demonstrates how the outer hair cell’s adaptation alters the temporal pattern of 2TS by modulating mechano-electrical gain and phase.
Collapse
|
12
|
Yellamsetty A, Bidelman GM. Brainstem correlates of concurrent speech identification in adverse listening conditions. Brain Res 2019; 1714:182-192. [PMID: 30796895 DOI: 10.1016/j.brainres.2019.02.025] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2018] [Revised: 01/07/2019] [Accepted: 02/19/2019] [Indexed: 01/20/2023]
Abstract
When two voices compete, listeners can segregate and identify concurrent speech sounds using pitch (fundamental frequency, F0) and timbre (harmonic) cues. Speech perception is also hindered by the signal-to-noise ratio (SNR). How clear and degraded concurrent speech sounds are represented at early, pre-attentive stages of the auditory system is not well understood. To this end, we measured scalp-recorded frequency-following responses (FFR) from the EEG while human listeners heard two concurrently presented, steady-state (time-invariant) vowels whose F0 differed by zero or four semitones (ST) presented diotically in either clean (no noise) or noise-degraded (+5dB SNR) conditions. Listeners also performed a speeded double vowel identification task in which they were required to identify both vowels correctly. Behavioral results showed that speech identification accuracy increased with F0 differences between vowels, and this perceptual F0 benefit was larger for clean compared to noise degraded (+5dB SNR) stimuli. Neurophysiological data demonstrated more robust FFR F0 amplitudes for single compared to double vowels and considerably weaker responses in noise. F0 amplitudes showed speech-on-speech masking effects, along with a non-linear constructive interference at 0ST, and suppression effects at 4ST. Correlations showed that FFR F0 amplitudes failed to predict listeners' identification accuracy. In contrast, FFR F1 amplitudes were associated with faster reaction times, although this correlation was limited to noise conditions. The limited number of brain-behavior associations suggests subcortical activity mainly reflects exogenous processing rather than perceptual correlates of concurrent speech perception. Collectively, our results demonstrate that FFRs reflect pre-attentive coding of concurrent auditory stimuli that only weakly predict the success of identifying concurrent speech.
Collapse
Affiliation(s)
- Anusha Yellamsetty
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Department of Communication Sciences & Disorders, University of South Florida, USA.
| | - Gavin M Bidelman
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA; University of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, USA.
| |
Collapse
|
13
|
Hegland EL, Strickland EA. The effects of preceding sound and stimulus duration on measures of suppression in younger and older adults. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 144:3548. [PMID: 30599663 PMCID: PMC6308016 DOI: 10.1121/1.5083824] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/21/2017] [Revised: 11/04/2018] [Accepted: 11/26/2018] [Indexed: 05/19/2023]
Abstract
Despite clinically normal audiometric thresholds, some older adults may experience difficulty in tasks such as understanding speech in a noisy environment. One potential reason may be reduced cochlear nonlinearity. A sensitive measure of cochlear nonlinearity is two-tone suppression, which is a reduction in the auditory system's response to one tone in the presence of a second tone. Previous research has been mixed on whether suppression decreases with age in humans. Studies of efferent cochlear gain reduction also suggest that stimulus duration should be considered in measuring suppression. In the present study, suppression was first measured psychoacoustically using stimuli that were too short to result in gain reduction. The potential effect of efferent cochlear gain reduction was then measured by using longer stimuli and presenting tonal or noise precursors before the shorter stimuli. Younger adults (ages 19-22 yr) and older adults (ages 57+ yr) with clinically normal hearing were tested. Suppression estimates decreased with longer stimuli or preceding sound which included the signal frequency, but did not decrease with preceding sound at the suppressor frequency. On average, the older group had lower suppression than the younger group, but this difference was not statistically significant.
Collapse
Affiliation(s)
- Erica L Hegland
- Department of Speech, Language, and Hearing Sciences, Purdue University, West Lafayette, Indiana 47907, USA
| | - Elizabeth A Strickland
- Department of Speech, Language, and Hearing Sciences, Purdue University, West Lafayette, Indiana 47907, USA
| |
Collapse
|
14
|
The Interplay Between Spike-Time and Spike-Rate Modes in the Auditory Nerve Encodes Tone-In-Noise Threshold. J Neurosci 2018; 38:5727-5738. [PMID: 29793977 DOI: 10.1523/jneurosci.3103-17.2018] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2017] [Revised: 05/04/2018] [Accepted: 05/10/2018] [Indexed: 11/21/2022] Open
Abstract
Auditory nerve fibers (ANFs) encode pure tones through two modes of coding, spike time and spike rate, depending on the tone frequency. In response to a low-frequency tone, ANF firing is phase locked to the sinusoidal waveform. Because time coding vanishes with an increase in the tone frequency, high-frequency tone coding relies on the spike rate of the ANFs. Adding a continuous broadband noise to a tone compresses the rate intensity function of ANFs and shifts its dynamic range toward higher intensities. Therefore, the ANFs with high-threshold/low-spontaneous rate (SR) are thought to contribute to behavioral tone detection in noise. However, this theory relies on the discharge rate of the ANFs. The direct comparison with the masking threshold through spike timing, irrespective of the spontaneous rate, has not so far been investigated. Taking advantage of a unique proxy to quantify the spike synchrony (i.e., the shuffle autocorrelogram), we show in female gerbils that high-SR ANFs are more adapted to encode low-frequency thresholds through temporal code, giving them a strong robustness in noise. By comparing behavioral thresholds measured using prepulse inhibition of the acoustical startle reflex with population thresholds calculated from ANFs pooled per octave band, we show that threshold-based spike timing provides a better estimate of behavioral thresholds in the low-frequency range, whereas the high-frequency behavioral thresholds rely on the spiking rate, particularly in noise. This emphasizes the complementarity of temporal and rate modes to code tone-in-noise thresholds over a large range of frequencies.SIGNIFICANCE STATEMENT There is a general agreement that high-threshold/low-spontaneous rate (SR) auditory nerve fibers (ANFs) are of prime importance for tone detection in noise. However, this theory is based on the discharge rate of the fibers. Comparing the behavioral thresholds and single ANF thresholds shows that this is only true in the high-frequency range of tone stimulations. In the low-frequency range of tones (up to 2.7 kHz in the gerbil), the most sensitive ANFs (high-SR fibers) carry neural information through a spike-timing mode, even for noise in which tones do not induce a noticeable increment in the spike rate. This emphasizes the interplay between spike-time and spike-rate modes in the auditory nerve to encode tone-in-noise threshold over a large range of tone frequencies.
Collapse
|
15
|
Specialization of the auditory system for the processing of bio-sonar information in the frequency domain: Mustached bats. Hear Res 2018; 361:1-22. [DOI: 10.1016/j.heares.2018.01.012] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/23/2017] [Revised: 01/18/2018] [Accepted: 01/22/2018] [Indexed: 11/20/2022]
|
16
|
Milekhina ON, Nechaev DI, Popov VV, Supin AY. Compressive Nonlinearity in the Auditory System: Manifestation in the Action of Complex Sound Signals. BIOL BULL+ 2017. [DOI: 10.1134/s1062359017060073] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|
17
|
Verschooten E, Strickland EA, Verhaert N, Joris PX. Assessment of Ipsilateral Efferent Effects in Human via ECochG. Front Neurosci 2017; 11:331. [PMID: 28642679 PMCID: PMC5462931 DOI: 10.3389/fnins.2017.00331] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2017] [Accepted: 05/26/2017] [Indexed: 11/16/2022] Open
Abstract
Development of electrophysiological means to assess the medial olivocochlear (MOC) system in humans is important to further our understanding of the function of that system and for the refinement and validation of psychoacoustical and otoacoustic emission methods which are thought to probe the MOC. Based on measurements in anesthetized animals it has been hypothesized that the MOC-reflex (MOCR) can enhance the response to signals in noise, and several lines of evidence support such a role in humans. A difficulty in these studies is the isolation of efferent effects. Efferent activation can be triggered by acoustic stimulation of the contralateral or ipsilateral ear, but ipsilateral stimulation is thought to be more effective. However, ipsilateral stimulation complicates interpretation of effects since these sounds can affect the perception of other ipsilateral sounds by mechanisms not involving olivocochlear efferents. We assessed the ipsilaterally evoked MOCR in human using a transtympanic procedure to record mass-potentials from the cochlear promontory or the niche of the round window. Averaged compound action potential (CAP) responses to masked probe tones of 4 kHz with and without a precursor (designed to activate the MOCR but not the stapedius reflex) were extracted with a polarity alternating paradigm. The masker was either a simultaneous narrow band noise masker or a short (20-ms) tonal ON- or OFF-frequency forward masker. The subjects were screened for normal hearing (audiogram, tympanogram, threshold stapedius reflex) and psychoacoustically tested for the presence of a precursor effect. We observed a clear reduction of CAP amplitude by the precursor, for different masking conditions. Even without an MOCR, this is expected because the precursor will affect the response to subsequent stimuli via neural adaptation. To determine whether the precursor also activated the efferent system, we measured the CAP over a range of masker levels, with or without precursor, and for different types of masker. The results show CAP reduction consistent with the type of gain reduction caused by the MOCR. These results generally support psychoacoustical paradigms designed to probe the efferent system as indeed activating the MOCR system, but not all observations are consistent with this mechanism.
Collapse
Affiliation(s)
- Eric Verschooten
- Laboratory of Auditory Neurophysiology, Department of Neurosciences, University of LeuvenLeuven, Belgium
| | - Elizabeth A. Strickland
- Department of Speech, Language, and Hearing Sciences, Purdue UniversityWest Lafayette, IN, United States
| | - Nicolas Verhaert
- ExpORL Research Group, Department of Neurosciences, University of LeuvenLeuven, Belgium
| | - Philip X. Joris
- Laboratory of Auditory Neurophysiology, Department of Neurosciences, University of LeuvenLeuven, Belgium
| |
Collapse
|
18
|
Keine C, Rübsamen R, Englitz B. Inhibition in the auditory brainstem enhances signal representation and regulates gain in complex acoustic environments. eLife 2016; 5. [PMID: 27855778 PMCID: PMC5148601 DOI: 10.7554/elife.19295] [Citation(s) in RCA: 26] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/03/2016] [Accepted: 11/17/2016] [Indexed: 12/30/2022] Open
Abstract
Inhibition plays a crucial role in neural signal processing, shaping and limiting responses. In the auditory system, inhibition already modulates second order neurons in the cochlear nucleus, e.g. spherical bushy cells (SBCs). While the physiological basis of inhibition and excitation is well described, their functional interaction in signal processing remains elusive. Using a combination of in vivo loose-patch recordings, iontophoretic drug application, and detailed signal analysis in the Mongolian Gerbil, we demonstrate that inhibition is widely co-tuned with excitation, and leads only to minor sharpening of the spectral response properties. Combinations of complex stimuli and neuronal input-output analysis based on spectrotemporal receptive fields revealed inhibition to render the neuronal output temporally sparser and more reproducible than the input. Overall, inhibition plays a central role in improving the temporal response fidelity of SBCs across a wide range of input intensities and thereby provides the basis for high-fidelity signal processing.
Collapse
Affiliation(s)
- Christian Keine
- Faculty of Bioscience, Pharmacy and Psychology, University of Leipzig, Leipzig, Germany
| | - Rudolf Rübsamen
- Faculty of Bioscience, Pharmacy and Psychology, University of Leipzig, Leipzig, Germany
| | - Bernhard Englitz
- Department of Neurophysiology, Donders Center for Neuroscience, Radboud University, Nijmegen, Netherlands
| |
Collapse
|
19
|
Wang Y, Gong Q, Zhang T. The influence of probe level on the tuning of stimulus frequency otoacoustic emissions and behavioral test in human. Biomed Eng Online 2016; 15:51. [PMID: 27160830 PMCID: PMC4862048 DOI: 10.1186/s12938-016-0167-0] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2015] [Accepted: 04/28/2016] [Indexed: 12/03/2022] Open
Abstract
Background Frequency selectivity (FS) of the auditory system is established at the level of the cochlea and it is important for the perception of complex sounds. Although direct measurements of cochlear FS require surgical preparation, it can also be estimated with the measurements of otoacoustic emissions or behavioral tests, including stimulus frequency otoacoustic emission suppression tuning curves (SFOAE STCs) or psychophysical tuning curves (PTCs). These two methods result in similar estimates of FS at low probe levels. As the compressive nonlinearity of cochlea is strongly dependent on the stimulus intensity, the sharpness of tuning curves which is relevant to the cochlear nonlinearity will change as a function of probe level. The present study aims to investigate the influence of different probe levels on the relationship between SFOAE STCs and PTCs. Methods The study included 15 young subjects with normal hearing. SFOAE STCs and PTCs were recorded at low and moderate probe levels for frequencies centred at 1, 2, and 4 kHz. The ratio or the difference of the characteristic parameters between the two methods was calculated at each probe level. The effect of probe level on the ratio or the difference between the parameters of SFOAE STCs and PTCs was then statistically analysed. Results The tuning of SFOAE STCs was significantly positively correlated with the tuning of the PTCs at both low and moderate probe levels; yet, at the moderate probe level, the SFOAE STCs were consistently broader than the PTCs. The mean ratio of sharpness of tuning at low probe levels was constantly around 1 while around 1.5 at moderate probe levels. Conclusions Probe level had a significant effect on the sharpness of tuning between the two methods of estimating FS. SFOAE STC seems a good alternative measurement of PTC for FS assessment at low probe levels. At moderate probe levels, SFOAE STC and PTC were not equivalent measures of the FS in terms of their bandwidths. Because SFOAE STCs are not biased by higher levels auditory processing, they may represent cochlear FS better than PTCs.
Collapse
Affiliation(s)
- Yao Wang
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, 100084, China
| | - Qin Gong
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, 100084, China. .,Research Center of Biomedical Engineering, Graduate School at Shenzhen, Tsinghua University, Shenzhen, 518055, China.
| | - Tao Zhang
- Tsinghua National Laboratory for Information Science and Technology (TNList), Tsinghua University, Beijing, 100084, China
| |
Collapse
|
20
|
Hots J, Jarzombek K, Verhey JL. Mid-bandwidth loudness depression in hearing-impaired listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 139:2334. [PMID: 27250129 DOI: 10.1121/1.4947090] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
The loudness of a bandpass-filtered noise depends on its bandwidth. For bandwidths larger than a critical bandwidth, loudness increases as the bandwidth increases, an effect commonly referred to as spectral loudness summation. For bandwidths smaller than the critical bandwidth, it was shown recently for normal-hearing listeners that loudness decreases as the bandwidth increases. This study investigated if listeners with a hearing impairment of primarily cochlear origin also showed this effect. Levels at equal loudness between a 1500-Hz pure-tone reference and noise-band targets centered at 1500 Hz were measured for bandwidths in the range from 15 to 1620 Hz. The reference level was adjusted individually on the basis of the audiogram. The average level difference at equal loudness increased from 0 dB at 15 Hz up to a maximum of about 4 dB at 810 Hz. Thus, the mid-bandwidth loudness depression is also observed for hearing-impaired listeners.
Collapse
Affiliation(s)
- Jan Hots
- Department of Experimental Audiology, Otto von Guericke University Magdeburg, Leipziger Straße 44, 39120 Magdeburg, Germany
| | - Katrin Jarzombek
- Department of Experimental Audiology, Otto von Guericke University Magdeburg, Leipziger Straße 44, 39120 Magdeburg, Germany
| | - Jesko L Verhey
- Department of Experimental Audiology, Otto von Guericke University Magdeburg, Leipziger Straße 44, 39120 Magdeburg, Germany
| |
Collapse
|
21
|
Suppression Measured from Chinchilla Auditory-Nerve-Fiber Responses Following Noise-Induced Hearing Loss: Adaptive-Tracking and Systems-Identification Approaches. ADVANCES IN EXPERIMENTAL MEDICINE AND BIOLOGY 2016; 894:285-295. [PMID: 27080669 PMCID: PMC5069700 DOI: 10.1007/978-3-319-25474-6_30] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 12/23/2022]
Abstract
The compressive nonlinearity of cochlear signal transduction, reflecting outer-hair-cell function, manifests as suppressive spectral interactions; e.g., two-tone suppression. Moreover, for broadband sounds, there are multiple interactions between frequency components. These frequency-dependent nonlinearities are important for neural coding of complex sounds, such as speech. Acoustic-trauma-induced outer-hair-cell damage is associated with loss of nonlinearity, which auditory prostheses attempt to restore with, e.g., "multi-channel dynamic compression" algorithms.Neurophysiological data on suppression in hearing-impaired (HI) mammals are limited. We present data on firing-rate suppression measured in auditory-nerve-fiber responses in a chinchilla model of noise-induced hearing loss, and in normal-hearing (NH) controls at equal sensation level. Hearing-impaired (HI) animals had elevated single-fiber excitatory thresholds (by ~ 20-40 dB), broadened frequency tuning, and reduced-magnitude distortion-product otoacoustic emissions; consistent with mixed inner- and outer-hair-cell pathology. We characterized suppression using two approaches: adaptive tracking of two-tone-suppression threshold (62 NH, and 35 HI fibers), and Wiener-kernel analyses of responses to broadband noise (91 NH, and 148 HI fibers). Suppression-threshold tuning curves showed sensitive low-side suppression for NH and HI animals. High-side suppression thresholds were elevated in HI animals, to the same extent as excitatory thresholds. We factored second-order Wiener-kernels into excitatory and suppressive sub-kernels to quantify the relative strength of suppression. We found a small decrease in suppression in HI fibers, which correlated with broadened tuning. These data will help guide novel amplification strategies, particularly for complex listening situations (e.g., speech in noise), in which current hearing aids struggle to restore intelligibility.
Collapse
|
22
|
Okamoto H, Kakigi R. Encoding of frequency-modulation (FM) rates in human auditory cortex. Sci Rep 2015; 5:18143. [PMID: 26656920 PMCID: PMC4677350 DOI: 10.1038/srep18143] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2015] [Accepted: 11/13/2015] [Indexed: 11/09/2022] Open
Abstract
Frequency-modulated sounds play an important role in our daily social life. However, it currently remains unclear whether frequency modulation rates affect neural activity in the human auditory cortex. In the present study, using magnetoencephalography, we investigated the auditory evoked N1m and sustained field responses elicited by temporally repeated and superimposed frequency-modulated sweeps that were matched in the spectral domain, but differed in frequency modulation rates (1, 4, 16, and 64 octaves per sec). The results obtained demonstrated that the higher rate frequency-modulated sweeps elicited the smaller N1m and the larger sustained field responses. Frequency modulation rate had a significant impact on the human brain responses, thereby providing a key for disentangling a series of natural frequency-modulated sounds such as speech and music.
Collapse
Affiliation(s)
- Hidehiko Okamoto
- Department of Integrative Physiology, National Institute for Physiological Sciences, Okazaki, Japan.,Department of Physiological Sciences, School of Life Science, SOKENDAI (The Graduate University for Advanced Studies), Hayama, Japan
| | - Ryusuke Kakigi
- Department of Integrative Physiology, National Institute for Physiological Sciences, Okazaki, Japan.,Department of Physiological Sciences, School of Life Science, SOKENDAI (The Graduate University for Advanced Studies), Hayama, Japan
| |
Collapse
|
23
|
Feng L, Oxenham AJ. New perspectives on the measurement and time course of auditory enhancement. J Exp Psychol Hum Percept Perform 2015; 41:1696-708. [PMID: 26280269 DOI: 10.1037/xhp0000115] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
A target sound can become more audible and may "pop out" from a simultaneously presented masker if the masker is presented first by itself, as a precursor. This phenomenon, known as auditory enhancement, may reflect the general perceptual principle of contrast enhancement, which facilitates adaptation to ongoing acoustic conditions and the detection of new events. Little is known about the mechanisms underlying enhancement, and potential confounding factors have made the size of the effect and its time course a point of contention. Here we measured enhancement as a function of precursor duration and delay between precursor offset and target onset, using 2 single-interval pitch comparison tasks, which involve either same-different or up-down judgments, to avoid the potential confounds of earlier studies. Although these 2 tasks elicit different levels of performance and may reflect different underlying mechanisms, they produced similar amounts of enhancement. The effect decreased with decreasing precursor duration, but remained present for precursors as short as 62.5 ms, and decreased with increasing gap between the precursor and target, but remained measurable 1 s after the precursor. Additional conditions, examining the effect of precursor/masker similarity and the possible role of grouping and cueing, suggest multiple sources of auditory enhancement.
Collapse
Affiliation(s)
- Lei Feng
- Department of Otolaryngology, University of Minnesota
| | | |
Collapse
|
24
|
Stimulus Frequency Otoacoustic Emissions Provide No Evidence for the Role of Efferents in the Enhancement Effect. J Assoc Res Otolaryngol 2015; 16:613-29. [PMID: 26153415 DOI: 10.1007/s10162-015-0534-8] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2015] [Accepted: 06/18/2015] [Indexed: 10/23/2022] Open
Abstract
Auditory enhancement refers to the perceptual phenomenon that a target sound is heard out more readily from a background sound if the background is presented alone first. Here we used stimulus-frequency otoacoustic emissions (SFOAEs) to test the hypothesis that activation of the medial olivocochlear efferent system contributes to auditory enhancement effects. The SFOAEs were used as a tool to measure changes in cochlear responses to a target component and the neighboring components of a multitone background between conditions producing enhancement and conditions producing no enhancement. In the "enhancement" condition, the target and multitone background were preceded by a precursor stimulus with a spectral notch around the signal frequency; in the control (no-enhancement) condition, the target and multitone background were presented without the precursor. In an experiment using a wideband multitone stimulus known to produce significant psychophysical enhancement effects, SFOAEs showed no changes consistent with enhancement, but some aspects of the results indicated possible contamination of the SFOAE magnitudes by the activation of the middle-ear-muscle reflex. The same SFOAE measurements performed using narrower-band stimuli at lower sound levels also showed no SFOAE changes consistent with either absolute or relative enhancement despite robust psychophysical enhancement effects observed in the same listeners with the same stimuli. The results suggest that cochlear efferent control does not play a significant role in auditory enhancement effects.
Collapse
|
25
|
Poblete V, Espic F, King S, Stern RM, Huenupán F, Fredes J, Yoma NB. A perceptually-motivated low-complexity instantaneous linear channel normalization technique applied to speaker verification. COMPUT SPEECH LANG 2015. [DOI: 10.1016/j.csl.2014.10.006] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
26
|
Pannunzi M, Pérez-Bellido A, Pereda-Baños A, López-Moliner J, Deco G, Soto-Faraco S. Deconstructing multisensory enhancement in detection. J Neurophysiol 2014; 113:1800-18. [PMID: 25520431 DOI: 10.1152/jn.00341.2014] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
The mechanisms responsible for the integration of sensory information from different modalities have become a topic of intense interest in psychophysics and neuroscience. Many authors now claim that early, sensory-based cross-modal convergence improves performance in detection tasks. An important strand of supporting evidence for this claim is based on statistical models such as the Pythagorean model or the probabilistic summation model. These models establish statistical benchmarks representing the best predicted performance under the assumption that there are no interactions between the two sensory paths. Following this logic, when observed detection performances surpass the predictions of these models, it is often inferred that such improvement indicates cross-modal convergence. We present a theoretical analyses scrutinizing some of these models and the statistical criteria most frequently used to infer early cross-modal interactions during detection tasks. Our current analysis shows how some common misinterpretations of these models lead to their inadequate use and, in turn, to contradictory results and misleading conclusions. To further illustrate the latter point, we introduce a model that accounts for detection performances in multimodal detection tasks but for which surpassing of the Pythagorean or probabilistic summation benchmark can be explained without resorting to early cross-modal interactions. Finally, we report three experiments that put our theoretical interpretation to the test and further propose how to adequately measure multimodal interactions in audiotactile detection tasks.
Collapse
Affiliation(s)
| | | | | | - Joan López-Moliner
- Universitat de Barcelona, Barcelona, Spain; Institute for Brain, Cognition and Behaviour (IR3C), Barcelona, Spain; and
| | - Gustavo Deco
- Universitat Pompeu Fabra, Barcelona, Spain; Institució Catalana de Recerca i Estudis Avançats (ICREA), Barcelona, Spain
| | - Salvador Soto-Faraco
- Universitat Pompeu Fabra, Barcelona, Spain; Institució Catalana de Recerca i Estudis Avançats (ICREA), Barcelona, Spain
| |
Collapse
|
27
|
Koka K, Tollin DJ. Linear coding of complex sound spectra by discharge rate in neurons of the medial nucleus of the trapezoid body (MNTB) and its inputs. Front Neural Circuits 2014; 8:144. [PMID: 25565971 PMCID: PMC4267272 DOI: 10.3389/fncir.2014.00144] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2014] [Accepted: 11/25/2014] [Indexed: 11/25/2022] Open
Abstract
The interaural level difference (ILD) cue to sound location is first encoded in the lateral superior olive (LSO). ILD sensitivity results because the LSO receives excitatory input from the ipsilateral cochlear nucleus and inhibitory input indirectly from the contralateral cochlear nucleus via glycinergic neurons of the ipsilateral medial nucleus of the trapezoid body (MNTB). It is hypothesized that in order for LSO neurons to encode ILDs, the sound spectra at both ears must be accurately encoded via spike rate by their afferents. This spectral-coding hypothesis has not been directly tested in MNTB, likely because MNTB neurons have been mostly described and studied recently in regards to their abilities to encode temporal aspects of sounds, not spectral. Here, we test the hypothesis that MNTB neurons and their inputs from the cochlear nucleus and auditory nerve code sound spectra via discharge rate. The Random Spectral Shape (RSS) method was used to estimate how the levels of 100-ms duration spectrally stationary stimuli were weighted, both linearly and non-linearly, across a wide band of frequencies. In general, MNTB neurons, and their globular bushy cell inputs, were found to be well-modeled by a linear weighting of spectra demonstrating that the pathways through the MNTB can accurately encode sound spectra including those resulting from the acoustical cues to sound location provided by head-related directional transfer functions (DTFs). Together with the anatomical and biophysical specializations for timing in the MNTB-LSO complex, these mechanisms may allow ILDs to be computed for complex stimuli with rapid spectrotemporally-modulated envelopes such as speech and animal vocalizations and moving sound sources.
Collapse
Affiliation(s)
- Kanthaiah Koka
- Department of Physiology and Biophysics, University of Colorado School of Medicine Aurora, CO, USA
| | - Daniel J Tollin
- Department of Physiology and Biophysics, University of Colorado School of Medicine Aurora, CO, USA
| |
Collapse
|
28
|
Effect of the attachment of the tectorial membrane on cochlear micromechanics and two-tone suppression. Biophys J 2014; 106:1398-405. [PMID: 24655515 DOI: 10.1016/j.bpj.2014.01.034] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2013] [Revised: 01/16/2014] [Accepted: 01/23/2014] [Indexed: 11/21/2022] Open
Abstract
The mechanical stimulation of the outer hair cell hair bundle (HB) is a key step in nonlinear cochlear amplification. We show how two-tone suppression (TTS), a hallmark of cochlear nonlinearity, can be used as an indirect measure of HB stimulation. Using two different nonlinear computational models of the cochlea, we investigate the effect of altering the mechanical load applied by the tectorial membrane (TM) on the outer hair cell HB. In the first model (TM-A model), the TM is attached to the spiral limbus (as in wild-type animals); in the second model (TM-D model), the TM is detached from the spiral limbus (mimicking the cochlea of Otoa(EGFP/EGFP) mutant mice). As in recent experiments, model simulations demonstrate that the absence of the TM attachment does not preclude cochlear amplification. However, detaching the TM alters the mechanical load applied by the TM on the HB at low frequencies and therefore affects TTS by low-frequency suppressors. For low-frequency suppressors, the suppression threshold obtained with the TM-A model corresponds to a constant suppressor displacement on the basilar membrane (as in experiments with wild-type animals), whereas it corresponds to a constant suppressor velocity with the TM-D model. The predictions with the TM-D model could be tested by measuring TTS on the basilar membrane of the Otoa(EGFP/EGFP) mice to improve our understanding of the fundamental workings of the cochlea.
Collapse
|
29
|
Charaziak KK, Siegel JH. Estimating cochlear frequency selectivity with stimulus-frequency otoacoustic emissions in chinchillas. J Assoc Res Otolaryngol 2014; 15:883-96. [PMID: 25230801 DOI: 10.1007/s10162-014-0487-3] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2014] [Accepted: 09/03/2014] [Indexed: 11/29/2022] Open
Abstract
It has been suggested that the tuning of the cochlear filters can be derived from measures of otoacoustic emissions (OAEs). Two approaches have been proposed to estimate cochlear frequency selectivity using OAEs evoked with a single tone (stimulus-frequency (SF)) OAEs: based on SFOAE group delays (SF-GDs) and on SFOAE suppression tuning curves (SF-STCs). The aim of this study was to evaluate whether either SF-GDs or SF-STCs obtained with low probe levels (30 dB SPL) correlate with more direct measures of cochlear tuning (compound action potential suppression tuning curves (CAP-STCs)) in chinchillas. The SFOAE-based estimates of tuning covaried with CAP-STCs tuning for >3 kHz probe frequencies, indicating that these measures are related to cochlear frequency selectivity. However, the relationship may be too weak to predict tuning with either SFOAE method in an individual. The SF-GD prediction of tuning was sharper than CAP-STC tuning. On the other hand, SF-STCs were consistently broader than CAP-STCs implying that SFOAEs may have less restricted region of generation in the cochlea than CAPs. Inclusion of <3 kHz data in a statistical model resulted in no significant or borderline significant covariation among the three methods: neither SFOAE test appears to reliably estimate an individual's CAP-STC tuning at low-frequencies. At the group level, SF-GDs and CAP-STCs showed similar tuning at low frequencies, while SF-STCs were over five times broader than the CAP-STCs indicating that low-frequency SFOAE may originate over a very broad region of the cochlea extending ≥5 mm basal to the tonotopic place of the probe.
Collapse
Affiliation(s)
- Karolina K Charaziak
- Department of Communication Sciences and Disorders, School of Communication, Northwestern University, 2240 Campus Drive, Evanston, IL, 60208-2952, USA,
| | | |
Collapse
|
30
|
Brown MC. Single-unit labeling of medial olivocochlear neurons: the cochlear frequency map for efferent axons. J Neurophysiol 2014; 111:2177-86. [PMID: 24598524 DOI: 10.1152/jn.00045.2014] [Citation(s) in RCA: 28] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Medial olivocochlear (MOC) neurons are efferent neurons that project axons from the brain to the cochlea. Their action on outer hair cells reduces the gain of the "cochlear amplifier," which shifts the dynamic range of hearing and reduces the effects of noise masking. The MOC effects in one ear can be elicited by sound in that ipsilateral ear or by sound in the contralateral ear. To study how MOC neurons project onto the cochlea to mediate these effects, single-unit labeling in guinea pigs was used to study the mapping of MOC neurons for neurons responsive to ipsilateral sound vs. those responsive to contralateral sound. MOC neurons were sharply tuned to sound frequency with a well-defined characteristic frequency (CF). However, their labeled termination spans in the organ of Corti ranged from narrow to broad, innervating between 14 and 69 outer hair cells per axon in a "patchy" pattern. For units responsive to ipsilateral sound, the midpoint of innervation was mapped according to CF in a relationship generally similar to, but with more variability than, that of auditory-nerve fibers. Thus, based on CF mappings, most of the MOC terminations miss outer hair cells involved in the cochlear amplifier for their CF, which are located more basally. Compared with ipsilaterally responsive neurons, contralaterally responsive neurons had an apical offset in termination and a larger span of innervation (an average of 10.41% cochlear distance), suggesting that when contralateral sound activates the MOC reflex, the actions are different than those for ipsilateral sound.
Collapse
Affiliation(s)
- M Christian Brown
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear Infirmary, and Department of Otology and Laryngology, Harvard Medical School, Boston, Massachusetts
| |
Collapse
|
31
|
Zhou Y, Wang X. Spatially extended forward suppression in primate auditory cortex. Eur J Neurosci 2013; 39:919-933. [PMID: 24372934 DOI: 10.1111/ejn.12460] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2013] [Revised: 10/22/2013] [Accepted: 11/20/2013] [Indexed: 11/26/2022]
Abstract
When auditory neurons are stimulated with a pair of sounds, the preceding sound can inhibit the neural responses to the succeeding sound. This phenomenon, referred to as 'forward suppression', has been linked to perceptual forward masking. Previous studies investigating forward suppression typically measured the interaction between masker and probe sounds using a fixed sound location. However, in natural environments, interacting sounds often come from different spatial locations. The present study investigated two questions regarding forward suppression in the primary auditory cortex and adjacent caudal field of awake marmoset monkeys. First, what is the relationship between the location of a masker and its effectiveness in inhibiting neural response to a probe? Second, does varying the location of a masker change the spectral profile of forward suppression? We found that a masker can inhibit a neuron's response to a probe located at a preferred location even when the masker is located at a non-preferred location of a neuron. This is especially so for neurons in the caudal field. Furthermore, we found that the strongest forward suppression is observed when a masker's frequency is close to the best frequency of a neuron, regardless of the location of the masker. These results reveal, for the first time, the stability of forward masking in cortical processing of multiple sounds presented from different locations. They suggest that forward suppression in the auditory cortex is spectrally specific and spatially broad with respect to the frequency and location of the masker, respectively.
Collapse
Affiliation(s)
- Yi Zhou
- Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, Johns Hopkins University, School of Medicine, Baltimore, MD, 21205, USA
| | | |
Collapse
|
32
|
Abstract
To enhance weak sounds while compressing the dynamic intensity range, auditory sensory cells amplify sound-induced vibrations in a nonlinear, intensity-dependent manner. In the course of this process, instantaneous waveform distortion is produced, with two conspicuous kinds of interwoven consequences, the introduction of new sound frequencies absent from the original stimuli, which are audible and detectable in the ear canal as otoacoustic emissions, and the possibility for an interfering sound to suppress the response to a probe tone, thereby enhancing contrast among frequency components. We review how the diverse manifestations of auditory nonlinearity originate in the gating principle of their mechanoelectrical transduction channels; how they depend on the coordinated opening of these ion channels ensured by connecting elements; and their links to the dynamic behavior of auditory sensory cells. This paper also reviews how the complex properties of waves traveling through the cochlea shape the manifestations of auditory nonlinearity. Examination methods based on the detection of distortions open noninvasive windows on the modes of activity of mechanosensitive structures in auditory sensory cells and on the distribution of sites of nonlinearity along the cochlear tonotopic axis, helpful for deciphering cochlear molecular physiology in hearing-impaired animal models. Otoacoustic emissions enable fast tests of peripheral sound processing in patients. The study of auditory distortions also contributes to the understanding of the perception of complex sounds.
Collapse
Affiliation(s)
- Paul Avan
- Laboratory of Neurosensory Biophysics, University of Auvergne, School of Medicine, Clermont-Ferrand, France; Institut National de la Santé et de la Recherche Médicale (INSERM), UMR 1107, Clermont-Ferrand, France; Centre Jean Perrin, Clermont-Ferrand, France; Department of Otolaryngology, County Hospital, Krems an der Donau, Austria; Laboratory of Genetics and Physiology of Hearing, Department of Neuroscience, Institut Pasteur, Paris, France; Collège de France, Genetics and Cell Physiology, Paris, France
| | - Béla Büki
- Laboratory of Neurosensory Biophysics, University of Auvergne, School of Medicine, Clermont-Ferrand, France; Institut National de la Santé et de la Recherche Médicale (INSERM), UMR 1107, Clermont-Ferrand, France; Centre Jean Perrin, Clermont-Ferrand, France; Department of Otolaryngology, County Hospital, Krems an der Donau, Austria; Laboratory of Genetics and Physiology of Hearing, Department of Neuroscience, Institut Pasteur, Paris, France; Collège de France, Genetics and Cell Physiology, Paris, France
| | - Christine Petit
- Laboratory of Neurosensory Biophysics, University of Auvergne, School of Medicine, Clermont-Ferrand, France; Institut National de la Santé et de la Recherche Médicale (INSERM), UMR 1107, Clermont-Ferrand, France; Centre Jean Perrin, Clermont-Ferrand, France; Department of Otolaryngology, County Hospital, Krems an der Donau, Austria; Laboratory of Genetics and Physiology of Hearing, Department of Neuroscience, Institut Pasteur, Paris, France; Collège de France, Genetics and Cell Physiology, Paris, France
| |
Collapse
|
33
|
Charaziak KK, Souza P, Siegel JH. Stimulus-frequency otoacoustic emission suppression tuning in humans: comparison to behavioral tuning. J Assoc Res Otolaryngol 2013; 14:843-62. [PMID: 24013802 DOI: 10.1007/s10162-013-0412-1] [Citation(s) in RCA: 31] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/18/2012] [Accepted: 08/11/2013] [Indexed: 11/30/2022] Open
Abstract
As shown by the work of Kemp and Chum in 1980, stimulus-frequency otoacoustic emission suppression tuning curves (SFOAE STCs) have potential to objectively estimate behaviorally measured tuning curves. To date, this potential has not been tested. This study aims to do so by comparing SFOAE STCs and behavioral measures of tuning (simultaneous masking psychophysical tuning curves, PTCs) in 10 normal-hearing listeners for frequency ranges centered around 1,000 and 4,000 Hz at low probe levels. Additionally, SFOAE STCs were collected for varying conditions (probe level and suppression criterion) to identify the optimal parameters for comparison with behavioral data and to evaluate how these conditions affect the features of SFOAE STCs. SFOAE STCs qualitatively resembled PTCs: they demonstrated band-pass characteristics and asymmetric shapes with steeper high-frequency sides than low, but unlike PTCs they were consistently tuned to frequencies just above the probe frequency. When averaged across subjects the shapes of SFOAE STCs and PTCs showed agreement for most recording conditions, suggesting that PTCs are predominantly shaped by the frequency-selective filtering and suppressive effects of the cochlea. Individual SFOAE STCs often demonstrated irregular shapes (e.g., "double-tips"), particularly for the 1,000-Hz probe, which were not observed for the same subject's PTC. These results show the limited utility of SFOAE STCs to assess tuning in an individual. The irregularly shaped SFOAE STCs may be attributed to contributions from SFOAE sources distributed over a region of the basilar membrane extending beyond the probe characteristic place, as suggested by a repeatable pattern of SFOAE residual phase shifts observed in individual data.
Collapse
Affiliation(s)
- Karolina K Charaziak
- Department of Communication Sciences and Disorders, Northwestern University, School of Communication, 2240 Campus Drive, Evanston, IL, 602080-2952, USA,
| | | | | |
Collapse
|
34
|
Understanding the neurophysiological basis of auditory abilities for social communication: a perspective on the value of ethological paradigms. Hear Res 2013; 305:3-9. [PMID: 23994815 DOI: 10.1016/j.heares.2013.08.008] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/25/2013] [Revised: 08/11/2013] [Accepted: 08/19/2013] [Indexed: 11/21/2022]
Abstract
Acoustic communication between animals requires them to detect, discriminate, and categorize conspecific or heterospecific vocalizations in their natural environment. Laboratory studies of the auditory-processing abilities that facilitate these tasks have typically employed a broad range of acoustic stimuli, ranging from natural sounds like vocalizations to "artificial" sounds like pure tones and noise bursts. However, even when using vocalizations, laboratory studies often test abilities like categorization in relatively artificial contexts. Consequently, it is not clear whether neural and behavioral correlates of these tasks (1) reflect extensive operant training, which drives plastic changes in auditory pathways, or (2) the innate capacity of the animal and its auditory system. Here, we review a number of recent studies, which suggest that adopting more ethological paradigms utilizing natural communication contexts are scientifically important for elucidating how the auditory system normally processes and learns communication sounds. Additionally, since learning the meaning of communication sounds generally involves social interactions that engage neuromodulatory systems differently than laboratory-based conditioning paradigms, we argue that scientists need to pursue more ethological approaches to more fully inform our understanding of how the auditory system is engaged during acoustic communication. This article is part of a Special Issue entitled "Communication Sounds and the Brain: New Directions and Perspectives".
Collapse
|
35
|
Palmer AR, Shackleton TM, Sumner CJ, Zobay O, Rees A. Classification of frequency response areas in the inferior colliculus reveals continua not discrete classes. J Physiol 2013; 591:4003-25. [PMID: 23753527 PMCID: PMC3764642 DOI: 10.1113/jphysiol.2013.255943] [Citation(s) in RCA: 44] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2013] [Accepted: 06/05/2013] [Indexed: 11/28/2022] Open
Abstract
A differential response to sound frequency is a fundamental property of auditory neurons. Frequency analysis in the cochlea gives rise to V-shaped tuning functions in auditory nerve fibres, but by the level of the inferior colliculus (IC), the midbrain nucleus of the auditory pathway, neuronal receptive fields display diverse shapes that reflect the interplay of excitation and inhibition. The origin and nature of these frequency receptive field types is still open to question. One proposed hypothesis is that the frequency response class of any given neuron in the IC is predominantly inherited from one of three major afferent pathways projecting to the IC, giving rise to three distinct receptive field classes. Here, we applied subjective classification, principal component analysis, cluster analysis, and other objective statistical measures, to a large population (2826) of frequency response areas from single neurons recorded in the IC of the anaesthetised guinea pig. Subjectively, we recognised seven frequency response classes (V-shaped, non-monotonic Vs, narrow, closed, tilt down, tilt up and double-peaked), that were represented at all frequencies. We could identify similar classes using our objective classification tools. Importantly, however, many neurons exhibited properties intermediate between these classes, and none of the objective methods used here showed evidence of discrete response classes. Thus receptive field shapes in the IC form continua rather than discrete classes, a finding consistent with the integration of afferent inputs in the generation of frequency response areas. The frequency disposition of inhibition in the response areas of some neurons suggests that across-frequency inputs originating at or below the level of the IC are involved in their generation.
Collapse
Affiliation(s)
- Alan R Palmer
- Medical Research Council Institute of Hearing Research, University Park, Nottingham NG7 2RD, UK.
| | | | | | | | | |
Collapse
|
36
|
Rasetshwane DM, Gorga MP, Neely ST. Signal-processing strategy for restoration of cross-channel suppression in hearing-impaired listeners. IEEE Trans Biomed Eng 2013; 61:64-75. [PMID: 23925364 DOI: 10.1109/tbme.2013.2276351] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Because frequency components interact nonlinearly with each other inside the cochlea, the loudness growth of tones is relatively simple in comparison to the loudness growth of complex sounds. The term suppression refers to a reduction in the response growth of one tone in the presence of a second tone. Suppression is a salient feature of normal cochlear processing and contributes to psychophysical masking. Suppression is evident in many measurements of cochlear function in subjects with normal hearing, including distortion-product otoacoustic emissions (DPOAEs). Suppression is also evident, to a lesser extent, in subjects with mild-to-moderate hearing loss. This paper describes a hearing-aid signal-processing strategy that aims to restore both loudness growth and two-tone suppression in hearing-impaired listeners. The prescription of gain for this strategy is based on measurements of loudness by a method known as categorical loudness scaling. The proposed signal-processing strategy reproduces measured DPOAE suppression tuning curves and generalizes to any number of frequency components. The restoration of both normal suppression and normal loudness has the potential to improve hearing-aid performance and user satisfaction.
Collapse
|
37
|
Laback B, Necciari T, Balazs P, Savel S, Ystad S. Simultaneous masking additivity for short Gaussian-shaped tones: spectral effects. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 134:1160-1171. [PMID: 23927115 DOI: 10.1121/1.4812773] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/02/2023]
Abstract
Laback et al. [(2011). J. Acoust. Soc. Am. 129, 888-897] investigated the additivity of nonsimultaneous masking using short Gaussian-shaped tones as maskers and target. The present study involved Gaussian stimuli to measure the additivity of simultaneous masking for combinations of up to four spectrally separated maskers. According to most basilar membrane measurements, the maskers should be processed linearly at the characteristic frequency (CF) of the target. Assuming also compression of the target, all masker combinations should produce excess masking (exceeding linear additivity). The results for a pair of maskers flanking the target indeed showed excess masking. The amount of excess masking could be predicted by a model assuming summation of masker-evoked excitations in intensity units at the target CF and compression of the target, using compressive input/output functions derived from the nonsimultaneous masking study. However, the combinations of lower-frequency maskers showed much less excess masking than predicted by the model. This cannot easily be attributed to factors like off-frequency listening, combination tone perception, or between-masker suppression. It was better predicted, however, by assuming weighted intensity summation of masker excitations. The optimum weights for the lower-frequency maskers were smaller than one, consistent with partial masker compression as indicated by recent psychoacoustic data.
Collapse
Affiliation(s)
- Bernhard Laback
- Acoustics Research Institute, Austrian Academy of Sciences, Wohllebengasse 12-14, A-1040 Vienna, Austria.
| | | | | | | | | |
Collapse
|
38
|
Versteegh CPC, van der Heijden M. The spatial buildup of compression and suppression in the mammalian cochlea. J Assoc Res Otolaryngol 2013; 14:523-45. [PMID: 23690278 PMCID: PMC3705085 DOI: 10.1007/s10162-013-0393-0] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2013] [Accepted: 04/23/2013] [Indexed: 11/25/2022] Open
Abstract
We recorded responses of the gerbil basilar membrane (BM) to wideband tone complexes. The intensity of one component was varied and the effects on the amplitude and phase of the others were assessed. This suppression paradigm enabled us to vary probe frequency and suppressor frequency independently, allowing the use of simple scaling arguments to analyze the spatial buildup of the nonlinear interaction between traveling waves. Most suppressors had the same effects on probe amplitude and phase as did wideband intensity increments. The main exception were suppressors above the characteristic frequency (CF) of the recording location, for which the frequency range of most affected probes was not constant, but shifted upward with suppressor frequency. BM displacement reliably predicted the effectiveness of low-side suppressors, but not high-side suppressors. We found “anti-suppression” of probes well below CF, i.e., suppressor-induced enhancement of probe response amplitude. Large (>1 cycle) phase effects occurred for above-CF probes. Phase shifts varied nonmonotonically, but systematically, with suppressor level, probe frequency, and suppressor frequency, reconciling apparent discrepancies in the literature. The analysis of spatial buildup revealed an accumulation of local effects on the propagation of the traveling wave, with larger BM displacement reducing the local forward gain. The propagation speed of the wave was also affected. With larger BM displacement, the basal portion of the wave slowed down, while the apical part sped up. This framework of spatial buildup of local effects unifies the widely different effects of overall intensity, low-side suppressors, and high-side suppressors on BM responses.
Collapse
|
39
|
Chang TR, Chiu TW, Sun X, Poon PWF. Modeling complex responses of FM-sensitive cells in the auditory midbrain using a committee machine. Brain Res 2013; 1536:44-52. [PMID: 23665390 DOI: 10.1016/j.brainres.2013.04.058] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2012] [Revised: 04/30/2013] [Accepted: 04/30/2013] [Indexed: 11/26/2022]
Abstract
Frequency modulation (FM) is an important building block of complex sounds that include speech signals. Exploring the neural mechanisms of FM coding with computer modeling could help understand how speech sounds are processed in the brain. Here, we modeled the single unit responses of auditory neurons recorded from the midbrain of anesthetized rats. These neurons displayed spectral temporal receptive fields (STRFs) that had multiple-trigger features, and were more complex than those with single-trigger features. Their responses have not been modeled satisfactorily with simple artificial neural networks, unlike neurons with simple-trigger features. To improve model performance, here we tested an approach with the committee machine. For a given neuron, the peri-stimulus time histogram (PSTH) was first generated in response to a repeated random FM tone, and peaks in the PSTH were segregated into groups based on the similarity of their pre-spike FM trigger features. Each group was then modeled using an artificial neural network with simple architecture, and, when necessary, by increasing the number of neurons in the hidden layer. After initial training, the artificial neural networks with their optimized weighting coefficients were pooled into a committee machine for training. Finally, the model performance was tested by prediction of the response of the same cell to a novel FM tone. The results showed improvement over simple artificial neural networks, supporting that trigger-feature-based modeling can be extended to cells with complex responses. This article is part of a Special Issue entitled Neural Coding 2012. This article is part of a Special Issue entitled Neural Coding 2012.
Collapse
Affiliation(s)
- T R Chang
- Department of Computer Science and Information Engineering, Southern Taiwan University of Science and Technology, Tainan, Taiwan.
| | | | | | | |
Collapse
|
40
|
Dau T, Piechowiak T, Ewert SD. Modeling within- and across-channel processes in comodulation masking release. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 133:350-364. [PMID: 23297908 DOI: 10.1121/1.4768882] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
The relative contributions of within-channel and across-channel processes to perceptual comodulation masking release (CMR) were investigated in the framework of an auditory processing model. A generalized version of the computational auditory signal processing and perception model [CASP; Jepsen et al., J. Acoust. Soc. Am. 124, 422-438 (2008)] was used and extended by an across-channel modulation processing stage according to Piechowiak et al. [J. Acoust. Soc. Am. 121, 2111-2126 (2007)]. Five experimental paradigms were considered: CMR with a broadband noise masker as a function of the masker spectrum level; CMR with four widely spaced flanking bands (FBs) varying in overall level; CMR with one FB varying in frequency and level relative to the on-frequency band (OFB); CMR with one FB varying in frequency; and CMR as a function of the number of FBs. The predictions suggest that at least three different mechanisms contribute to overall CMR in the considered conditions: (1) a within-channel process based on changes in the envelope characteristic due to the addition of the signal to the masker; (2) a within-channel process based on nonlinear peripheral processing of the OFB's envelope caused by the FB(s); and (3) an across-channel process that is robust across presentation levels but relatively small (2-5 dB).
Collapse
Affiliation(s)
- Torsten Dau
- Centre for Applied Hearing Research, Department of Electrical Engineering, Technical University of Denmark, DK-2800 Kgs. Lyngby, Denmark.
| | | | | |
Collapse
|
41
|
Abstract
The ferret (Mustela putorius) is a medium-sized, carnivorous mammal with good low-frequency hearing; it is relatively easy to train, and there is therefore a good body of behavioural data detailing its detection thresholds and localization abilities. However, despite extensive studies of the physiology of the central nervous system of the ferret, even extending to the prefrontal cortex, little is known of the functioning of the auditory periphery. Here, we provide an insight into this peripheral function by detailing responses of single auditory nerve fibres. Our expectation was that the ferret auditory nerve responsiveness would be similar that of its near relative, the cat. However, by comparing a range of variables (the frequency tuning, the variation of rate-level functions with spontaneous rate, and the high-frequency cut-off of phase locking) across several species, we show that the auditory nerve (and hence cochlea) in the ferret is more similar to that of the guinea-pig and chinchilla than to that of the cat. Animal models of hearing are often chosen on the basis of the similarity of their audiogram to that of the human, particularly in the low-frequency region. We show here that whereas the ferret hears well at low frequencies, this is likely to occur via fibres with higher characteristic frequencies. These qualitative differences in response characteristics in auditory nerve fibres are important in interpreting data across all of auditory science, as it has been argued recently that tuning in animals is broader than in humans.
Collapse
|
42
|
Zhou Y, Wang X. Level dependence of spatial processing in the primate auditory cortex. J Neurophysiol 2012; 108:810-26. [PMID: 22592309 DOI: 10.1152/jn.00500.2011] [Citation(s) in RCA: 36] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Sound localization in both humans and monkeys is tolerant to changes in sound levels. The underlying neural mechanism, however, is not well understood. This study reports the level dependence of individual neurons' spatial receptive fields (SRFs) in the primary auditory cortex (A1) and the adjacent caudal field in awake marmoset monkeys. We found that most neurons' excitatory SRF components were spatially confined in response to broadband noise stimuli delivered from the upper frontal sound field. Approximately half the recorded neurons exhibited little change in spatial tuning width over a ~20-dB change in sound level, whereas the remaining neurons showed either expansion or contraction in their tuning widths. Increased sound levels did not alter the percent distribution of tuning width for neurons collected in either cortical field. The population-averaged responses remained tuned between 30- and 80-dB sound pressure levels for neuronal groups preferring contralateral, midline, and ipsilateral locations. We further investigated the spatial extent and level dependence of the suppressive component of SRFs using a pair of sequentially presented stimuli. Forward suppression was observed when the stimuli were delivered from "far" locations, distant to the excitatory center of an SRF. In contrast to spatially confined excitation, the strength of suppression typically increased with stimulus level at both the excitatory center and far regions of an SRF. These findings indicate that although the spatial tuning of individual neurons varied with stimulus levels, their ensemble responses were level tolerant. Widespread spatial suppression may play an important role in limiting the sizes of SRFs at high sound levels in the auditory cortex.
Collapse
Affiliation(s)
- Yi Zhou
- Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, School of Medicine, Johns Hopkins University, Baltimore, Maryland 21205-2195, USA.
| | | |
Collapse
|
43
|
Abstract
Combination sensitivity in central auditory neurons is a form of spectrotemporal integration in which excitatory responses to sounds at one frequency are facilitated by sounds within a distinctly different frequency band. Combination-sensitive neurons respond selectively to acoustic elements of sonar echoes or social vocalizations. In mustached bats, this response property originates in high-frequency representations of the inferior colliculus (IC) and depends on low and high frequency-tuned glycinergic inputs. To identify the source of these inputs, we combined glycine immunohistochemistry with retrograde tract tracing. Tracers were deposited at high-frequency (>56 kHz), combination-sensitive recording sites in IC. Most glycine-immunopositive, retrogradely labeled cells were in ipsilateral ventral and intermediate nuclei of the lateral lemniscus (VNLL and INLL), with some double labeling in ipsilateral lateral and medial superior olivary nuclei (LSO and MSO). Generally, double-labeled cells were in expected high-frequency tonotopic areas, but some VNLL and INLL labeling appeared to be in low-frequency representations. To test whether these nuclei provide low frequency-tuned input to the high-frequency IC, we combined retrograde tracing from IC combination-sensitive sites with anterograde tracing from low frequency-tuned sites in the anteroventral cochlear nucleus (AVCN). Only VNLL and INLL contained retrogradely labeled cells near (≤50 μm) anterogradely labeled boutons. These cells likely receive excitatory low-frequency input from AVCN. Results suggest that combination-sensitive facilitation arises through convergence of high-frequency glycinergic inputs from VNLL, INLL, or MSO and low-frequency glycinergic inputs from VNLL or INLL. This work establishes an anatomical basis for spectrotemporal integration in the auditory midbrain and a functional role for monaural nuclei of the lateral lemniscus.
Collapse
|
44
|
Chen Z, Hu G, Glasberg BR, Moore BCJ. A new method of calculating auditory excitation patterns and loudness for steady sounds. Hear Res 2011; 282:204-15. [PMID: 21851853 DOI: 10.1016/j.heares.2011.08.001] [Citation(s) in RCA: 31] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/13/2011] [Revised: 07/20/2011] [Accepted: 08/03/2011] [Indexed: 11/26/2022]
Abstract
A new method for calculating auditory excitation patterns and loudness for steady sounds is described. The method is based on a nonlinear filterbank in which each filter is the sum of a broad passive filter and a sharp active filter. All filters have a rounded-exponential shape. For each center frequency (CF), the gain of the active filter is controlled by the output of the passive filter. The parameters of the model were derived from large sets of previously published notched-noise masking data obtained from human subjects. Excitation patterns derived using the new filterbank include the effects of basilar membrane compression. Loudness can be calculated as the area under the excitation pattern when plotted in intensity-like units on an ERB(N)-number (Cam) scale; no transformation from excitation to specific loudness is required. The method predicts the standard equal-loudness contours and loudness as a function of bandwidth with good accuracy. With some additional assumptions, the method also gives reasonably accurate predictions of partial loudness.
Collapse
Affiliation(s)
- Zhangli Chen
- Department of Biomedical Engineering, Medical School, Tsinghua University, Beijing 100084, China
| | | | | | | |
Collapse
|
45
|
Culling JF. Subcomponent cues in binaural unmasking. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2011; 129:3846-3855. [PMID: 21682408 DOI: 10.1121/1.3560944] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
The addition of a signal in the N0Sπ binaural configuration gives rise to fluctuations in interaural phase and amplitude. Sensitivity to these individual cues was measured by applying sinusoidal amplitude modulation (AM) or quasi-frequency modulation (QFM) to a band of noise. Discrimination between interaurally in-phase and out-of-phase modulation was measured using an adaptive task for narrow bands of noise at center frequencies from 250 to 1500 Hz, for modulation rates of 2-40 Hz, and with or without flanking bands of diotic noise. Discrimination thresholds increased steeply for QFM with increasing center frequency, but increased only modestly for AM, and mainly for modulation rates below 10 Hz. Flanking bands of noise increased thresholds for AM, but had no consistent effect for QFM. The results suggest that two underlying mechanisms may support binaural unmasking: one most sensitive to interaural amplitude modulations that is susceptible to across-frequency interference, and a second, most sensitive to interaural phase modulations that is immune to such effects.
Collapse
Affiliation(s)
- John F Culling
- School of Psychology, Cardiff University, Tower Building, Park Place, Cardiff CF10 3AT, United Kingdom. CullingJ@cf. ac.uk
| |
Collapse
|
46
|
Rodriguez J, Neely ST. Temporal aspects of suppression in distortion-product otoacoustic emissions. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2011; 129:3082-3089. [PMID: 21568411 PMCID: PMC3108389 DOI: 10.1121/1.3575553] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/28/2011] [Revised: 03/11/2011] [Accepted: 03/15/2011] [Indexed: 05/30/2023]
Abstract
This study examined the time course of cochlear suppression using a tone-burst suppressor to measure decrement of distortion-product otoacoustic emissions (DPOAEs). Seven normal-hearing subjects with ages ranging from 19 to 28 yr participated in the study. Each subject had audiometric thresholds ≤ 15 dB HL [re ANSI (2004) Specifications for Audiometers] for standard octave and inter-octave frequencies from 0.25 to 8 kHz. DPOAEs were elicited by primary tones with f(2) = 4.0 kHz and f(1) = 3.333 kHz (f(2)/f(1) = 1.2). For the f(2), L(2) combination, suppression was measured for three suppressor frequencies: One suppressor below f(2) (3.834 kHz) and two above f(2) (4.166 and 4.282 kHz) at three levels (55, 60, and 65 dB SPL). DPOAE decrement as a function of L(3) for the tone-burst suppressor was similar to decrements obtained with longer duration suppressors. Onset- and setoff- latencies were ≤ 4 ms, in agreement with previous physiological findings in auditory-nerve fiber studies that suggest suppression results from a nearly instantaneous compression of the waveform. Persistence of suppression was absent for the below-frequency suppressor (f(3) = 3.834 kHz) and was ≤ 3 ms for the two above-frequency suppressors (f(3) = 4.166 and 4.282 kHz).
Collapse
Affiliation(s)
- Joyce Rodriguez
- Starkey Hearing Research Center, 2150 Shattuck Avenue, Suite 408, Berkeley, California 94704, USA.
| | | |
Collapse
|
47
|
Gorga MP, Neely ST, Kopun J, Tan H. Growth of suppression in humans based on distortion-product otoacoustic emission measurements. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2011; 129:801-6. [PMID: 21361439 PMCID: PMC3070999 DOI: 10.1121/1.3523287] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/27/2010] [Revised: 11/01/2010] [Accepted: 11/04/2010] [Indexed: 05/30/2023]
Abstract
Distortion-product otoacoustic emissions (DPOAEs) were used to describe suppression growth in normal-hearing humans. Data were collected at eight f(2) frequencies ranging from 0.5 to 8 kHz for L(2) levels ranging from 10 to 60 dB sensation level. For each f(2) and L(2) combination, suppression was measured for nine or eleven suppressor frequencies (f(3)) whose levels varied from -20 to 85 dB sound pressure level (SPL). Suppression grew nearly linearly when f(3) ≈ f(2), grew more rapidly for f(3) < f(2), and grew more slowly for f(3) > f(2). These results are consistent with physiological and mechanical data from lower animals, as well as previous DPOAE data from humans, although no previous DPOAE study has described suppression growth for as wide a range of frequencies and levels. These trends were evident for all f(2) and L(2) combinations; however, some exceptions were noted. Specifically, suppression growth rate was less steep as a function of f(3) for f(2) frequencies ≤ 1 kHz. Thus, despite the qualitative similarities across frequency, there were quantitative differences related to f(2), suggesting that there may be subtle differences in suppression for frequencies above 1 kHz compared to frequencies below 1 kHz.
Collapse
Affiliation(s)
- Michael P Gorga
- Boys Town National Research Hospital, 555 North 30th Street, Omaha, Nebraska 68131, USA.
| | | | | | | |
Collapse
|
48
|
|
49
|
Lutfi RA, Liu CJ. A method for evaluating the relation between sound source segregation and masking. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2011; 129:EL34-EL38. [PMID: 21302979 PMCID: PMC3037974 DOI: 10.1121/1.3519871] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/28/2010] [Accepted: 10/27/2010] [Indexed: 05/30/2023]
Abstract
Sound source segregation refers to the ability to hear as separate entities two or more sound sources comprising a mixture. Masking refers to the ability of one sound to make another sound difficult to hear. Often in studies, masking is assumed to result from a failure of segregation, but this assumption may not always be correct. Here a method is offered to identify the relation between masking and sound source segregation in studies and an example is given of its application.
Collapse
Affiliation(s)
- Robert A Lutfi
- Auditory Behavioral Research Laboratory, University of Wisconsin, Madison, Wisconsin 53706, USA.
| | | |
Collapse
|
50
|
Reiss LAJ, Ramachandran R, May BJ. Effects of signal level and background noise on spectral representations in the auditory nerve of the domestic cat. J Assoc Res Otolaryngol 2010; 12:71-88. [PMID: 20824483 DOI: 10.1007/s10162-010-0232-5] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/23/2009] [Accepted: 08/09/2010] [Indexed: 12/22/2022] Open
Abstract
Background noise poses a significant obstacle for auditory perception, especially among individuals with hearing loss. To better understand the physiological basis of this perceptual impediment, the present study evaluated the effects of background noise on the auditory nerve representation of head-related transfer functions (HRTFs). These complex spectral shapes describe the directional filtering effects of the head and torso. When a broadband sound passes through the outer ear en route to the tympanic membrane, the HRTF alters its spectrum in a manner that establishes the perceived location of the sound source. HRTF-shaped noise shares many of the acoustic features of human speech, while communicating biologically relevant localization cues that are generalized across mammalian species. Previous studies have used parametric manipulations of random spectral shapes to elucidate HRTF coding principles at various stages of the cat's auditory system. This study extended that body of work by examining the effects of sound level and background noise on the quality of spectral coding in the auditory nerve. When fibers were classified by their spontaneous rates, the coding properties of the more numerous low-threshold, high-spontaneous rate fibers were found to degrade at high presentation levels and in low signal-to-noise ratios. Because cats are known to maintain accurate directional hearing under these challenging listening conditions, behavioral performance may be disproportionally based on the enhanced dynamic range of the less common high-threshold, low-spontaneous rate fibers.
Collapse
Affiliation(s)
- Lina A J Reiss
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore, MD, 21205, USA.
| | | | | |
Collapse
|