1
|
Bader K, Dalhoff E, Dierkes L, Braun LH, Gummer AW, Zelle D. Reliable Long-Term Serial Evaluation of Cochlear Function Using Pulsed Distortion-Product Otoacoustic Emissions: Analyzing Levels and Pressure Time Courses. Ear Hear 2024:00003446-990000000-00284. [PMID: 38809242 DOI: 10.1097/aud.0000000000001522] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/30/2024]
Abstract
OBJECTIVES To date, there is no international standard on how to use distortion-product otoacoustic emissions (DPOAEs) in serial measurements to accurately detect changes in the function of the cochlear amplifier due, for example, to ototoxic therapies, occupational noise, or the development of regenerative therapies. The use of clinically established standard DPOAE protocols for serial monitoring programs appears to be hampered by multiple factors, including probe placement and calibration effects, signal-processing complexities associated with multiple sites of emission generation as well as suboptimal selection of stimulus parameters. DESIGN Pulsed DPOAEs were measured seven times within 3 months for f2 = 1 to 14 kHz and L2 = 25 to 80 dB SPL in 20 ears of 10 healthy participants with normal hearing (mean age = 32.1 ± 9.7 years). L1 values were computed from individual optimal-path parameters derived from the corresponding individual DPOAE level map in the first test session. Three different DPOAE metrics for evaluating the functional state of the cochlear amplifier were investigated with respect to their test-retest reliability: (1) the interference-free, nonlinear-distortion component level (LOD), (2) the time course of the DPOAE-envelope levels, LDP(t), and (3) the squared, zero-lag correlation coefficient () between the time courses of the DPOAE-envelope pressures, pDP(t), measured in two sessions. The latter two metrics include the two main DPOAE components and their state of interference. RESULTS Collated over all sessions and frequencies, the median absolute difference for LOD was 1.93 dB and for LDP(t) was 2.52 dB; the median of was 0.988. For the low (f2 = 1 to 3 kHz), mid (f2 = 4 to 9 kHz), and high (f2 = 10 to 14 kHz) frequency ranges, the test-retest reliability of LOD increased with increasing signal to noise ratio (SNR). CONCLUSIONS On the basis of the knowledge gained from this study on the test-retest reliability of pulsed DPOAE signals and the current literature, we propose a DPOAE protocol for future serial monitoring applications that takes into account the following factors: (1) separation of DPOAE components, (2) use of individually optimal stimulus parameters, (3) SNR of at least 15 dB, (4) accurate pressure calibration, (5) consideration of frequency- and level-dependent test-retest reliabilities and corresponding reference ranges, and (6) stimulus levels L2 that are as low as possible with sufficient SNR to capture the nonlinear functional state of the cochlear amplifier operating at its highest gain.
Collapse
Affiliation(s)
- Katharina Bader
- Department of Otolaryngology, Head and Neck Surgery, Eberhard-Karls-University Tübingen, Tübingen, Germany
| | - Ernst Dalhoff
- Section of Physiological Acoustics and Communication, Department of Otolaryngology, Eberhard-Karls-University Tübingen, Tübingen, Germany
| | - Linda Dierkes
- Department of Otolaryngology, Head and Neck Surgery, Eberhard-Karls-University Tübingen, Tübingen, Germany
- Section of Physiological Acoustics and Communication, Department of Otolaryngology, Eberhard-Karls-University Tübingen, Tübingen, Germany
| | - Lore Helene Braun
- Department of Radiooncology, Eberhard-Karls-University Tübingen, Tübingen, Germany
| | - Anthony W Gummer
- Section of Physiological Acoustics and Communication, Department of Otolaryngology, Eberhard-Karls-University Tübingen, Tübingen, Germany
| | - Dennis Zelle
- Section of Physiological Acoustics and Communication, Department of Otolaryngology, Eberhard-Karls-University Tübingen, Tübingen, Germany
- Earlab GmbH, Tübingen, Germany
| |
Collapse
|
2
|
Jones SD, Stewart HJ, Westermann G. A maturational frequency discrimination deficit may explain developmental language disorder. Psychol Rev 2024; 131:695-715. [PMID: 37498700 PMCID: PMC11115354 DOI: 10.1037/rev0000436] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2022] [Revised: 05/04/2023] [Accepted: 05/10/2023] [Indexed: 07/29/2023]
Abstract
Auditory perceptual deficits are widely observed among children with developmental language disorder (DLD). Yet, the nature of these deficits and the extent to which they explain speech and language problems remain controversial. In this study, we hypothesize that disruption to the maturation of the basilar membrane may impede the optimization of the auditory pathway from brainstem to cortex, curtailing high-resolution frequency sensitivity and the efficient spectral decomposition and encoding of natural speech. A series of computational simulations involving deep convolutional neural networks that were trained to encode, recognize, and retrieve naturalistic speech are presented to demonstrate the strength of this account. These neural networks were built on top of biologically truthful inner ear models developed to model human cochlea function, which-in the key innovation of the present study-were scheduled to mature at different rates over time. Delaying cochlea maturation qualitatively replicated the linguistic behavior and neurophysiology of individuals with language learning difficulties in a number of ways, resulting in (a) delayed language acquisition profiles, (b) lower spoken word recognition accuracy, (c) word finding and retrieval difficulties, (d) "fuzzy" and intersecting speech encodings and signatures of immature neural optimization, and (e) emergent working memory and attentional deficits. These simulations illustrate many negative cascading effects that a primary maturational frequency discrimination deficit may have on early language development and generate precise and testable hypotheses for future research into the nature and cost of auditory processing deficits in children with language learning difficulties. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
Collapse
|
3
|
Griffiths CS, Lebert JM, Sollini J, Bizley JK. Gradient boosted decision trees reveal nuances of auditory discrimination behavior. PLoS Comput Biol 2024; 20:e1011985. [PMID: 38626220 PMCID: PMC11051626 DOI: 10.1371/journal.pcbi.1011985] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2023] [Revised: 04/26/2024] [Accepted: 03/09/2024] [Indexed: 04/18/2024] Open
Abstract
Animal psychophysics can generate rich behavioral datasets, often comprised of many 1000s of trials for an individual subject. Gradient-boosted models are a promising machine learning approach for analyzing such data, partly due to the tools that allow users to gain insight into how the model makes predictions. We trained ferrets to report a target word's presence, timing, and lateralization within a stream of consecutively presented non-target words. To assess the animals' ability to generalize across pitch, we manipulated the fundamental frequency (F0) of the speech stimuli across trials, and to assess the contribution of pitch to streaming, we roved the F0 from word token to token. We then implemented gradient-boosted regression and decision trees on the trial outcome and reaction time data to understand the behavioral factors behind the ferrets' decision-making. We visualized model contributions by implementing SHAPs feature importance and partial dependency plots. While ferrets could accurately perform the task across all pitch-shifted conditions, our models reveal subtle effects of shifting F0 on performance, with within-trial pitch shifting elevating false alarms and extending reaction times. Our models identified a subset of non-target words that animals commonly false alarmed to. Follow-up analysis demonstrated that the spectrotemporal similarity of target and non-target words rather than similarity in duration or amplitude waveform was the strongest predictor of the likelihood of false alarming. Finally, we compared the results with those obtained with traditional mixed effects models, revealing equivalent or better performance for the gradient-boosted models over these approaches.
Collapse
Affiliation(s)
| | - Jules M. Lebert
- Ear Institute, University College London, London, United Kingdom
| | - Joseph Sollini
- Ear Institute, University College London, London, United Kingdom
- Hearing Sciences, University of Nottingham, Nottingham, United Kingdom
| | | |
Collapse
|
4
|
Joris PX, Verschooten E, Mc Laughlin M, Versteegh C, van der Heijden M. Frequency selectivity in monkey auditory nerve studied with suprathreshold multicomponent stimuli. Hear Res 2024; 443:108964. [PMID: 38277882 DOI: 10.1016/j.heares.2024.108964] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/17/2023] [Revised: 01/15/2024] [Accepted: 01/20/2024] [Indexed: 01/28/2024]
Abstract
Data from non-human primates can help extend observations from non-primate species to humans. Here we report measurements on the auditory nerve of macaque monkeys in the context of a controversial topic important to human hearing. A range of techniques have been used to examine the claim, which is not generally accepted, that human frequency tuning is sharper than traditionally thought, and sharper than in commonly used animal models. Data from single auditory-nerve fibers occupy a pivotal position to examine this claim, but are not available for humans. A previous study reported sharper tuning in auditory-nerve fibers of macaque relative to the cat. A limitation of these and other single-fiber data is that frequency selectivity was measured with tonal threshold-tuning curves, which do not directly assess spectral filtering and whose shape is sharpened by cochlear nonlinearity. Our aim was to measure spectral filtering with wideband suprathreshold stimuli in the macaque auditory nerve. We obtained responses of single nerve fibers of anesthetized macaque monkeys and cats to a suprathreshold, wideband, multicomponent stimulus designed to allow characterization of spectral filtering at any cochlear locus. Quantitatively the differences between the two species are smaller than in previous studies, but consistent with these studies the filters obtained show a trend of sharper tuning in macaque, relative to the cat, for fibers in the basal half of the cochlea. We also examined differences in group delay measured on the phase data near the characteristic frequency versus in the low-frequency tail. The phase data are consistent with the interpretation of sharper frequency tuning in monkey in the basal half of the cochlea. We conclude that use of suprathreshold, wide-band stimuli supports the interpretation of sharper frequency selectivity in macaque nerve fibers relative to the cat, although the difference is less marked than apparent from the assessment with tonal threshold-based data.
Collapse
Affiliation(s)
- P X Joris
- Lab of Auditory Neurophysiology, KU Leuven, O&N2 KU Leuven, Herestraat 49 bus 1021, Leuven B-3000, Belgium.
| | - E Verschooten
- Lab of Auditory Neurophysiology, KU Leuven, O&N2 KU Leuven, Herestraat 49 bus 1021, Leuven B-3000, Belgium
| | - M Mc Laughlin
- Lab of Auditory Neurophysiology, KU Leuven, O&N2 KU Leuven, Herestraat 49 bus 1021, Leuven B-3000, Belgium
| | - Cpc Versteegh
- Department of Neuroscience, Erasmus MC, Rotterdam, the Netherlands
| | | |
Collapse
|
5
|
Deloche F, Parida S, Sivaprakasam A, Heinz MG. Estimation of Cochlear Frequency Selectivity Using a Convolution Model of Forward-Masked Compound Action Potentials. J Assoc Res Otolaryngol 2024; 25:35-51. [PMID: 38278969 PMCID: PMC10907335 DOI: 10.1007/s10162-023-00922-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2023] [Accepted: 12/09/2023] [Indexed: 01/28/2024] Open
Abstract
PURPOSE Frequency selectivity is a fundamental property of the peripheral auditory system; however, the invasiveness of auditory nerve (AN) experiments limits its study in the human ear. Compound action potentials (CAPs) associated with forward masking have been suggested as an alternative to assess cochlear frequency selectivity. Previous methods relied on an empirical comparison of AN and CAP tuning curves in animal models, arguably not taking full advantage of the information contained in forward-masked CAP waveforms. METHODS To improve the estimation of cochlear frequency selectivity based on the CAP, we introduce a convolution model to fit forward-masked CAP waveforms. The model generates masking patterns that, when convolved with a unitary response, can predict the masking of the CAP waveform induced by Gaussian noise maskers. Model parameters, including those characterizing frequency selectivity, are fine-tuned by minimizing waveform prediction errors across numerous masking conditions, yielding robust estimates. RESULTS The method was applied to click-evoked CAPs at the round window of anesthetized chinchillas using notched-noise maskers with various notch widths and attenuations. The estimated quality factor Q10 as a function of center frequency is shown to closely match the average quality factor obtained from AN fiber tuning curves, without the need for an empirical correction factor. CONCLUSION This study establishes a moderately invasive method for estimating cochlear frequency selectivity with potential applicability to other animal species or humans. Beyond the estimation of frequency selectivity, the proposed model proved to be remarkably accurate in fitting forward-masked CAP responses and could be extended to study more complex aspects of cochlear signal processing (e.g., compressive nonlinearities).
Collapse
Affiliation(s)
- François Deloche
- Department of Speech, Language, and Hearing Sciences, Purdue University, 715 Clinic Drive, West Lafayette, 47907, IN, USA.
| | - Satyabrata Parida
- Department of Speech, Language, and Hearing Sciences, Purdue University, 715 Clinic Drive, West Lafayette, 47907, IN, USA
- Weldon School of Biomedical Engineering, Purdue University, 206 S. Martin Jischke Drive, West Lafayette, 47907, IN, USA
| | - Andrew Sivaprakasam
- Weldon School of Biomedical Engineering, Purdue University, 206 S. Martin Jischke Drive, West Lafayette, 47907, IN, USA
| | - Michael G Heinz
- Department of Speech, Language, and Hearing Sciences, Purdue University, 715 Clinic Drive, West Lafayette, 47907, IN, USA
- Weldon School of Biomedical Engineering, Purdue University, 206 S. Martin Jischke Drive, West Lafayette, 47907, IN, USA
| |
Collapse
|
6
|
Li YH, Joris PX. Case reopened: A temporal basis for harmonic pitch templates in the early auditory system?a). THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 154:3986-4003. [PMID: 38149819 DOI: 10.1121/10.0023969] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/14/2023] [Accepted: 12/04/2023] [Indexed: 12/28/2023]
Abstract
A fundamental assumption of rate-place models of pitch is the existence of harmonic templates in the central nervous system (CNS). Shamma and Klein [(2000). J. Acoust. Soc. Am. 107, 2631-2644] hypothesized that these templates have a temporal basis. Coincidences in the temporal fine-structure of neural spike trains, even in response to nonharmonic, stochastic stimuli, would be sufficient for the development of harmonic templates. The physiological plausibility of this hypothesis is tested. Responses to pure tones, low-pass noise, and broadband noise from auditory nerve fibers and brainstem "high-sync" neurons are studied. Responses to tones simulate the output of fibers with infinitely sharp filters: for these responses, harmonic structure in a coincidence matrix comparing pairs of spike trains is indeed found. However, harmonic template structure is not observed in coincidences across responses to broadband noise, which are obtained from nerve fibers or neurons with enhanced synchronization. Using a computer model based on that of Shamma and Klein, it is shown that harmonic templates only emerge when consecutive processing steps (cochlear filtering, lateral inhibition, and temporal enhancement) are implemented in extreme, physiologically implausible form. It is concluded that current physiological knowledge does not support the hypothesis of Shamma and Klein (2000).
Collapse
Affiliation(s)
- Yi-Hsuan Li
- Laboratory of Auditory Neurophysiology, Medical School, Campus Gasthuisberg, University of Leuven, B-3000 Leuven, Belgium
| | - Philip X Joris
- Laboratory of Auditory Neurophysiology, Medical School, Campus Gasthuisberg, University of Leuven, B-3000 Leuven, Belgium
| |
Collapse
|
7
|
Vinay, Moore BCJ. Exploiting individual differences to assess the role of place and phase locking cues in auditory frequency discrimination at 2 kHz. Sci Rep 2023; 13:13801. [PMID: 37612303 PMCID: PMC10447419 DOI: 10.1038/s41598-023-40571-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Accepted: 08/13/2023] [Indexed: 08/25/2023] Open
Abstract
The relative role of place and temporal mechanisms in auditory frequency discrimination was assessed for a centre frequency of 2 kHz. Four measures of frequency discrimination were obtained for 63 normal-hearing participants: detection of frequency modulation using modulation rates of 2 Hz (FM2) and 20 Hz (FM20); detection of a change in frequency across successive pure tones (difference limen for frequency, DLF); and detection of changes in the temporal fine structure of bandpass filtered complex tones centred at 2 kHz (TFS). Previous work has suggested that: FM2 depends on the use of both temporal and place cues; FM20 depends primarily on the use of place cues because the temporal mechanism cannot track rapid changes in frequency; DLF depends primarily on temporal cues; TFS depends exclusively on temporal cues. This led to the following predicted patterns of the correlations of scores across participants: DLF and TFS should be highly correlated; FM2 should be correlated with DLF and TFS; FM20 should not be correlated with DLF or TFS. The results were broadly consistent with these predictions and with the idea that frequency discrimination at 2 kHz depends partly or primarily on temporal cues except for frequency modulation detection at a high rate.
Collapse
Affiliation(s)
- Vinay
- Audiology Group, Department of Neuromedicine and Movement Science, Faculty of Medicine and Health Sciences, Norwegian University of Science and Technology (NTNU), Tungasletta 2, 7491, Trondheim, Norway.
| | - Brian C J Moore
- Cambridge Hearing Group, Department of Psychology, University of Cambridge, Cambridge, UK
| |
Collapse
|
8
|
Whiteford KL, Oxenham AJ. Sensitivity to Frequency Modulation is Limited Centrally. J Neurosci 2023; 43:3687-3695. [PMID: 37028932 PMCID: PMC10198444 DOI: 10.1523/jneurosci.0995-22.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2022] [Revised: 03/23/2023] [Accepted: 03/31/2023] [Indexed: 04/09/2023] Open
Abstract
Modulations in both amplitude and frequency are prevalent in natural sounds and are critical in defining their properties. Humans are exquisitely sensitive to frequency modulation (FM) at the slow modulation rates and low carrier frequencies that are common in speech and music. This enhanced sensitivity to slow-rate and low-frequency FM has been widely believed to reflect precise, stimulus-driven phase locking to temporal fine structure in the auditory nerve. At faster modulation rates and/or higher carrier frequencies, FM is instead thought to be coded by coarser frequency-to-place mapping, where FM is converted to amplitude modulation (AM) via cochlear filtering. Here, we show that patterns of human FM perception that have classically been explained by limits in peripheral temporal coding are instead better accounted for by constraints in the central processing of fundamental frequency (F0) or pitch. We measured FM detection in male and female humans using harmonic complex tones with an F0 within the range of musical pitch but with resolved harmonic components that were all above the putative limits of temporal phase locking (>8 kHz). Listeners were more sensitive to slow than fast FM rates, even though all components were beyond the limits of phase locking. In contrast, AM sensitivity remained better at faster than slower rates, regardless of carrier frequency. These findings demonstrate that classic trends in human FM sensitivity, previously attributed to auditory nerve phase locking, may instead reflect the constraints of a unitary code that operates at a more central level of processing.SIGNIFICANCE STATEMENT Natural sounds involve dynamic frequency and amplitude fluctuations. Humans are particularly sensitive to frequency modulation (FM) at slow rates and low carrier frequencies, which are prevalent in speech and music. This sensitivity has been ascribed to encoding of stimulus temporal fine structure (TFS) via phase-locked auditory nerve activity. To test this long-standing theory, we measured FM sensitivity using complex tones with a low F0 but only high-frequency harmonics beyond the limits of phase locking. Dissociating the F0 from TFS showed that FM sensitivity is limited not by peripheral encoding of TFS but rather by central processing of F0, or pitch. The results suggest a unitary code for FM detection limited by more central constraints.
Collapse
Affiliation(s)
- Kelly L Whiteford
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455
| |
Collapse
|
9
|
Oxenham AJ. Questions and controversies surrounding the perception and neural coding of pitch. Front Neurosci 2023; 16:1074752. [PMID: 36699531 PMCID: PMC9868815 DOI: 10.3389/fnins.2022.1074752] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2022] [Accepted: 12/16/2022] [Indexed: 01/12/2023] Open
Abstract
Pitch is a fundamental aspect of auditory perception that plays an important role in our ability to understand speech, appreciate music, and attend to one sound while ignoring others. The questions surrounding how pitch is represented in the auditory system, and how our percept relates to the underlying acoustic waveform, have been a topic of inquiry and debate for well over a century. New findings and technological innovations have led to challenges of some long-standing assumptions and have raised new questions. This article reviews some recent developments in the study of pitch coding and perception and focuses on the topic of how pitch information is extracted from peripheral representations based on frequency-to-place mapping (tonotopy), stimulus-driven auditory-nerve spike timing (phase locking), or a combination of both. Although a definitive resolution has proved elusive, the answers to these questions have potentially important implications for mitigating the effects of hearing loss via devices such as cochlear implants.
Collapse
Affiliation(s)
- Andrew J. Oxenham
- Center for Applied and Translational Sensory Science, University of Minnesota Twin Cities, Minneapolis, MN, United States
- Department of Psychology, University of Minnesota Twin Cities, Minneapolis, MN, United States
| |
Collapse
|
10
|
Shofner WP. Cochlear tuning and the peripheral representation of harmonic sounds in mammals. J Comp Physiol A Neuroethol Sens Neural Behav Physiol 2023; 209:145-161. [PMID: 35867137 DOI: 10.1007/s00359-022-01560-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2022] [Revised: 06/24/2022] [Accepted: 07/01/2022] [Indexed: 02/07/2023]
Abstract
Albert Feng was a prominent comparative neurophysiologist whose research provided numerous contributions towards understanding how the spectral and temporal characteristics of vocalizations underlie sound communication in frogs and bats. The present study is dedicated to Al's memory and compares the spectral and temporal representations of stochastic, complex sounds which underlie the perception of pitch strength in humans and chinchillas. Specifically, the pitch strengths of these stochastic sounds differ between humans and chinchillas, suggesting that humans and chinchillas may be using different cues. Outputs of auditory filterbank models based on human and chinchilla cochlear tuning were examined. Excitation patterns of harmonics are enhanced in humans as compared with chinchillas. In contrast, summary correlograms are degraded in humans as compared with chinchillas. Comparing summary correlograms and excitation patterns with corresponding behavioral data on pitch strength suggests that the dominant cue for pitch strength in humans is spectral (i.e., harmonic) structure, whereas the dominant cue for chinchillas is temporal (i.e., envelope) structure. The results support arguments that the broader cochlear tuning in non-human mammals emphasizes temporal cues for pitch perception, whereas the sharper cochlear tuning in humans emphasizes spectral cues.
Collapse
Affiliation(s)
- William P Shofner
- Department of Speech, Language and Hearing Sciences, Indiana University, 2631 East Discovery Parkway, Bloomington, IN, 47408, USA.
| |
Collapse
|
11
|
Guinan JJ. Cochlear amplification in the short-wave region by outer hair cells changing organ-of-Corti area to amplify the fluid traveling wave. Hear Res 2022. [DOI: 10.1016/j.heares.2022.108641] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
|
12
|
Guérit F, Middlebrooks JC, Richardson ML, Arneja A, Harland AJ, Gransier R, Wouters J, Carlyon RP. Tonotopic Selectivity in Cats and Humans: Electrophysiology and Psychophysics. J Assoc Res Otolaryngol 2022; 23:513-534. [PMID: 35697952 PMCID: PMC9437197 DOI: 10.1007/s10162-022-00851-5] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2021] [Accepted: 05/02/2022] [Indexed: 01/06/2023] Open
Abstract
We describe a scalp-recorded measure of tonotopic selectivity, the "cortical onset response" (COR) and compare the results between humans and cats. The COR results, in turn, were compared with psychophysical masked-detection thresholds obtained using similar stimuli and obtained from both species. The COR consisted of averaged responses elicited by 50-ms tone-burst probes presented at 1-s intervals against a continuous noise masker. The noise masker had a bandwidth of 1 or 1/8th octave, geometrically centred on 4000 Hz for humans and on 8000 Hz for cats. The probe frequency was either - 0.5, - 0.25, 0, 0.25 or 0.5 octaves re the masker centre frequency. The COR was larger for probe frequencies more distant from the centre frequency of the masker, and this effect was greater for the 1/8th-octave than for the 1-octave masker. This pattern broadly reflected the masked excitation patterns obtained psychophysically with similar stimuli in both species. However, the positive signal-to-noise ratio used to obtain reliable COR measures meant that some aspects of the data differed from those obtained psychophysically, in a way that could be partly explained by the upward spread of the probe's excitation pattern. Our psychophysical measurements also showed that the auditory filter width obtained at 8000 Hz using notched-noise maskers was slightly wider in cat than previous measures from humans. We argue that although conclusions from COR measures differ in some ways from conclusions based on psychophysics, the COR measures provide an objective, noninvasive, valid measure of tonotopic selectivity that does not require training and that may be applied to acoustic and cochlear-implant experiments in humans and laboratory animals.
Collapse
Affiliation(s)
- Francois Guérit
- grid.5335.00000000121885934Cambridge Hearing Group, MRC Cognition & Brain Sciences Unit, University of Cambridge, Cambridge, England
| | - John C. Middlebrooks
- grid.266093.80000 0001 0668 7243Department of Otolaryngology, University of California at Irvine, Irvine, CA USA
- grid.266093.80000 0001 0668 7243Department of Neurobiology and Behavior, University of California at Irvine, Irvine, CA USA
- grid.266093.80000 0001 0668 7243Department of Cognitive Sciences, University of California at Irvine, Irvine, CA USA
- grid.266093.80000 0001 0668 7243Department of Biomedical Engineering, University of California at Irvine, Irvine, CA USA
| | - Matthew L. Richardson
- grid.266093.80000 0001 0668 7243Department of Otolaryngology, University of California at Irvine, Irvine, CA USA
| | - Akshat Arneja
- grid.266093.80000 0001 0668 7243Department of Cognitive Sciences, University of California at Irvine, Irvine, CA USA
| | - Andrew J. Harland
- grid.5335.00000000121885934Cambridge Hearing Group, MRC Cognition & Brain Sciences Unit, University of Cambridge, Cambridge, England
| | - Robin Gransier
- Dept. of Neurosciences, ExpORL, Leuven, Louvain, KU Belgium
| | - Jan Wouters
- Dept. of Neurosciences, ExpORL, Leuven, Louvain, KU Belgium
| | - Robert P. Carlyon
- grid.5335.00000000121885934Cambridge Hearing Group, MRC Cognition & Brain Sciences Unit, University of Cambridge, Cambridge, England
| |
Collapse
|
13
|
Leschke J, Orellana GR, Shera CA, Oxenham AJ. Auditory filter shapes derived from forward and simultaneous masking at low frequencies: Implications for human cochlear tuning. Hear Res 2022; 420:108500. [PMID: 35405591 PMCID: PMC9167757 DOI: 10.1016/j.heares.2022.108500] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/17/2021] [Revised: 03/08/2022] [Accepted: 03/28/2022] [Indexed: 01/04/2023]
Abstract
Behavioral forward-masking thresholds with a spectrally notched-noise masker and a fixed low-level probe tone have been shown to provide accurate estimates of cochlear tuning. Estimates using simultaneous masking are similar but generally broader, presumably due to nonlinear cochlear suppression effects. So far, estimates with forward masking have been limited to frequencies of 1 kHz and above. This study used spectrally notched noise under forward and simultaneous masking to estimate frequency selectivity between 200 and 1000 Hz for young adult listeners with normal hearing. Estimates of filter tuning at 1000 Hz were in agreement with previous studies. Estimated tuning broadened below 1000 Hz, with the filter quality factor based on the equivalent rectangular bandwidth (QERB) decreasing more rapidly with decreasing frequency than predicted by previous equations, in line with earlier predictions based on otoacoustic-emission latencies. Estimates from simultaneous masking remained broader than those from forward masking by approximately the same ratio. The new data provide a way to compare human cochlear tuning estimates with auditory-nerve tuning curves from other species across most of the auditory frequency range.
Collapse
Affiliation(s)
| | | | - Christopher A. Shera
- Auditory Research Center, Caruso Department of Otolaryngology, University of Southern California, Los Angeles, CA 90033
| | - Andrew J. Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, MN 55455
| |
Collapse
|
14
|
Joris PX. In praise of adventitious sounds. Hear Res 2022; 425:108592. [DOI: 10.1016/j.heares.2022.108592] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/04/2021] [Revised: 07/13/2022] [Accepted: 07/26/2022] [Indexed: 11/04/2022]
|
15
|
Sumner CJ, Bergevin C, Oxenham AJ, Shera CA. WHAT MAKES HUMAN HEARING SPECIAL? FRONTIERS FOR YOUNG MINDS 2022; 10:708921. [PMID: 37465203 PMCID: PMC10353771 DOI: 10.3389/frym.2022.708921] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 07/20/2023]
Abstract
Humans and many other animals can hear a wide range of sounds. We can hear low and high notes and both quiet and loud sounds. We are also very good at telling the difference between sounds that are similar, like the speech sounds "argh" and "ah," and picking apart sounds that are mixed together, like when an orchestra is playing. But how do human hearing abilities compare to those of other animals? In this article, we discover how the inner ear determines hearing abilities. Many other mammals can hear very high notes that we cannot, and some can hear quiet sounds that we cannot. However, humans may be better than any other species at distinguishing similar sounds. We know this because, milliseconds after the sounds around us go into our ears, other sounds come out: sounds that are actually produced by those same ears!
Collapse
Affiliation(s)
| | - Christopher Bergevin
- Department of Physics and Astronomy and Centre for Vision Research, York University, Toronto, ON, Canada
| | - Andrew J. Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, MN, United States
- Department of Otolaryngology, University of Minnesota, Minneapolis, MN, United States
| | - Christopher A. Shera
- Caruso Department of Otolaryngology, University of Southern California, Los Angeles, CA, United States
- Department of Physics and Astronomy, University of Southern California, Los Angeles, CA, United States
| |
Collapse
|
16
|
Easy and Hard Auditory Tasks Distinguished by Otoacoustic Emissions and Event-related Potentials: Insights into Efferent System Activity. Neuroscience 2022; 491:87-97. [DOI: 10.1016/j.neuroscience.2022.03.029] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2021] [Revised: 03/17/2022] [Accepted: 03/24/2022] [Indexed: 11/22/2022]
|
17
|
In Vivo Basilar Membrane Time Delays in Humans. Brain Sci 2022; 12:brainsci12030400. [PMID: 35326357 PMCID: PMC8946056 DOI: 10.3390/brainsci12030400] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/17/2022] [Revised: 03/11/2022] [Accepted: 03/11/2022] [Indexed: 12/10/2022] Open
Abstract
To date, objective measurements and psychophysical experiments have been used to measure frequency dependent basilar membrane (BM) delays in humans; however, in vivo measurements have not been made. This study aimed to measure BM delays by performing intracochlear electrocochleography in cochlear implant recipients. Sixteen subjects with various degrees of hearing abilities were selected. Postoperative Computer Tomography was performed to determine electrode locations. Electrical potentials in response to acoustic tone pips at 0.25, 0.5, 1, 2, and 4 kHz and clicks were recorded with electrodes at the frequency specific region. The electrode array was inserted up to the characteristic cochlear frequency region of 250 Hz for 6 subjects. Furthermore, the array was inserted in the region of 500 Hz for 15 subjects, and 1, 2, and 4 kHz were reached in all subjects. Intracochlear electrocochleography for each frequency-specific tone pip and clicks showed detectable responses in all subjects. The latencies differed among the cochlear location and the cochlear microphonic (CM) onset latency increased with decreasing frequency and were consistent with click derived band technique. Accordingly, BM delays in humans could be derived. The BM delays increased systematically along the cochlea from basal to apical end and were in accordance with Ruggero and Temchin, 2007.
Collapse
|
18
|
Guest DR, Oxenham AJ. Human discrimination and modeling of high-frequency complex tones shed light on the neural codes for pitch. PLoS Comput Biol 2022; 18:e1009889. [PMID: 35239639 PMCID: PMC8923464 DOI: 10.1371/journal.pcbi.1009889] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2021] [Revised: 03/15/2022] [Accepted: 02/02/2022] [Indexed: 11/24/2022] Open
Abstract
Accurate pitch perception of harmonic complex tones is widely believed to rely on temporal fine structure information conveyed by the precise phase-locked responses of auditory-nerve fibers. However, accurate pitch perception remains possible even when spectrally resolved harmonics are presented at frequencies beyond the putative limits of neural phase locking, and it is unclear whether residual temporal information, or a coarser rate-place code, underlies this ability. We addressed this question by measuring human pitch discrimination at low and high frequencies for harmonic complex tones, presented either in isolation or in the presence of concurrent complex-tone maskers. We found that concurrent complex-tone maskers impaired performance at both low and high frequencies, although the impairment introduced by adding maskers at high frequencies relative to low frequencies differed between the tested masker types. We then combined simulated auditory-nerve responses to our stimuli with ideal-observer analysis to quantify the extent to which performance was limited by peripheral factors. We found that the worsening of both frequency discrimination and F0 discrimination at high frequencies could be well accounted for (in relative terms) by optimal decoding of all available information at the level of the auditory nerve. A Python package is provided to reproduce these results, and to simulate responses to acoustic stimuli from the three previously published models of the human auditory nerve used in our analyses.
Collapse
Affiliation(s)
- Daniel R. Guest
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota, United States of America
| | - Andrew J. Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota, United States of America
| |
Collapse
|
19
|
Wagner JD, Gelman A, Hancock KE, Chung Y, Delgutte B. Rabbits use both spectral and temporal cues to discriminate the fundamental frequency of harmonic complexes with missing fundamentals. J Neurophysiol 2022; 127:290-312. [PMID: 34879207 PMCID: PMC8759963 DOI: 10.1152/jn.00366.2021] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/03/2023] Open
Abstract
The pitch of harmonic complex tones (HCTs) common in speech, music, and animal vocalizations plays a key role in the perceptual organization of sound. Unraveling the neural mechanisms of pitch perception requires animal models, but little is known about complex pitch perception by animals, and some species appear to use different pitch mechanisms than humans. Here, we tested rabbits' ability to discriminate the fundamental frequency (F0) of HCTs with missing fundamentals, using a behavioral paradigm inspired by foraging behavior in which rabbits learned to harness a spatial gradient in F0 to find the location of a virtual target within a room for a food reward. Rabbits were initially trained to discriminate HCTs with F0s in the range 400-800 Hz and with harmonics covering a wide frequency range (800-16,000 Hz) and then tested with stimuli differing in spectral composition to test the role of harmonic resolvability (experiment 1) or in F0 range (experiment 2) or in both F0 and spectral content (experiment 3). Together, these experiments show that rabbits can discriminate HCTs over a wide F0 range (200-1,600 Hz) encompassing the range of conspecific vocalizations and can use either the spectral pattern of harmonics resolved by the cochlea for higher F0s or temporal envelope cues resulting from interaction between unresolved harmonics for lower F0s. The qualitative similarity of these results to human performance supports the use of rabbits as an animal model for studies of pitch mechanisms, providing species differences in cochlear frequency selectivity and F0 range of vocalizations are taken into account.NEW & NOTEWORTHY Understanding the neural mechanisms of pitch perception requires experiments in animal models, but little is known about pitch perception by animals. Here we show that rabbits, a popular animal in auditory neuroscience, can discriminate complex sounds differing in pitch using either spectral cues or temporal cues. The results suggest that the role of spectral cues in pitch perception by animals may have been underestimated by predominantly testing low frequencies in the range of human voice.
Collapse
Affiliation(s)
- Joseph D. Wagner
- 1Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, Massachusetts,3Department of Biomedical Engineering, Boston University, Boston, Massachusetts
| | - Alice Gelman
- 1Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, Massachusetts
| | - Kenneth E. Hancock
- 1Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, Massachusetts,2Department of Otolaryngology, Head and Neck Surgery, Harvard Medical School, Boston, Massachusetts
| | - Yoojin Chung
- 1Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, Massachusetts,2Department of Otolaryngology, Head and Neck Surgery, Harvard Medical School, Boston, Massachusetts
| | - Bertrand Delgutte
- 1Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, Massachusetts,2Department of Otolaryngology, Head and Neck Surgery, Harvard Medical School, Boston, Massachusetts
| |
Collapse
|
20
|
Altoè A, Charaziak KK, Dewey JB, Moleti A, Sisto R, Oghalai JS, Shera CA. The Elusive Cochlear Filter: Wave Origin of Cochlear Cross-Frequency Masking. J Assoc Res Otolaryngol 2021; 22:623-640. [PMID: 34677710 DOI: 10.1007/s10162-021-00814-2] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2020] [Accepted: 08/23/2021] [Indexed: 11/28/2022] Open
Abstract
The mammalian cochlea achieves its remarkable sensitivity, frequency selectivity, and dynamic range by spatially segregating the different frequency components of sound via nonlinear processes that remain only partially understood. As a consequence of the wave-based nature of cochlear processing, the different frequency components of complex sounds interact spatially and nonlinearly, mutually suppressing one another as they propagate. Because understanding nonlinear wave interactions and their effects on hearing appears to require mathematically complex or computationally intensive models, theories of hearing that do not deal specifically with cochlear mechanics have often neglected the spatial nature of suppression phenomena. Here we describe a simple framework consisting of a nonlinear traveling-wave model whose spatial response properties can be estimated from basilar-membrane (BM) transfer functions. Without invoking jazzy details of organ-of-Corti mechanics, the model accounts well for the peculiar frequency-dependence of suppression found in two-tone suppression experiments. In particular, our analysis shows that near the peak of the traveling wave, the amplitude of the BM response depends primarily on the nonlinear properties of the traveling wave in more basal (high-frequency) regions. The proposed framework provides perhaps the simplest representation of cochlear signal processing that accounts for the spatially distributed effects of nonlinear wave propagation. Shifting the perspective from local filters to non-local, spatially distributed processes not only elucidates the character of cochlear signal processing, but also has important consequences for interpreting psychophysical experiments.
Collapse
Affiliation(s)
- Alessandro Altoè
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA.
| | - Karolina K Charaziak
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA
| | - James B Dewey
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA
| | - Arturo Moleti
- Department of Physics, University of Roma Tor Vergata, Rome, Italy
| | - Renata Sisto
- DIMEILA, INAIL, Monte Porzio Catone, Rome, Italy
| | - John S Oghalai
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA
| | - Christopher A Shera
- Caruso Department of Otolaryngology Head & Neck Surgery, University of Southern California, CA, Los Angeles, USA.,Department of Physics and Astronomy, University of Southern California, Los Angeles, CA, USA
| |
Collapse
|
21
|
Heil P, Mohamed ESI, Matysiak A. Towards a unifying basis of auditory thresholds: Thresholds for multicomponent stimuli. Hear Res 2021; 410:108349. [PMID: 34530356 DOI: 10.1016/j.heares.2021.108349] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/17/2020] [Revised: 08/23/2021] [Accepted: 08/30/2021] [Indexed: 11/25/2022]
Abstract
Sounds consisting of multiple simultaneous or consecutive components can be detected by listeners when the stimulus levels of the components are lower than those needed to detect the individual components alone. The mechanisms underlying such spectral, spectrotemporal, temporal, or across-ear integration are not completely understood. Here, we report threshold measurements from human subjects for multicomponent stimuli (tone complexes, tone sequences, diotic or dichotic tones) and for their individual sinusoidal components in quiet. We examine whether the data are compatible with the detection model developed by Heil, Matysiak, and Neubauer (HMN model) to account for temporal integration (Heil et al. 2017), and we compare its performance to that of the statistical summation model (Green 1958), the model commonly used to account for spectral and spectrotemporal integration. In addition, we compare the performance of both models with respect to previously published thresholds for sequences of identical tones and for diotic tones. The HMN model is similar to the statistical summation model but is based on the assumption that the decision variable is a number of sensory events generated by the components via independent Poisson point processes. The rate of events is low without stimulation and increases with stimulation. The increase is proportional to the time-varying amplitude envelope of the bandpass-filtered component(s) raised to an exponent of 3. For an ideal observer, the decision variable is the sum of the events from all channels carrying information, for as long as they carry information. We find that the HMN model provides a better account of the thresholds for multicomponent stimuli than the statistical summation model, and it offers a unifying account of spectral, spectrotemporal, temporal, and across-ear integration at threshold.
Collapse
Affiliation(s)
- Peter Heil
- Department of Systems Physiology of Learning, Leibniz Institute for Neurobiology, Magdeburg 39118, Germany; Center for Behavioral Brain Sciences, Magdeburg, Germany.
| | - Esraa S I Mohamed
- Department of Systems Physiology of Learning, Leibniz Institute for Neurobiology, Magdeburg 39118, Germany
| | - Artur Matysiak
- Research Group Comparative Neuroscience, Leibniz Institute for Neurobiology, Magdeburg, Germany
| |
Collapse
|
22
|
Homma NY, Bajo VM. Lemniscal Corticothalamic Feedback in Auditory Scene Analysis. Front Neurosci 2021; 15:723893. [PMID: 34489635 PMCID: PMC8417129 DOI: 10.3389/fnins.2021.723893] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2021] [Accepted: 07/30/2021] [Indexed: 12/15/2022] Open
Abstract
Sound information is transmitted from the ear to central auditory stations of the brain via several nuclei. In addition to these ascending pathways there exist descending projections that can influence the information processing at each of these nuclei. A major descending pathway in the auditory system is the feedback projection from layer VI of the primary auditory cortex (A1) to the ventral division of medial geniculate body (MGBv) in the thalamus. The corticothalamic axons have small glutamatergic terminals that can modulate thalamic processing and thalamocortical information transmission. Corticothalamic neurons also provide input to GABAergic neurons of the thalamic reticular nucleus (TRN) that receives collaterals from the ascending thalamic axons. The balance of corticothalamic and TRN inputs has been shown to refine frequency tuning, firing patterns, and gating of MGBv neurons. Therefore, the thalamus is not merely a relay stage in the chain of auditory nuclei but does participate in complex aspects of sound processing that include top-down modulations. In this review, we aim (i) to examine how lemniscal corticothalamic feedback modulates responses in MGBv neurons, and (ii) to explore how the feedback contributes to auditory scene analysis, particularly on frequency and harmonic perception. Finally, we will discuss potential implications of the role of corticothalamic feedback in music and speech perception, where precise spectral and temporal processing is essential.
Collapse
Affiliation(s)
- Natsumi Y. Homma
- Center for Integrative Neuroscience, University of California, San Francisco, San Francisco, CA, United States
- Coleman Memorial Laboratory, Department of Otolaryngology – Head and Neck Surgery, University of California, San Francisco, San Francisco, CA, United States
| | - Victoria M. Bajo
- Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford, United Kingdom
| |
Collapse
|
23
|
Compression and amplification algorithms in hearing aids impair the selectivity of neural responses to speech. Nat Biomed Eng 2021; 6:717-730. [PMID: 33941898 PMCID: PMC7612903 DOI: 10.1038/s41551-021-00707-y] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2020] [Accepted: 02/25/2021] [Indexed: 02/07/2023]
Abstract
In quiet environments, hearing aids improve the perception of low-intensity sounds. However, for high-intensity sounds in background noise, the aids often fail to provide a benefit to the wearer. Here, by using large-scale single-neuron recordings from hearing-impaired gerbils — an established animal model of human hearing — we show that hearing aids restore the sensitivity of neural responses to speech, but not their selectivity. Rather than reflecting a deficit in supra-threshold auditory processing, the low selectivity is a consequence of hearing-aid compression (which decreases the spectral and temporal contrasts of incoming sound) and of amplification (which distorts neural responses, regardless of whether hearing is impaired). Processing strategies that avoid the trade-off between neural sensitivity and selectivity should improve the performance of hearing aids.
Collapse
|
24
|
Tani T, Koike-Tani M, Tran MT, Shribak M, Levic S. Postnatal structural development of mammalian Basilar Membrane provides anatomical basis for the maturation of tonotopic maps and frequency tuning. Sci Rep 2021; 11:7581. [PMID: 33828185 PMCID: PMC8027603 DOI: 10.1038/s41598-021-87150-w] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2020] [Accepted: 03/16/2021] [Indexed: 02/01/2023] Open
Abstract
The basilar membrane (BM) of the mammalian cochlea constitutes a spiraling acellular ribbon that is intimately attached to the organ of Corti. Its graded stiffness, increasing from apex to the base of the cochlea provides the mechanical basis for sound frequency analysis. Despite its central role in auditory signal transduction, virtually nothing is known about the BM's structural development. Using polarized light microscopy, the present study characterized the architectural transformations of freshly dissected BM at time points during postnatal development and maturation. The results indicate that the BM structural elements increase progressively in size, becoming radially aligned and more tightly packed with maturation and reach the adult structural signature by postnatal day 20 (P20). The findings provide insight into structural details and developmental changes of the mammalian BM, suggesting that BM is a dynamic structure that changes throughout the life of an animal.
Collapse
Affiliation(s)
- Tomomi Tani
- Marine Biological Laboratory, Eugene Bell Center, Woods Hole, MA, USA
- Biomedical Research Institute, National Institute of Advanced Industrial Science and Technology, Ikeda, Osaka, Japan
| | - Maki Koike-Tani
- Marine Biological Laboratory, Eugene Bell Center, Woods Hole, MA, USA
- Integrated Frontier Research for Medical Science Division, Institute for Open and Transdisciplinary Research Initiatives (OTRI), Osaka University, Suita, Osaka, Japan
| | - Mai Thi Tran
- Marine Biological Laboratory, Eugene Bell Center, Woods Hole, MA, USA
- College of Engineering and Computer Science, VinUniversity, Gia Lam District, Hanoi, Vietnam
| | - Michael Shribak
- Marine Biological Laboratory, Eugene Bell Center, Woods Hole, MA, USA
| | - Snezana Levic
- Marine Biological Laboratory, Eugene Bell Center, Woods Hole, MA, USA.
- Sensory Neuroscience Research Group, School of Pharmacy and Biomolecular Sciences, University of Brighton, Huxley Building, Brighton, BN2 4GJ, UK.
- Brighton and Sussex Medical School, University of Sussex, Brighton, BN1 9PX, UK.
| |
Collapse
|
25
|
Liu W, Luque M, Li H, Schrott-Fischer A, Glueckert R, Tylstedt S, Rajan G, Ladak H, Agrawal S, Rask-Andersen H. Spike Generators and Cell Signaling in the Human Auditory Nerve: An Ultrastructural, Super-Resolution, and Gene Hybridization Study. Front Cell Neurosci 2021; 15:642211. [PMID: 33796009 PMCID: PMC8008129 DOI: 10.3389/fncel.2021.642211] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2020] [Accepted: 02/22/2021] [Indexed: 11/23/2022] Open
Abstract
Background: The human auditory nerve contains 30,000 nerve fibers (NFs) that relay complex speech information to the brain with spectacular acuity. How speech is coded and influenced by various conditions is not known. It is also uncertain whether human nerve signaling involves exclusive proteins and gene manifestations compared with that of other species. Such information is difficult to determine due to the vulnerable, "esoteric," and encapsulated human ear surrounded by the hardest bone in the body. We collected human inner ear material for nanoscale visualization combining transmission electron microscopy (TEM), super-resolution structured illumination microscopy (SR-SIM), and RNA-scope analysis for the first time. Our aim was to gain information about the molecular instruments in human auditory nerve processing and deviations, and ways to perform electric modeling of prosthetic devices. Material and Methods: Human tissue was collected during trans-cochlear procedures to remove petro-clival meningioma after ethical permission. Cochlear neurons were processed for electron microscopy, confocal microscopy (CM), SR-SIM, and high-sensitive in situ hybridization for labeling single mRNA transcripts to detect ion channel and transporter proteins associated with nerve signal initiation and conductance. Results: Transport proteins and RNA transcripts were localized at the subcellular level. Hemi-nodal proteins were identified beneath the inner hair cells (IHCs). Voltage-gated ion channels (VGICs) were expressed in the spiral ganglion (SG) and axonal initial segments (AISs). Nodes of Ranvier (NR) expressed Nav1.6 proteins, and encoding genes critical for inter-cellular coupling were disclosed. Discussion: Our results suggest that initial spike generators are located beneath the IHCs in humans. The first NRs appear at different places. Additional spike generators and transcellular communication may boost, sharpen, and synchronize afferent signals by cell clusters at different frequency bands. These instruments may be essential for the filtering of complex sounds and may be challenged by various pathological conditions.
Collapse
Affiliation(s)
- Wei Liu
- Section of Otolaryngology, Department of Surgical Sciences, Head and Neck Surgery, Uppsala University Hospital, Uppsala, Sweden
| | - Maria Luque
- Department of Otorhinolaryngology, Medical University of Innsbruck, Innsbruck, Austria
| | - Hao Li
- Section of Otolaryngology, Department of Surgical Sciences, Head and Neck Surgery, Uppsala University Hospital, Uppsala, Sweden
| | | | - Rudolf Glueckert
- Department of Otorhinolaryngology, Medical University of Innsbruck, Innsbruck, Austria
| | - Sven Tylstedt
- Department of Olaryngology, Västerviks Hospital, Västervik, Sweden
| | - Gunesh Rajan
- Department of Otolaryngology, Head & Neck Surgery, Luzerner Kantonsspital, Luzern, Switzerland
- Department of Otolaryngology, Head & Neck Surgery, Division of Surgery, Medical School, University of Western Australia, Perth, WA, Australia
| | - Hanif Ladak
- Department of Otolaryngology-Head and Neck Surgery, Department of Medical Biophysics and Department of Electrical and Computer Engineering, Western University, London, ON, Canada
| | - Sumit Agrawal
- Department of Otolaryngology-Head and Neck Surgery, Western University, London, ON, Canada
| | - Helge Rask-Andersen
- Section of Otolaryngology, Department of Surgical Sciences, Head and Neck Surgery, Uppsala University Hospital, Uppsala, Sweden
| |
Collapse
|
26
|
de Cheveigné A. Harmonic Cancellation-A Fundamental of Auditory Scene Analysis. Trends Hear 2021; 25:23312165211041422. [PMID: 34698574 PMCID: PMC8552394 DOI: 10.1177/23312165211041422] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2021] [Revised: 07/23/2021] [Accepted: 07/09/2021] [Indexed: 11/16/2022] Open
Abstract
This paper reviews the hypothesis of harmonic cancellation according to which an interfering sound is suppressed or canceled on the basis of its harmonicity (or periodicity in the time domain) for the purpose of Auditory Scene Analysis. It defines the concept, discusses theoretical arguments in its favor, and reviews experimental results that support it, or not. If correct, the hypothesis may draw on time-domain processing of temporally accurate neural representations within the brainstem, as required also by the classic equalization-cancellation model of binaural unmasking. The hypothesis predicts that a target sound corrupted by interference will be easier to hear if the interference is harmonic than inharmonic, all else being equal. This prediction is borne out in a number of behavioral studies, but not all. The paper reviews those results, with the aim to understand the inconsistencies and come up with a reliable conclusion for, or against, the hypothesis of harmonic cancellation within the auditory system.
Collapse
Affiliation(s)
- Alain de Cheveigné
- Laboratoire des systèmes perceptifs, CNRS, Paris, France
- Département d’études cognitives, École normale supérieure, PSL
University, Paris, France
- UCL Ear Institute, London, UK
| |
Collapse
|
27
|
Cheatham MA. Comparing spontaneous and stimulus frequency otoacoustic emissions in mice with tectorial membrane defects. Hear Res 2020; 400:108143. [PMID: 33340968 DOI: 10.1016/j.heares.2020.108143] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/24/2020] [Revised: 11/20/2020] [Accepted: 12/02/2020] [Indexed: 01/28/2023]
Abstract
The global standing-wave model for generation of spontaneous otoacoustic emissions (SOAEs) suggests that they are amplitude-stabilized standing waves and that the spacing between SOAEs corresponds to the interval over which the phase changes by one cycle as determined from the phase-gradient delays of stimulus frequency otoacoustic emissions (SFOAEs). Because data characterizing the relationship between spontaneous and evoked emissions in nonhuman mammals are limited, we examined SOAEs and SFOAEs in tectorial membrane (TM) mutants and their controls. Computations indicate that the spacing between adjacent SOAEs is predicted by the SFOAE phase-gradient delays for TM mutants lacking Ceacam16, where SOAE frequencies are greater than ~20 kHz and the mutants retain near-normal hearing when young. Mice with a missense mutation in Tecta (TectaY1870C/+), as well as mice lacking Otoancorin (Otoa-/-), were also examined. Although these mutants exhibit hearing loss, they generate SOAEs with average frequencies of 11 kHz in TectaY1870C/+ and 6 kHz in Otoa-/-. In these animals, the spacing between adjacent SOAEs is larger than predicted by the SFOAE phase delays. It is also demonstrated that mice do not exhibit the strong frequency-dependence in signal coding that characterizes species with good low-frequency hearing. In fact, a transition occurs near the apical end of the mouse cochlea rather than at the mid-point along the cochlear partition. Hence, disagreements with the standing-wave model are not easily explained by a transition in tuning ratios between apical and basal regions of the cochlea, especially for SOAEs generated in TectaY1870C/+mice.
Collapse
Affiliation(s)
- Mary Ann Cheatham
- The Knowles Hearing Center, Roxelyn and Richard Pepper Department of Communication Sciences and Disorders, Northwestern University, 2-240 Frances Searle Building, 2240 Campus Drive, Evanston IL 60208, USA.
| |
Collapse
|
28
|
Rahman M, Willmore BDB, King AJ, Harper NS. Simple transformations capture auditory input to cortex. Proc Natl Acad Sci U S A 2020; 117:28442-28451. [PMID: 33097665 PMCID: PMC7668077 DOI: 10.1073/pnas.1922033117] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022] Open
Abstract
Sounds are processed by the ear and central auditory pathway. These processing steps are biologically complex, and many aspects of the transformation from sound waveforms to cortical response remain unclear. To understand this transformation, we combined models of the auditory periphery with various encoding models to predict auditory cortical responses to natural sounds. The cochlear models ranged from detailed biophysical simulations of the cochlea and auditory nerve to simple spectrogram-like approximations of the information processing in these structures. For three different stimulus sets, we tested the capacity of these models to predict the time course of single-unit neural responses recorded in ferret primary auditory cortex. We found that simple models based on a log-spaced spectrogram with approximately logarithmic compression perform similarly to the best-performing biophysically detailed models of the auditory periphery, and more consistently well over diverse natural and synthetic sounds. Furthermore, we demonstrated that including approximations of the three categories of auditory nerve fiber in these simple models can substantially improve prediction, particularly when combined with a network encoding model. Our findings imply that the properties of the auditory periphery and central pathway may together result in a simpler than expected functional transformation from ear to cortex. Thus, much of the detailed biological complexity seen in the auditory periphery does not appear to be important for understanding the cortical representation of sound.
Collapse
Affiliation(s)
- Monzilur Rahman
- Department of Physiology, Anatomy and Genetics, University of Oxford, OX1 3PT Oxford, United Kingdom
| | - Ben D B Willmore
- Department of Physiology, Anatomy and Genetics, University of Oxford, OX1 3PT Oxford, United Kingdom
| | - Andrew J King
- Department of Physiology, Anatomy and Genetics, University of Oxford, OX1 3PT Oxford, United Kingdom
| | - Nicol S Harper
- Department of Physiology, Anatomy and Genetics, University of Oxford, OX1 3PT Oxford, United Kingdom
| |
Collapse
|
29
|
Engler S, de Kleine E, Avan P, van Dijk P. Frequency selectivity of tonal language native speakers probed by suppression tuning curves of spontaneous otoacoustic emissions. Hear Res 2020; 398:108100. [PMID: 33125981 DOI: 10.1016/j.heares.2020.108100] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/12/2020] [Revised: 09/28/2020] [Accepted: 10/19/2020] [Indexed: 10/23/2022]
Abstract
Native acquisition of a tonal language (TL) is related to enhanced abilities of pitch perception and production, compared to non-tonal language (NTL) native speakers. Moreover, differences in brain responses to both linguistically relevant and non-relevant pitch changes have been described in TL native speakers. It is so far unclear to which extent differences are present at the peripheral processing level of the cochlea. To determine possible differences in cochlear frequency selectivity between Asian TL speakers and Caucasian NTL speakers, suppression tuning curves (STCs) of spontaneous otoacoustic emissions (SOAEs) were examined in both groups. By presenting pure tones, SOAE levels were suppressed and STCs were derived. SOAEs with center frequencies higher than 4.5 kHz were recorded only in female TL native speakers, which correlated with better high-frequency tone detection thresholds. The suppression thresholds at the tip of the STC and filter quality coefficient Q10dB did not differ significantly between both language groups. Thus, the characteristics of the STCs of SOAEs do not support the presence of differences in peripheral auditory processing between TL and NTL native speakers.
Collapse
Affiliation(s)
- Sina Engler
- Department of Otorhinolaryngology/Head and Neck Surgery, University of Groningen, University Medical Center Groningen, the Netherlands; Graduate School of Medical Sciences, Research School of Behavioural and Cognitive Neurosciences, University of Groningen, the Netherlands.
| | - Emile de Kleine
- Department of Otorhinolaryngology/Head and Neck Surgery, University of Groningen, University Medical Center Groningen, the Netherlands; Graduate School of Medical Sciences, Research School of Behavioural and Cognitive Neurosciences, University of Groningen, the Netherlands
| | - Paul Avan
- Laboratory of Neurosensory Biophysics, University Clermont Auvergne, Laboratory of Neurosensory Biophysics, UMR INSERM 1107, Clermont-Ferrand, France; School of Medicine, 28 Place Henri Dunant, Clermont-Ferrand 63000, France
| | - Pim van Dijk
- Department of Otorhinolaryngology/Head and Neck Surgery, University of Groningen, University Medical Center Groningen, the Netherlands; Graduate School of Medical Sciences, Research School of Behavioural and Cognitive Neurosciences, University of Groningen, the Netherlands
| |
Collapse
|
30
|
Whiteford KL, Kreft HA, Oxenham AJ. The role of cochlear place coding in the perception of frequency modulation. eLife 2020; 9:58468. [PMID: 32996463 PMCID: PMC7556860 DOI: 10.7554/elife.58468] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2020] [Accepted: 09/29/2020] [Indexed: 12/17/2022] Open
Abstract
Natural sounds convey information via frequency and amplitude modulations (FM and AM). Humans are acutely sensitive to the slow rates of FM that are crucial for speech and music. This sensitivity has long been thought to rely on precise stimulus-driven auditory-nerve spike timing (time code), whereas a coarser code, based on variations in the cochlear place of stimulation (place code), represents faster FM rates. We tested this theory in listeners with normal and impaired hearing, spanning a wide range of place-coding fidelity. Contrary to predictions, sensitivity to both slow and fast FM correlated with place-coding fidelity. We also used incoherent AM on two carriers to simulate place coding of FM and observed poorer sensitivity at high carrier frequencies and fast rates, two properties of FM detection previously ascribed to the limits of time coding. The results suggest a unitary place-based neural code for FM across all rates and carrier frequencies.
Collapse
Affiliation(s)
- Kelly L Whiteford
- Department of Psychology, University of Minnesota, Minneapolis, United States
| | - Heather A Kreft
- Department of Psychology, University of Minnesota, Minneapolis, United States
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, United States
| |
Collapse
|
31
|
Wilson US, Browning-Kamins J, Boothalingam S, Moleti A, Sisto R, Dhar S. Relationship Between Behavioral and Stimulus Frequency Otoacoustic Emissions Delay-Based Tuning Estimates. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2020; 63:1958-1968. [PMID: 32464079 PMCID: PMC7839027 DOI: 10.1044/2020_jslhr-19-00386] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/09/2019] [Revised: 02/18/2020] [Accepted: 02/20/2020] [Indexed: 06/11/2023]
Abstract
Purpose The phase delay of stimulus frequency otoacoustic emissions (SFOAEs) has been proposed as a noninvasive, objective, and fast source for estimating cochlear mechanical tuning. However, the implementation of SFOAEs clinically has been thwarted by the gaps in understanding of the stability of SFOAE delay-based tuning estimates and their relationship to behavioral measures of tuning. Therefore, the goals of this study were (a) to investigate the relationship between delay-based tuning estimates from SFOAEs and simultaneously masked psychophysical tuning curves (PTCs) and (b) to assess the across- and within-session repeatability of tuning estimates from behavioral and OAE measures. Method Three sets of behavioral and OAE measurements were collected in 24 normal-hearing, young adults for two probe frequencies, 1 and 4 kHz. For each participant, delay-based tuning estimates were derived from the phase gradient of SFOAEs. SFOAE-based and behavioral estimates of tuning obtained using the fast-swept PTC paradigm were compared within and across sessions. Results In general, tuning estimates were sharper at 4 kHz compared to 1 kHz for both PTCs and SFOAEs. Statistical analyses revealed a significant correlation between SFOAE delay-based tuning and PTCs at 4 kHz, but not 1 kHz. Lastly, SFOAE delay-based tuning estimates showed better intra- and intersession repeatability compared to PTCs. Conclusions SFOAE phase-gradient delays reflect aspects of cochlear mechanical tuning, in that a frequency dependence similar to that of basilar membrane tuning was observed. Furthermore, the significant correlation with PTCs at 4 kHz and the high repeatability of SFOAE-based tuning measures offer promise of an objective, nonbehavioral assay of tuning in human ears.
Collapse
Affiliation(s)
- Uzma Shaheen Wilson
- Roxelyn and Richard Pepper Department of Communication Sciences & Disorders, Northwestern University, Evanston, IL
| | - Jenna Browning-Kamins
- Roxelyn and Richard Pepper Department of Communication Sciences & Disorders, Northwestern University, Evanston, IL
| | - Sriram Boothalingam
- Roxelyn and Richard Pepper Department of Communication Sciences & Disorders, Northwestern University, Evanston, IL
- Waisman Center, University of Wisconsin-Madison, Madison, WI
| | - Arturo Moleti
- Physics Department, University of Roma Tor Vergata, Rome, Italy
| | - Renata Sisto
- Istituto Nazionale Assicurazione Infortuni sul Lavoro Research, Rome, Italy
| | - Sumitrajit Dhar
- Roxelyn and Richard Pepper Department of Communication Sciences & Disorders, Northwestern University, Evanston, IL
- Knowles Hearing Center, Northwestern University, Evanston, IL
| |
Collapse
|
32
|
Russell IJ, Lukashkina VA, Levic S, Cho YW, Lukashkin AN, Ng L, Forrest D. Emilin 2 promotes the mechanical gradient of the cochlear basilar membrane and resolution of frequencies in sound. SCIENCE ADVANCES 2020; 6:eaba2634. [PMID: 32577518 PMCID: PMC7286672 DOI: 10.1126/sciadv.aba2634] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/19/2019] [Accepted: 04/02/2020] [Indexed: 06/11/2023]
Abstract
The detection of different frequencies in sound is accomplished with remarkable precision by the basilar membrane (BM), an elastic, ribbon-like structure with graded stiffness along the cochlear spiral. Sound stimulates a wave of displacement along the BM with maximal magnitude at precise, frequency-specific locations to excite neural signals that carry frequency information to the brain. Perceptual frequency discrimination requires fine resolution of this frequency map, but little is known of the intrinsic molecular features that demarcate the place of response on the BM. To investigate the role of BM microarchitecture in frequency discrimination, we deleted extracellular matrix protein emilin 2, which disturbed the filamentous organization in the BM. Emilin2 -/- mice displayed broadened mechanical and neural frequency tuning with multiple response peaks that are shifted to lower frequencies than normal. Thus, emilin 2 confers a stiffness gradient on the BM that is critical for accurate frequency resolution.
Collapse
Affiliation(s)
- Ian J. Russell
- Sensory Neuroscience Research Group, School of Pharmacy and Biomolecular Sciences, University of Brighton, Huxley Building, Brighton BN2 4GJ, UK
| | - Victoria A. Lukashkina
- Sensory Neuroscience Research Group, School of Pharmacy and Biomolecular Sciences, University of Brighton, Huxley Building, Brighton BN2 4GJ, UK
| | - Snezana Levic
- Sensory Neuroscience Research Group, School of Pharmacy and Biomolecular Sciences, University of Brighton, Huxley Building, Brighton BN2 4GJ, UK
- Brighton and Sussex Medical School, University of Sussex, Brighton BN1 9PX, UK
| | - Young-Wook Cho
- Laboratory of Endocrinology and Receptor Biology, NIDDK, National Institutes of Health, 10 Center Drive, Bethesda, MD 20892, USA
| | - Andrei N. Lukashkin
- Sensory Neuroscience Research Group, School of Pharmacy and Biomolecular Sciences, University of Brighton, Huxley Building, Brighton BN2 4GJ, UK
| | - Lily Ng
- Laboratory of Endocrinology and Receptor Biology, NIDDK, National Institutes of Health, 10 Center Drive, Bethesda, MD 20892, USA
| | - Douglas Forrest
- Laboratory of Endocrinology and Receptor Biology, NIDDK, National Institutes of Health, 10 Center Drive, Bethesda, MD 20892, USA
| |
Collapse
|
33
|
Vencovský V, Vetešník A, Gummer AW. Nonlinear reflection as a cause of the short-latency component in stimulus-frequency otoacoustic emissions simulated by the methods of compression and suppression. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:3992. [PMID: 32611132 DOI: 10.1121/10.0001394] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/19/2020] [Accepted: 05/22/2020] [Indexed: 06/11/2023]
Abstract
Stimulus-frequency otoacoustic emissions (SFOAEs) are generated by coherent reflection of forward traveling waves by perturbations along the basilar membrane. The strongest wavelets are backscattered near the place where the traveling wave reaches its maximal amplitude (tonotopic place). Therefore, the SFOAE group delay might be expected to be twice the group delay estimated in the cochlear filters. However, experimental data have yielded steady-state SFOAE components with near-zero latency. A cochlear model is used to show that short-latency SFOAE components can be generated due to nonlinear reflection of the compressor or suppressor tones used in SFOAE measurements. The simulations indicate that suppressors produce more pronounced short-latency components than compressors. The existence of nonlinear reflection components due to suppressors can also explain why SFOAEs can still be detected when suppressors are presented more than half an octave above the probe-tone frequency. Simulations of the SFOAE suppression tuning curves showed that phase changes in the SFOAE residual as the suppressor frequency increases are mostly determined by phase changes of the nonlinear reflection component.
Collapse
Affiliation(s)
- Václav Vencovský
- Department of Radioelectronics, Czech Technical University in Prague, Technická 2, 166 27 Prague, Czech Republic
| | - Aleš Vetešník
- Department of Nuclear Chemistry, Czech Technical University in Prague, Břehová 7, 115 19 Prague, Czech Republic
| | - Anthony W Gummer
- Department of Otolaryngology, Section of Physiological Acoustics and Communication, Eberhard-Karls-University Tübingen, Elfriede-Aulhorn-Strasse 5, 72076 Tübingen, Germany
| |
Collapse
|
34
|
Sumner CJ, Sumner S. Signal detection: applying analysis methods from psychology to animal behaviour. Philos Trans R Soc Lond B Biol Sci 2020; 375:20190480. [PMID: 32420861 DOI: 10.1098/rstb.2019.0480] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
Conspecific acceptance thresholds (Reeve 1989 Am. Nat. 133, 407-435), which have been widely applied to explain ecological behaviour in animals, proposed how sensory information, prior information and the costs of decisions determine actions. Signal detection theory (Green & Swets 1966 Signal detection theory and psychophysics; SDT), which forms the basis of CAT models, has been widely used in psychological studies to partition the ability to discriminate sensory information from the action made as a result of it. In this article, we will review the application of SDT in interpreting the behaviour of laboratory animals trained in operant conditioning tasks and then consider its potential in ecological studies of animal behaviour in natural environments. Focusing on the nest-mate recognition systems exhibited by social insects, we show how the quantitative application of SDT has the potential to transform acceptance rate data into independent indices of cue sensitivity and decision criterion (also known as the acceptance threshold). However, further tests of the assumptions underlying SDT analysis are required. Overall, we argue that SDT, as conventionally applied in psychological studies, may provide clearer insights into the mechanistic basis of decision making and information processing in behavioural ecology. This article is part of the theme issue 'Signal detection theory in recognition systems: from evolving models to experimental tests'.
Collapse
Affiliation(s)
- Christian J Sumner
- Department of Psychology, Nottingham Trent University, 50 Shakespeare Street, Nottingham NG1 4FQ, UK
| | - Seirian Sumner
- Centre for Biodiversity and Environmental Research, Department of Genetics Evolution and Environment, Division of Biosciences, University College London, Gower Street, London WC1E 6BT, UK
| |
Collapse
|
35
|
Maxwell BN, Richards VM, Carney LH. Neural fluctuation cues for simultaneous notched-noise masking and profile-analysis tasks: Insights from model midbrain responses. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:3523. [PMID: 32486827 PMCID: PMC7229985 DOI: 10.1121/10.0001226] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/10/2019] [Revised: 04/05/2020] [Accepted: 04/21/2020] [Indexed: 05/19/2023]
Abstract
Results of simultaneous notched-noise masking are commonly interpreted as reflecting the bandwidth of underlying auditory filters. This interpretation assumes that listeners detect a tone added to notched-noise based on an increase in energy at the output of an auditory filter. Previous work challenged this assumption by showing that randomly and independently varying (roving) the levels of each stimulus interval does not substantially worsen listener thresholds [Lentz, Richards, and Matiasek (1999). J. Acoust. Soc. Am. 106, 2779-2792]. Lentz et al. further challenged this assumption by showing that filter bandwidths based on notched-noise results were different from those based on a profile-analysis task [Green (1983). Am. Psychol. 38, 133-142; (1988). (Oxford University Press, New York)], although these estimates were later reconciled by emphasizing spectral peaks of the profile-analysis stimulus [Lentz (2006). J. Acoust. Soc. Am. 120, 945-956]. Here, a single physiological model is shown to account for performance in fixed- and roving-level notched-noise tasks and the Lentz et al. profile-analysis task. This model depends on peripheral neural fluctuation cues that are transformed into the average rates of model inferior colliculus neurons. Neural fluctuations are influenced by peripheral filters, synaptic adaptation, cochlear amplification, and saturation of inner hair cells, an element not included in previous theories of envelope-based cues for these tasks. Results suggest reevaluation of the interpretation of performance in these paradigms.
Collapse
Affiliation(s)
- Braden N Maxwell
- Departments of Biomedical Engineering and Neuroscience, 601 Elmwood Avenue, University of Rochester, Rochester, New York 14642, USA
| | - Virginia M Richards
- Department of Cognitive Sciences, University of California, 3151 Social Science Plaza, Irvine, California 92697-5100, USA
| | - Laurel H Carney
- Departments of Biomedical Engineering and Neuroscience, 601 Elmwood Avenue, University of Rochester, Rochester, New York 14642, USA
| |
Collapse
|
36
|
Mehta AH, Oxenham AJ. Effect of lowest harmonic rank on fundamental-frequency difference limens varies with fundamental frequency. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:2314. [PMID: 32359332 PMCID: PMC7166120 DOI: 10.1121/10.0001092] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/19/2019] [Revised: 03/25/2020] [Accepted: 03/27/2020] [Indexed: 06/11/2023]
Abstract
This study investigated the relationship between fundamental frequency difference limens (F0DLs) and the lowest harmonic number present over a wide range of F0s (30-2000 Hz) for 12-component harmonic complex tones that were presented in either sine or random phase. For fundamental frequencies (F0s) between 100 and 400 Hz, a transition from low (∼1%) to high (∼5%) F0DLs occurred as the lowest harmonic number increased from about seven to ten, in line with earlier studies. At lower and higher F0s, the transition between low and high F0DLs occurred at lower harmonic numbers. The worsening performance at low F0s was reasonably well predicted by the expected decrease in spectral resolution below about 500 Hz. At higher F0s, the degradation in performance at lower harmonic numbers could not be predicted by changes in spectral resolution but remained relatively good (<2%-3%) in some conditions, even when all harmonics were above 8 kHz, confirming that F0 can be extracted from harmonics even when temporal envelope or fine-structure cues are weak or absent.
Collapse
Affiliation(s)
- Anahita H Mehta
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| |
Collapse
|
37
|
Deloche F. Fine-grained statistical structure of speech. PLoS One 2020; 15:e0230233. [PMID: 32196513 PMCID: PMC7083313 DOI: 10.1371/journal.pone.0230233] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2019] [Accepted: 02/25/2020] [Indexed: 12/04/2022] Open
Abstract
In spite of its acoustic diversity, the speech signal presents statistical regularities that can be exploited by biological or artificial systems for efficient coding. Independent Component Analysis (ICA) revealed that on small time scales (∼ 10 ms), the overall structure of speech is well captured by a time-frequency representation whose frequency selectivity follows the same power law in the high frequency range 1–8 kHz as cochlear frequency selectivity in mammals. Variations in the power-law exponent, i.e. different time-frequency trade-offs, have been shown to provide additional adaptation to phonetic categories. Here, we adopt a parametric approach to investigate the variations of the exponent at a finer level of speech. The estimation procedure is based on a measure that reflects the sparsity of decompositions in a set of Gabor dictionaries whose atoms are Gaussian-modulated sinusoids. We examine the variations of the exponent associated with the best decomposition, first at the level of phonemes, then at an intra-phonemic level. We show that this analysis offers a rich interpretation of the fine-grained statistical structure of speech, and that the exponent values can be related to key acoustic properties. Two main results are: i) for plosives, the exponent is lowered by the release bursts, concealing higher values during the opening phases; ii) for vowels, the exponent is bound to formant bandwidths and decreases with the degree of acoustic radiation at the lips. This work further suggests that an efficient coding strategy is to reduce frequency selectivity with sound intensity level, congruent with the nonlinear behavior of cochlear filtering.
Collapse
Affiliation(s)
- François Deloche
- Centre d’analyse et de mathématique sociales, CNRS, EHESS, Paris, France
- * E-mail:
| |
Collapse
|
38
|
Mehta AH, Lu H, Oxenham AJ. The Perception of Multiple Simultaneous Pitches as a Function of Number of Spectral Channels and Spectral Spread in a Noise-Excited Envelope Vocoder. J Assoc Res Otolaryngol 2020; 21:61-72. [PMID: 32048077 DOI: 10.1007/s10162-019-00738-y] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2018] [Accepted: 10/30/2019] [Indexed: 01/06/2023] Open
Abstract
Cochlear implant (CI) listeners typically perform poorly on tasks involving the pitch of complex tones. This limitation in performance is thought to be mainly due to the restricted number of active channels and the broad current spread that leads to channel interactions and subsequent loss of precise spectral information, with temporal information limited primarily to temporal-envelope cues. Little is known about the degree of spectral resolution required to perceive combinations of multiple pitches, or a single pitch in the presence of other interfering tones in the same spectral region. This study used noise-excited envelope vocoders that simulate the limited resolution of CIs to explore the perception of multiple pitches presented simultaneously. The results show that the resolution required for perceiving multiple complex pitches is comparable to that found in a previous study using single complex tones. Although relatively high performance can be achieved with 48 channels, performance remained near chance when even limited spectral spread (with filter slopes as steep as 144 dB/octave) was introduced to the simulations. Overall, these tight constraints suggest that current CI technology will not be able to convey the pitches of combinations of spectrally overlapping complex tones.
Collapse
Affiliation(s)
- Anahita H Mehta
- Department of Psychology, University of Minnesota, N218 Elliott Hall, 75 East River Parkway, Minneapolis, MN, 55455, USA.
| | - Hao Lu
- Department of Psychology, University of Minnesota, N218 Elliott Hall, 75 East River Parkway, Minneapolis, MN, 55455, USA
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, N218 Elliott Hall, 75 East River Parkway, Minneapolis, MN, 55455, USA
| |
Collapse
|
39
|
Robust Rate-Place Coding of Resolved Components in Harmonic and Inharmonic Complex Tones in Auditory Midbrain. J Neurosci 2020; 40:2080-2093. [PMID: 31996454 DOI: 10.1523/jneurosci.2337-19.2020] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2019] [Revised: 01/12/2020] [Accepted: 01/16/2020] [Indexed: 11/21/2022] Open
Abstract
Harmonic complex tones (HCTs) commonly occurring in speech and music evoke a strong pitch at their fundamental frequency (F0), especially when they contain harmonics individually resolved by the cochlea. When all frequency components of an HCT are shifted by the same amount, the pitch of the resulting inharmonic tone (IHCT) can also shift, although the envelope repetition rate is unchanged. A rate-place code, whereby resolved harmonics are represented by local maxima in firing rates along the tonotopic axis, has been characterized in the auditory nerve and primary auditory cortex, but little is known about intermediate processing stages. We recorded single-neuron responses to HCT and IHCT with varying F0 and sound level in the inferior colliculus (IC) of unanesthetized rabbits of both sexes. Many neurons showed peaks in firing rate when a low-numbered harmonic aligned with the neuron's characteristic frequency, demonstrating "rate-place" coding. The IC rate-place code was most prevalent for F0 > 800 Hz, was only moderately dependent on sound level over a 40 dB range, and was not sensitive to stimulus harmonicity. A spectral receptive-field model incorporating broadband inhibition better predicted the neural responses than a purely excitatory model, suggesting an enhancement of the rate-place representation by inhibition. Some IC neurons showed facilitation in response to HCT relative to pure tones, similar to cortical "harmonic template neurons" (Feng and Wang, 2017), but to a lesser degree. Our findings shed light on the transformation of rate-place coding of resolved harmonics along the auditory pathway.SIGNIFICANCE STATEMENT Harmonic complex tones are ubiquitous in speech and music and produce strong pitch percepts when they contain frequency components that are individually resolved by the cochlea. Here, we characterize a "rate-place" code for resolved harmonics in the auditory midbrain that is more robust across sound levels than the peripheral rate-place code and insensitive to the harmonic relationships among frequency components. We use a computational model to show that inhibition may play an important role in shaping the rate-place code. Our study fills a major gap in understanding the transformations in neural representations of resolved harmonics along the auditory pathway.
Collapse
|
40
|
Moore BCJ, Wan J, Varathanathan A, Naddell S, Baer T. No Effect of Musical Training on Frequency Selectivity Estimated Using Three Methods. Trends Hear 2019; 23:2331216519841980. [PMID: 31081487 DOI: 10.1177/2331216519841980] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022] Open
Abstract
It is widely believed that the frequency selectivity of the auditory system is largely determined by processes occurring in the cochlea. If so, musical training would not be expected to influence frequency selectivity. Consistent with this, auditory filter shapes for low center frequencies do not differ for musicians and nonmusicians. However, it has been reported that psychophysical tuning curves (PTCs) at 4000 Hz were sharper for musicians than for nonmusicians. This study explored the origin of the discrepancy across studies. Frequency selectivity was estimated for musicians and nonmusicians using three methods: fast PTCs with a masker that swept in frequency, "traditional" PTCs obtained using several fixed masker center frequencies, and the notched-noise method. The signal frequency was 4000 Hz. The data were fitted assuming that each side of the auditory filter had the shape of a rounded-exponential function. The sharpness of the auditory filters, estimated as the Q10 values, did not differ significantly between musicians and nonmusicians for any of the methods, but detection efficiency tended to be higher for the musicians. This is consistent with the idea that musicianship influences auditory proficiency but does not influence the peripheral processes that determine the frequency selectivity of the auditory system.
Collapse
Affiliation(s)
| | - Jie Wan
- 1 Department of Psychology, University of Cambridge, UK.,2 Research School of Behavioural and Cognitive Neurosciences, University of Groningen, the Netherlands
| | | | | | - Thomas Baer
- 1 Department of Psychology, University of Cambridge, UK
| |
Collapse
|
41
|
Su Y, Delgutte B. Pitch of harmonic complex tones: rate and temporal coding of envelope repetition rate in inferior colliculus of unanesthetized rabbits. J Neurophysiol 2019; 122:2468-2485. [PMID: 31664871 DOI: 10.1152/jn.00512.2019] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Harmonic complex tones (HCTs) found in speech, music, and animal vocalizations evoke strong pitch percepts at their fundamental frequencies. The strongest pitches are produced by HCTs that contain harmonics resolved by cochlear frequency analysis, but HCTs containing solely unresolved harmonics also evoke a weaker pitch at their envelope repetition rate (ERR). In the auditory periphery, neurons phase lock to the stimulus envelope, but this temporal representation of ERR degrades and gives way to rate codes along the ascending auditory pathway. To assess the role of the inferior colliculus (IC) in such transformations, we recorded IC neuron responses to HCT and sinusoidally modulated broadband noise (SAMN) with varying ERR from unanesthetized rabbits. Different interharmonic phase relationships of HCT were used to manipulate the temporal envelope without changing the power spectrum. Many IC neurons demonstrated band-pass rate tuning to ERR between 60 and 1,600 Hz for HCT and between 40 and 500 Hz for SAMN. The tuning was not related to the pure-tone best frequency of neurons but was dependent on the shape of the stimulus envelope, indicating a temporal rather than spectral origin. A phenomenological model suggests that the tuning may arise from peripheral temporal response patterns via synaptic inhibition. We also characterized temporal coding to ERR. Some IC neurons could phase lock to the stimulus envelope up to 900 Hz for either HCT or SAMN, but phase locking was weaker with SAMN. Together, the rate code and the temporal code represent a wide range of ERR, providing strong cues for the pitch of unresolved harmonics.NEW & NOTEWORTHY Envelope repetition rate (ERR) provides crucial cues for pitch perception of frequency components that are not individually resolved by the cochlea, but the neural representation of ERR for stimuli containing many harmonics is poorly characterized. Here we show that the pitch of stimuli with unresolved harmonics is represented by both a rate code and a temporal code for ERR in auditory midbrain neurons and propose possible underlying neural mechanisms with a computational model.
Collapse
Affiliation(s)
- Yaqing Su
- Eaton-Peabody Labs, Massachusetts Eye and Ear, Boston, Massachusetts.,Department of Biomedical Engineering, Boston University, Boston, Massachusetts
| | - Bertrand Delgutte
- Eaton-Peabody Labs, Massachusetts Eye and Ear, Boston, Massachusetts.,Department of Otolaryngology, Harvard Medical School, Boston, Massachusetts
| |
Collapse
|
42
|
Cochlear partition anatomy and motion in humans differ from the classic view of mammals. Proc Natl Acad Sci U S A 2019; 116:13977-13982. [PMID: 31235601 DOI: 10.1073/pnas.1900787116] [Citation(s) in RCA: 17] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Mammals detect sound through mechanosensitive cells of the cochlear organ of Corti that rest on the basilar membrane (BM). Motions of the BM and organ of Corti have been studied at the cochlear base in various laboratory animals, and the assumption has been that the cochleas of all mammals work similarly. In the classic view, the BM attaches to a stationary osseous spiral lamina (OSL), the tectorial membrane (TM) attaches to the limbus above the stationary OSL, and the BM is the major moving element, with a peak displacement near its center. Here, we measured the motion and studied the anatomy of the human cochlear partition (CP) at the cochlear base of fresh human cadaveric specimens. Unlike the classic view, we identified a soft-tissue structure between the BM and OSL in humans, which we name the CP "bridge." We measured CP transverse motion in humans and found that the OSL moved like a plate hinged near the modiolus, with motion increasing from the modiolus to the bridge. The bridge moved almost as much as the BM, with the maximum CP motion near the bridge-BM connection. BM motion accounts for 100% of CP volume displacement in the classic view, but accounts for only 27 to 43% in the base of humans. In humans, the TM-limbus attachment is above the moving bridge, not above a fixed structure. These results challenge long-held assumptions about cochlear mechanics in humans. In addition, animal apical anatomy (in SI Appendix) doesn't always fit the classic view.
Collapse
|
43
|
Norman-Haignere SV, Kanwisher N, McDermott JH, Conway BR. Divergence in the functional organization of human and macaque auditory cortex revealed by fMRI responses to harmonic tones. Nat Neurosci 2019; 22:1057-1060. [PMID: 31182868 PMCID: PMC6592717 DOI: 10.1038/s41593-019-0410-7] [Citation(s) in RCA: 29] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/10/2018] [Accepted: 04/19/2019] [Indexed: 12/02/2022]
Abstract
We report a difference between humans and macaque monkeys in the functional organization of cortical regions implicated in pitch perception: humans but not macaques showed regions with a strong preference for harmonic sounds compared to noise, measured with both synthetic tones and macaque vocalizations. In contrast, frequency-selective tonotopic maps were similar between the two species. This species difference may be driven by the unique demands of speech and music perception in humans.
Collapse
Affiliation(s)
- Sam V Norman-Haignere
- Zuckerman Institute for Mind, Brain and Behavior, Columbia University, New York, NY, USA. .,Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA. .,HHMI Postdoctoral Fellow of the Life Sciences Research Institute, Chevy Chase, MD, USA. .,Laboratoire des Systèmes Perceptifs, Département d'Études Cognitives, École Normale Supérieure, PSL University, CNRS, Paris, France.
| | - Nancy Kanwisher
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA.,McGovern Institute for Brain Research, Cambridge, MA, USA.,Center for Minds, Brains and Machines, Cambridge, MA, USA
| | - Josh H McDermott
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA.,McGovern Institute for Brain Research, Cambridge, MA, USA.,Center for Minds, Brains and Machines, Cambridge, MA, USA.,Program in Speech and Hearing Biosciences and Technology, Harvard University, Cambridge, MA, USA
| | - Bevil R Conway
- Laboratory of Sensorimotor Research, NEI, NIH, Bethesda, MD, USA. .,National Institute of Mental Health, NIH, Bethesda, MD, USA. .,National Institute of Neurological Disease and Stroke, NIH, Bethesda, MD, USA.
| |
Collapse
|
44
|
Graves JE, Oxenham AJ. Pitch discrimination with mixtures of three concurrent harmonic complexes. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 145:2072. [PMID: 31046318 PMCID: PMC6469983 DOI: 10.1121/1.5096639] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/13/2018] [Revised: 02/19/2019] [Accepted: 03/13/2019] [Indexed: 06/09/2023]
Abstract
In natural listening contexts, especially in music, it is common to hear three or more simultaneous pitches, but few empirical or theoretical studies have addressed how this is achieved. Place and pattern-recognition theories of pitch require at least some harmonics to be spectrally resolved for pitch to be extracted, but it is unclear how often such conditions exist when multiple complex tones are presented together. In three behavioral experiments, mixtures of three concurrent complexes were filtered into a single bandpass spectral region, and the relationship between the fundamental frequencies and spectral region was varied in order to manipulate the extent to which harmonics were resolved either before or after mixing. In experiment 1, listeners discriminated major from minor triads (a difference of 1 semitone in one note of the triad). In experiments 2 and 3, listeners compared the pitch of a probe tone with that of a subsequent target, embedded within two other tones. All three experiments demonstrated above-chance performance, even in conditions where the combinations of harmonic components were unlikely to be resolved after mixing, suggesting that fully resolved harmonics may not be necessary to extract the pitch from multiple simultaneous complexes.
Collapse
Affiliation(s)
- Jackson E Graves
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, Minnesota 55455, USA
| |
Collapse
|
45
|
The upper frequency limit for the use of phase locking to code temporal fine structure in humans: A compilation of viewpoints. Hear Res 2019; 377:109-121. [PMID: 30927686 PMCID: PMC6524635 DOI: 10.1016/j.heares.2019.03.011] [Citation(s) in RCA: 56] [Impact Index Per Article: 11.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/30/2018] [Revised: 02/09/2019] [Accepted: 03/13/2019] [Indexed: 11/27/2022]
Abstract
The relative importance of neural temporal and place coding in auditory perception is still a matter of much debate. The current article is a compilation of viewpoints from leading auditory psychophysicists and physiologists regarding the upper frequency limit for the use of neural phase locking to code temporal fine structure in humans. While phase locking is used for binaural processing up to about 1500 Hz, there is disagreement regarding the use of monaural phase-locking information at higher frequencies. Estimates of the general upper limit proposed by the contributors range from 1500 to 10000 Hz. The arguments depend on whether or not phase locking is needed to explain psychophysical discrimination performance at frequencies above 1500 Hz, and whether or not the phase-locked neural representation is sufficiently robust at these frequencies to provide useable information. The contributors suggest key experiments that may help to resolve this issue, and experimental findings that may cause them to change their minds. This issue is of crucial importance to our understanding of the neural basis of auditory perception in general, and of pitch perception in particular. Phase locking is used in binaural processing for frequencies up to ∼1500 Hz. Estimates of the general upper limit (inc. monaural processing) vary from 1500 to 10000 Hz. Direct recordings from human auditory nerve would determine peripheral limitation. Understanding of the central processing of temporal and place cues is needed to establish an upper limit.
Collapse
|
46
|
Walker KM, Gonzalez R, Kang JZ, McDermott JH, King AJ. Across-species differences in pitch perception are consistent with differences in cochlear filtering. eLife 2019; 8:41626. [PMID: 30874501 PMCID: PMC6435318 DOI: 10.7554/elife.41626] [Citation(s) in RCA: 20] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2018] [Accepted: 03/14/2019] [Indexed: 11/13/2022] Open
Abstract
Pitch perception is critical for recognizing speech, music and animal vocalizations, but its neurobiological basis remains unsettled, in part because of divergent results across species. We investigated whether species-specific differences exist in the cues used to perceive pitch and whether these can be accounted for by differences in the auditory periphery. Ferrets accurately generalized pitch discriminations to untrained stimuli whenever temporal envelope cues were robust in the probe sounds, but not when resolved harmonics were the main available cue. By contrast, human listeners exhibited the opposite pattern of results on an analogous task, consistent with previous studies. Simulated cochlear responses in the two species suggest that differences in the relative salience of the two pitch cues can be attributed to differences in cochlear filter bandwidths. The results support the view that cross-species variation in pitch perception reflects the constraints of estimating a sound’s fundamental frequency given species-specific cochlear tuning.
Collapse
Affiliation(s)
- Kerry Mm Walker
- Department of Physiology, Anatomy & Genetics, University of Oxford, Oxford, United Kingdom
| | - Ray Gonzalez
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, United States
| | - Joe Z Kang
- Department of Physiology, Anatomy & Genetics, University of Oxford, Oxford, United Kingdom
| | - Josh H McDermott
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, United States.,Program in Speech and Hearing Biosciences and Technology, Harvard University, Cambridge, United States
| | - Andrew J King
- Department of Physiology, Anatomy & Genetics, University of Oxford, Oxford, United Kingdom
| |
Collapse
|