1
|
Ou Y, Allen EJ, Kay KN, Oxenham AJ. Cortical substrates of perceptual confusion between pitch and timbre. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2025:2025.02.19.639197. [PMID: 40027705 PMCID: PMC11870464 DOI: 10.1101/2025.02.19.639197] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
Pitch and timbre are two fundamental perceptual attributes of sound that help us distinguish voices in speech and appreciate music. Brightness, one of the primary dimensions of timbre, is governed by different acoustic parameters compared to pitch, but the two can be confused perceptually when varied simultaneously. Here we combine human behavior and fMRI to provide evidence of a potential neural substrate to explain this important but poorly understood perceptual confusion. We identify orderly mappings of both pitch and brightness within auditory cortex and reveal two independent lines of evidence for cortical confusion between them. First, the preferred pitch of individual voxels decreases systematically as brightness increases, and vice versa, consistent with predictions based on perceptual confusion. Second, pitch and brightness mapping share a common high-low-high gradient across auditory cortex, implying a shared trajectory of cortical activation for changes in each dimension. The results provide a cortical substrate at both local and global scales for an established auditory perceptual phenomenon that is thought to reflect efficient coding of features ubiquitous in natural sound statistics.
Collapse
|
2
|
Zhang J, Liu Y, Wu P. An elastic piezoelectric nanomembrane with double noise reduction for high-quality bandpass acoustics. Nat Commun 2024; 15:8920. [PMID: 39414797 PMCID: PMC11484958 DOI: 10.1038/s41467-024-52787-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2024] [Accepted: 09/18/2024] [Indexed: 10/18/2024] Open
Abstract
Polymer piezoelectrics with high electromechanical energy conversion (HEEC) are very promising for flexible acoustoelectric devices. However, reducing thickness and improving ordered polarization and ferroelectricity while maintaining high mechanical strength pose enormous fabrication challenges for polymer piezoelectric membranes-additionally, noise management in the acoustoelectric conversion remains an open issue. Here, we present a hydro-levitation superspreading approach for fabricating polymer nanomembranes with ordered crystalline phases and sub-nanostructures on the water surface. The elastic piezoelectric nanomembrane (EPN) is only 335 nanometers thick and consists of a conductance-stable piezoelectric layer sandwiched between two elastic damping layers. Such an all-in-one EPN can reduce background noise with low autocorrelation in the environment, suppress spurious noise caused by poor circuit contact, and achieve bandpass filtering of acoustic signals at human voice frequencies. This nanomembrane holds promise in repairing the auditory system of patients with tympanic membrane perforation and in a wide range of other acoustoelectric conversion fields.
Collapse
Affiliation(s)
- Jialin Zhang
- State Key Laboratory for Modification of Chemical Fibers and Polymer Materials, College of Chemistry and Chemical Engineering, Donghua University, 2999 North Renmin Road, Shanghai, 201620, China
| | - Yanjun Liu
- State Key Laboratory for Modification of Chemical Fibers and Polymer Materials, College of Chemistry and Chemical Engineering, Donghua University, 2999 North Renmin Road, Shanghai, 201620, China.
| | - Peiyi Wu
- State Key Laboratory for Modification of Chemical Fibers and Polymer Materials, College of Chemistry and Chemical Engineering, Donghua University, 2999 North Renmin Road, Shanghai, 201620, China.
| |
Collapse
|
3
|
Bürgel M, Siedenburg K. Impact of interference on vocal and instrument recognition. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 156:922-938. [PMID: 39133041 DOI: 10.1121/10.0028152] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/15/2024] [Accepted: 07/12/2024] [Indexed: 08/13/2024]
Abstract
Voices arguably occupy a superior role in auditory processing. Specifically, studies have reported that singing voices are processed faster and more accurately and possess greater salience in musical scenes compared to instrumental sounds. However, the underlying acoustic features of this superiority and the generality of these effects remain unclear. This study investigates the impact of frequency micro-modulations (FMM) and the influence of interfering sounds on sound recognition. Thirty young participants, half with musical training, engage in three sound recognition experiments featuring short vocal and instrumental sounds in a go/no-go task. Accuracy and reaction times are measured for sounds from recorded samples and excerpts of popular music. Each sound is presented in separate versions with and without FMM, in isolation or accompanied by a piano. Recognition varies across sound categories, but no general vocal superiority emerges and no effects of FMM. When presented together with interfering sounds, all sounds exhibit degradation in recognition. However, whereas /a/ sounds stand out by showing a distinct robustness to interference (i.e., less degradation of recognition), /u/ sounds lack this robustness. Acoustical analysis implies that recognition differences can be explained by spectral similarities. Together, these results challenge the notion of general vocal superiority in auditory perception.
Collapse
Affiliation(s)
- Michel Bürgel
- Department of Medical Physics and Acoustics, University of Oldenburg, Oldenburg 26129, Germany
| | - Kai Siedenburg
- Department of Medical Physics and Acoustics, University of Oldenburg, Oldenburg 26129, Germany
- Signal Processing and Speech Communication Laboratory, Graz University of Technology, Graz 8010, Austria
| |
Collapse
|
4
|
McPherson MJ, McDermott JH. Relative pitch representations and invariance to timbre. Cognition 2023; 232:105327. [PMID: 36495710 PMCID: PMC10016107 DOI: 10.1016/j.cognition.2022.105327] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2022] [Revised: 09/13/2022] [Accepted: 11/10/2022] [Indexed: 12/12/2022]
Abstract
Information in speech and music is often conveyed through changes in fundamental frequency (f0), perceived by humans as "relative pitch". Relative pitch judgments are complicated by two facts. First, sounds can simultaneously vary in timbre due to filtering imposed by a vocal tract or instrument body. Second, relative pitch can be extracted in two ways: by measuring changes in constituent frequency components from one sound to another, or by estimating the f0 of each sound and comparing the estimates. We examined the effects of timbral differences on relative pitch judgments, and whether any invariance to timbre depends on whether judgments are based on constituent frequencies or their f0. Listeners performed up/down and interval discrimination tasks with pairs of spoken vowels, instrument notes, or synthetic tones, synthesized to be either harmonic or inharmonic. Inharmonic sounds lack a well-defined f0, such that relative pitch must be extracted from changes in individual frequencies. Pitch judgments were less accurate when vowels/instruments were different compared to when they were the same, and were biased by the associated timbre differences. However, this bias was similar for harmonic and inharmonic sounds, and was observed even in conditions where judgments of harmonic sounds were based on f0 representations. Relative pitch judgments are thus not invariant to timbre, even when timbral variation is naturalistic, and when such judgments are based on representations of f0.
Collapse
Affiliation(s)
- Malinda J McPherson
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA 02139, United States of America; Program in Speech and Hearing Biosciences and Technology, Harvard University, Boston, MA 02115, United States of America; McGovern Institute for Brain Research, MIT, Cambridge, MA 02139, United States of America.
| | - Josh H McDermott
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA 02139, United States of America; Program in Speech and Hearing Biosciences and Technology, Harvard University, Boston, MA 02115, United States of America; McGovern Institute for Brain Research, MIT, Cambridge, MA 02139, United States of America; Center for Brains Minds and Machines, MIT, Cambridge, MA 02139, United States of America
| |
Collapse
|
5
|
McAdams S, Thoret E, Wang G, Montrey M. Timbral cues for learning to generalize musical instrument identity across pitch register. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 153:797. [PMID: 36859162 DOI: 10.1121/10.0017100] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/30/2022] [Accepted: 01/12/2023] [Indexed: 06/18/2023]
Abstract
Timbre provides an important cue to identify musical instruments. Many timbral attributes covary with other parameters like pitch. This study explores listeners' ability to construct categories of instrumental sound sources from sounds that vary in pitch. Nonmusicians identified 11 instruments from the woodwind, brass, percussion, and plucked and bowed string families. In experiment 1, they were trained to identify instruments playing a pitch of C4, and in experiments 2 and 3, they were trained with a five-tone sequence (F#3-F#4), exposing them to the way timbre varies with pitch. Participants were required to reach a threshold of 75% correct identification in training. In the testing phase, successful listeners heard single tones (experiments 1 and 2) or three-tone sequences from (A3-D#4) (experiment 3) across each instrument's full pitch range to test their ability to generalize identification from the learned sound(s). Identification generalization over pitch varies a great deal across instruments. No significant differences were found between single-pitch and multi-pitch training or testing conditions. Identification rates can be predicted moderately well by spectrograms or modulation spectra. These results suggest that listeners use the most relevant acoustical invariance to identify musical instrument sounds, also using previous experience with the tested instruments.
Collapse
Affiliation(s)
- Stephen McAdams
- Schulich School of Music, McGill University, Montreal, Québec H3A 1E3, Canada
| | - Etienne Thoret
- Aix-Marseille University, Centre National de la Recherche Scientifique, Perception Representations Image Sound Music Laboratory, Unité Mixte de Recherche 7061, Laboratoire d'Informatique et Systèmes, Unité Mixte de Recherche 7020, 13009 Marseille, France
| | - Grace Wang
- Cognitive Science Program, McGill University, Montreal, Québec H3A 3R1, Canada
| | - Marcel Montrey
- Department of Psychology, McGill University, Montreal, Québec H3A 1G1, Canada
| |
Collapse
|
6
|
Siedenburg K, Graves J, Pressnitzer D. A unitary model of auditory frequency change perception. PLoS Comput Biol 2023; 19:e1010307. [PMID: 36634121 PMCID: PMC9876382 DOI: 10.1371/journal.pcbi.1010307] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2022] [Revised: 01/25/2023] [Accepted: 01/04/2023] [Indexed: 01/13/2023] Open
Abstract
Changes in the frequency content of sounds over time are arguably the most basic form of information about the behavior of sound-emitting objects. In perceptual studies, such changes have mostly been investigated separately, as aspects of either pitch or timbre. Here, we propose a unitary account of "up" and "down" subjective judgments of frequency change, based on a model combining auditory correlates of acoustic cues in a sound-specific and listener-specific manner. To do so, we introduce a generalized version of so-called Shepard tones, allowing symmetric manipulations of spectral information on a fine scale, usually associated to pitch (spectral fine structure, SFS), and on a coarse scale, usually associated timbre (spectral envelope, SE). In a series of behavioral experiments, listeners reported "up" or "down" shifts across pairs of generalized Shepard tones that differed in SFS, in SE, or in both. We observed the classic properties of Shepard tones for either SFS or SE shifts: subjective judgements followed the smallest log-frequency change direction, with cases of ambiguity and circularity. Interestingly, when both SFS and SE changes were applied concurrently (synergistically or antagonistically), we observed a trade-off between cues. Listeners were encouraged to report when they perceived "both" directions of change concurrently, but this rarely happened, suggesting a unitary percept. A computational model could accurately fit the behavioral data by combining different cues reflecting frequency changes after auditory filtering. The model revealed that cue weighting depended on the nature of the sound. When presented with harmonic sounds, listeners put more weight on SFS-related cues, whereas inharmonic sounds led to more weight on SE-related cues. Moreover, these stimulus-based factors were modulated by inter-individual differences, revealing variability across listeners in the detailed recipe for "up" and "down" judgments. We argue that frequency changes are tracked perceptually via the adaptive combination of a diverse set of cues, in a manner that is in fact similar to the derivation of other basic auditory dimensions such as spatial location.
Collapse
Affiliation(s)
- Kai Siedenburg
- Carl von Ossietzky University of Oldenburg, Dept. of Medical Physics and Acoustics, Oldenburg, Germany
- * E-mail:
| | - Jackson Graves
- Laboratoire des systèmes perceptifs, Dépt. d’études cognitives, École normale supérieure, PSL University, CNRS, Paris, France
| | - Daniel Pressnitzer
- Laboratoire des systèmes perceptifs, Dépt. d’études cognitives, École normale supérieure, PSL University, CNRS, Paris, France
| |
Collapse
|
7
|
Prabha A, Yadav J, Rani A, Singh V. Intelligent estimation of blood glucose level using wristband PPG signal and physiological parameters. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.103876] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
|
8
|
Abstract
Perception adapts to the properties of prior stimulation, as illustrated by phenomena such as visual color constancy or speech context effects. In the auditory domain, only little is known about adaptive processes when it comes to the attribute of auditory brightness. Here, we report an experiment that tests whether listeners adapt to spectral colorations imposed on naturalistic music and speech excerpts. Our results indicate consistent contrastive adaptation of auditory brightness judgments on a trial-by-trial basis. The pattern of results suggests that these effects tend to grow with an increase in the duration of the adaptor context but level off after around 8 trials of 2 s duration. A simple model of the response criterion yields a correlation of r = .97 with the measured data and corroborates the notion that brightness perception adapts on timescales that fall in the range of auditory short-term memory. Effects turn out to be similar for spectral filtering based on linear spectral filter slopes and filtering based on a measured transfer function from a commercially available hearing device. Overall, our findings demonstrate the adaptivity of auditory brightness perception under realistic acoustical conditions.
Collapse
Affiliation(s)
- Kai Siedenburg
- Department of Medical Physics and Acoustics, Carl von Ossietzky University of Oldenburg, Oldenburg, Germany.
| | - Feline Malin Barg
- Department of Medical Physics and Acoustics, Carl von Ossietzky University of Oldenburg, Oldenburg, Germany
| | - Henning Schepker
- Department of Medical Physics and Acoustics, Carl von Ossietzky University of Oldenburg, Oldenburg, Germany
- Starkey Hearing, Eden Prairie, MN, USA
| |
Collapse
|
9
|
Giordano N, Chatziioannou V. Status and future of modeling of musical instruments: Introduction to the JASA special issue. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 150:2294. [PMID: 34598593 DOI: 10.1121/10.0006439] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/30/2021] [Accepted: 09/07/2021] [Indexed: 06/13/2023]
Abstract
Over the last decades, physics-based modeling of musical instruments has seen increased attention. In 2020 and 2021, the Journal of the Acoustical Society of America accepted submissions for a special issue on the modeling of musical instruments. This article is intended as an introduction to the special issue. Our purpose is to discuss the role that modeling plays in the study of musical instruments, the kinds of things one hopes to learn from modeling studies, and how that work informs traditional experimental and theoretical studies of specific instruments. We also describe recent trends in modeling and make some observations about where we think the field is heading. Overall, our goal is to place the articles in the special issue into a context that helps the reader to better understand and appreciate the field.
Collapse
Affiliation(s)
- Nicholas Giordano
- Department of Physics, Auburn University, Auburn, Alabama 36849, USA
| | - Vasileios Chatziioannou
- Department of Music Acoustics, University of Music and Performing Arts Vienna, 1030 Vienna, Austria
| |
Collapse
|