1
|
Anikin A. Acoustic estimation of voice roughness. Atten Percept Psychophys 2025:10.3758/s13414-025-03060-3. [PMID: 40295423 DOI: 10.3758/s13414-025-03060-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 03/13/2025] [Indexed: 04/30/2025]
Abstract
Roughness is a perceptual characteristic of sound that was first applied to musical consonance and dissonance, but it is increasingly recognized as a central aspect of voice quality in human and animal communication. It may be particularly important for asserting social dominance or attracting attention in urgent signals such as screams. To ensure that the results of roughness research are valid and consistent across studies, we need standard methodology for measuring it. I review the literature on roughness estimation, from classic psychoacoustics to more recent approaches, and present two collections of 602 human vocal samples whose roughness was rated by 162 listeners in perceptual experiments. Two algorithms for estimating roughness acoustically from modulation spectra are then presented and optimized to match the human ratings. One uses a bank of gammatone or Butterworth filters to obtain an auditory spectrogram, and a faster algorithm begins with a conventional spectrogram obtained with Short-Time Fourier transform; both explain ~ 50% of variance in average human ratings per stimulus. The range of modulation frequencies most relevant to roughness perception is [50, 200] Hz; this range can be selected with simple cutoff points or with a lognormal weighting function. Modulation and roughness spectrograms are proposed as visual aids for studying the dynamics of roughness in longer recordings. The described algorithms are implemented in the function modulationSpectrum() from the open-source R library soundgen. The audio recordings and their ratings are freely available from https://osf.io/gvcpx/ and can be used for benchmarking other algorithms.
Collapse
Affiliation(s)
- Andrey Anikin
- Division of Cognitive Science, Department of Philosophy, Lund University, Box 192, SE- 221 00, Lund, Sweden.
| |
Collapse
|
2
|
Gélat P, Van't Wout E, Haqshenas R, Melbourne A, David AL, Mufti N, Henriques J, Thibaut de Maisières A, Jauniaux E. Evaluation of fetal exposure to environmental noise using a computer-generated model. Nat Commun 2025; 16:3916. [PMID: 40280911 PMCID: PMC12032418 DOI: 10.1038/s41467-025-58983-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2024] [Accepted: 04/08/2025] [Indexed: 04/29/2025] Open
Abstract
Acoustic noise can have profound effects on wellbeing, impacting the health of pregnant women and their fetus. Mounting evidence suggests neural memory traces are formed by auditory learning in utero. A better understanding of the fetal auditory environment is therefore critical to avoid exposure to damaging noise levels. Using anatomical data from MRI scans of pregnant patients ( N = 4 ) from 24 weeks of gestation, we develop a computational model to quantify fetal exposure to acoustic field. We obtain acoustic transfer characteristics across the human audio range and pressure maps in transverse planes passing through the uterus at 5 kHz, 10 kHz and 20 kHz, showcasing multiple scattering and modal patterns. Our calculations show that the sound transmitted in utero is attenuated by as little as 6 dB below 1 kHz, confirming results from animal studies that the maternal abdomen and pelvis do not shelter the fetus from external noise.
Collapse
Affiliation(s)
- Pierre Gélat
- Department of Surgical Biotechnology, Division of Surgery and Interventional Science, University College London, London, UK.
| | - Elwin Van't Wout
- Institute for Mathematical and Computational Engineering, Pontificia Universidad Católica de Chile, Santiago, Chile
| | - Reza Haqshenas
- Department of Mechanical Engineering, University College London, London, UK
| | - Andrew Melbourne
- School of Biomedical Engineering & Imaging Sciences, Faculty of Life Sciences & Medicine, King's College London, London, UK
- Elizabeth Garrett Anderson Institute for Women's Health, University College London, London, UK
| | - Anna L David
- Elizabeth Garrett Anderson Institute for Women's Health, University College London, London, UK
| | - Nada Mufti
- Elizabeth Garrett Anderson Institute for Women's Health, University College London, London, UK
| | - Julian Henriques
- Department of Media, Communications and Cultural Studies, Goldsmiths University of London, London, UK
| | | | - Eric Jauniaux
- Elizabeth Garrett Anderson Institute for Women's Health, University College London, London, UK
| |
Collapse
|
3
|
Bruder C, Larrouy-Maestri P. CoVox: A dataset of contrasting vocalizations. Behav Res Methods 2025; 57:142. [PMID: 40216652 PMCID: PMC11991967 DOI: 10.3758/s13428-025-02664-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 03/17/2025] [Indexed: 04/14/2025]
Abstract
The human voice is remarkably versatile and can vary greatly in sound depending on how it is used. An increasing number of studies have addressed the differences and similarities between the singing and the speaking voice. However, finding adequate stimuli material that is at the same time controlled and ecologically valid is challenging, and most datasets lack variability in terms of vocal styles performed by the same voice. Here, we describe a curated stimulus set of vocalizations where 22 female singers performed the same melody excerpts in three contrasting singing styles (as a lullaby, as a pop song, and as an opera aria) and spoke the text aloud in two speaking styles (as if speaking to an adult or to an infant). All productions were made with the songs' original lyrics, in Brazilian Portuguese, and with a/lu/sound. This ecologically valid dataset of 1320 vocalizations was validated through a forced-choice lab experiment (N = 25 for each stimulus) where lay listeners could recognize the intended vocalization style with high accuracy (proportion of correct recognition superior to 69% for all styles). We also provide acoustic characterization of the stimuli, depicting clear and contrasting acoustic profiles depending on the style of vocalization. All recordings are made freely available under a Creative Commons license and can be downloaded at https://osf.io/cgexn/ .
Collapse
Affiliation(s)
- Camila Bruder
- Max Planck Institute for Empirical Aesthetics, Grüneburgweg 14, 60322, Frankfurt Am Main, Germany.
| | - Pauline Larrouy-Maestri
- Max Planck Institute for Empirical Aesthetics, Grüneburgweg 14, 60322, Frankfurt Am Main, Germany.
- Center for Language, Music, and Emotion (CLaME), New York, NY, USA.
| |
Collapse
|
4
|
Fitch WT, Anikin A, Pisanski K, Valente D, Reby D. Formant analysis of vertebrate vocalizations: achievements, pitfalls, and promises. BMC Biol 2025; 23:92. [PMID: 40189499 PMCID: PMC11974057 DOI: 10.1186/s12915-025-02188-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2024] [Accepted: 03/06/2025] [Indexed: 04/09/2025] Open
Abstract
When applied to vertebrate vocalizations, source-filter theory, initially developed for human speech, has revolutionized our understanding of animal communication, resulting in major insights into the form and function of animal sounds. However, animal calls and human nonverbal vocalizations can differ qualitatively from human speech, often having more chaotic and higher-frequency sources, making formant measurement challenging. We review the considerable achievements of the "formant revolution" in animal vocal communication research, then highlight several important methodological problems in formant analysis. We offer concrete recommendations for effectively applying source-filter theory to non-speech vocalizations and discuss promising avenues for future research in this area.Brief Formants (vocal tract resonances) play key roles in animal communication, offering researchers exciting promise but also potential pitfalls.
Collapse
Affiliation(s)
- W Tecumseh Fitch
- Department of Behavioral and Cognitive Biology, University of Vienna, Djerassiplatz 1, Vienna, 1030, Austria.
| | - Andrey Anikin
- Division of Cognitive Science, Lund University, Helgonavägen 3, Lund, 223 62, Sweden
- ENES Bioacoustics Research Lab/Lyon Neuroscience Research Centre (CRNL), University of Saint-Etienne, CNRS UMR5292, INSERM UMR_S 1028, 21 rue du Dr. Paul Michelon, Saint-Étienne, 42100, France
| | - Katarzyna Pisanski
- ENES Bioacoustics Research Lab/Lyon Neuroscience Research Centre (CRNL), University of Saint-Etienne, CNRS UMR5292, INSERM UMR_S 1028, 21 rue du Dr. Paul Michelon, Saint-Étienne, 42100, France
- CNRS French National Centre for Scientific Research, DDL Dynamics of Language Lab, University of Lyon 2, Lyon, 69007, France
- Institute of Psychology, University of Wrocław, Poland, Jana Władysława Dawida 1, Wrocław, 50-529, Poland
| | - Daria Valente
- ENES Bioacoustics Research Lab/Lyon Neuroscience Research Centre (CRNL), University of Saint-Etienne, CNRS UMR5292, INSERM UMR_S 1028, 21 rue du Dr. Paul Michelon, Saint-Étienne, 42100, France
- Department of Life Sciences and Systems Biology, University of Torino, Via Accademia Albertina 13, Torino, 10123, Italy
| | - David Reby
- ENES Bioacoustics Research Lab/Lyon Neuroscience Research Centre (CRNL), University of Saint-Etienne, CNRS UMR5292, INSERM UMR_S 1028, 21 rue du Dr. Paul Michelon, Saint-Étienne, 42100, France.
- Institut Universitaire de France, 1 Rue Descartes, Paris, France.
| |
Collapse
|
5
|
del Olmo M, Schmal C, Herzel H. Exploring nonlinear phenomena in animal vocalizations through oscillator theory. Philos Trans R Soc Lond B Biol Sci 2025; 380:20240015. [PMID: 40176524 PMCID: PMC11966158 DOI: 10.1098/rstb.2024.0015] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2024] [Revised: 09/30/2024] [Accepted: 10/29/2024] [Indexed: 04/04/2025] Open
Abstract
Animal vocalizations comprise a rich array of complex sounds that exhibit nonlinear phenomena (NLP), which have fascinated researchers for decades. From the melodic songs of birds to the clicks and whistles of dolphins, many species have been found to produce nonlinear vocalizations, offering a valuable perspective on the mechanisms underlying sound production and potential adaptive functions. By leveraging on the principles of oscillator theory and nonlinear dynamics, animal vocalizations, which are based on coupled oscillators, can be described and conveniently classified. We review the basic ingredients for self-sustained oscillations and how different NLP can emerge. We discuss important terms in the context of oscillator theory: attractor types, phase space, bifurcations and Arnold tongue diagrams. Through a comparative analysis of observed NLP and bifurcation diagrams, our study reviews how the tools of nonlinear dynamics can provide insights into the intricate complexity of animal vocalizations, as well as into the evolutionary pressures and adaptive strategies that have shaped the diverse communication systems of the animal kingdom.This article is part of the theme issue, 'Nonlinear phenomena in vertebrate vocalizations: mechanisms and communicative functions'.
Collapse
Affiliation(s)
- Marta del Olmo
- Humboldt-Universität zu Berlin Institute for Theoretical Biology, Berlin, Germany
| | - Christoph Schmal
- Humboldt-Universität zu Berlin Institute for Theoretical Biology, Berlin, Germany
| | - Hanspeter Herzel
- Humboldt-Universität zu Berlin Institute for Theoretical Biology, Berlin, Germany
| |
Collapse
|
6
|
Anikin A, Herbst CT. How to analyse and manipulate nonlinear phenomena in voice recordings. Philos Trans R Soc Lond B Biol Sci 2025; 380:20240003. [PMID: 40176526 PMCID: PMC11966163 DOI: 10.1098/rstb.2024.0003] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2024] [Revised: 07/22/2024] [Accepted: 10/01/2024] [Indexed: 04/04/2025] Open
Abstract
We address two research applications in this methodological review: starting from an audio recording, the goal may be to characterize nonlinear phenomena (NLP) at the level of voice production or to test their perceptual effects on listeners. A crucial prerequisite for this work is the ability to detect NLP in acoustic signals, which can then be correlated with biologically relevant information about the caller and with listeners' reaction. NLP are often annotated manually, but this is labour-intensive and not very reliable, although we describe potentially helpful advanced visualization aids such as reassigned spectrograms and phasegrams. Objective acoustic features can also be useful, including general descriptives (harmonics-to-noise ratio, cepstral peak prominence, vocal roughness), statistics derived from nonlinear dynamics (correlation dimension) and NLP-specific measures (depth of modulation and subharmonics). On the perception side, playback studies can greatly benefit from tools for directly manipulating NLP in recordings. Adding frequency jumps, amplitude modulation and subharmonics is relatively straightforward. Creating biphonation, imitating chaos or removing NLP from a recording are more challenging, but feasible with parametric voice synthesis. We describe the most promising algorithms for analysing and manipulating NLP and provide detailed examples with audio files and R code in supplementary material.This article is part of the theme issue 'Nonlinear phenomena in vertebrate vocalizations: mechanisms and communicative functions'.
Collapse
Affiliation(s)
- Andrey Anikin
- Division of Cognitive Science, Lund University, Lund, Sweden
- ENES Bioacoustics Research Laboratory, Université Jean Monnet Saint-Étienne, Saint-Étienne, France
| | - Christian T. Herbst
- University of Vienna, Vienna, Austria
- Department of Communication Sciences and Disorders, College of Liberal Arts and Sciences, University of Iowa, Iowa City, Iowa, USA
| |
Collapse
|
7
|
Fitch WT. Applying nonlinear dynamics to the voice: a historical perspective. Philos Trans R Soc Lond B Biol Sci 2025; 380:20240024. [PMID: 40176512 PMCID: PMC11966167 DOI: 10.1098/rstb.2024.0024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2024] [Revised: 08/29/2024] [Accepted: 09/04/2024] [Indexed: 04/04/2025] Open
Abstract
The recognition that nonlinear phenomena, including subharmonics, bifurcations and deterministic chaos, are present in human and animal vocalizations is a relatively recent one. I give a brief history of this revolution in our understanding of the voice, based on interviews with some of the key players and personal experience. Most of the key concepts and mathematical principles of nonlinear dynamics were already well worked out in the early 1980s. In the early 1990s, physicist Hanspeter Herzel and colleagues in Berlin recognized that these principles are applicable to the human voice, initially to baby cries. The physics and physiology underlying many of these nonlinear phenomena had remained mysterious up until then. This insight was later generalized to animal vocalizations. Nonlinear phenomena play a relatively peripheral role in most human vocal communication but are a common feature of many animal vocalizations. The broad recognition of the existence of nonlinear vocalizations, and the quantitative study of their production and perception, has now fuelled important and exciting advances in our understanding of animal communication. I concentrate on how the core concepts came into focus, and on their initial application to an ever-wider circle of call types and species, and end with a brief prospectus for the future.This article is part of the theme issue 'Nonlinear phenomena in vertebrate vocalizations: mechanisms and communicative functions'.
Collapse
Affiliation(s)
- W. Tecumseh Fitch
- Department of Behavioral and Cognitive Biology, University of Vienna, Vienna, Austria
| |
Collapse
|
8
|
Anikin A, Reby D, Pisanski K. Nonlinear vocal phenomena and speech intelligibility. Philos Trans R Soc Lond B Biol Sci 2025; 380:20240254. [PMID: 40176514 PMCID: PMC11966171 DOI: 10.1098/rstb.2024.0254] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2024] [Revised: 08/27/2024] [Accepted: 09/04/2024] [Indexed: 04/04/2025] Open
Abstract
At some point in our evolutionary history, humans lost vocal membranes and air sacs, representing an unexpected simplification of the vocal apparatus relative to other great apes. One hypothesis is that these simplifications represent anatomical adaptations for speech because a simpler larynx provides a suitably stable and tonal vocal source with fewer nonlinear vocal phenomena (NLP). The key assumption that NLP reduce speech intelligibility is indirectly supported by studies of dysphonia, but it has not been experimentally tested. Here, we manipulate NLP in vocal stimuli ranging from single vowels to sentences, showing that the vocal source needs to be stable, but not necessarily tonal, for speech to be readily understood. When the task is to discriminate synthesized monophthong and diphthong vowels, continuous NLP (subharmonics, amplitude modulation and even deterministic chaos) actually improve vowel perception in high-pitched voices, likely because the resulting dense spectrum reveals formant transitions. Rough-sounding voices also remain highly intelligible when continuous NLP are added to recorded words and sentences. In contrast, voicing interruptions and pitch jumps dramatically reduce speech intelligibility, likely by interfering with voicing contrasts and normal intonation. We argue that NLP were not eliminated from the human vocal repertoire as we evolved for speech, but only brought under better control.This article is part of the theme issue 'Nonlinear phenomena in vertebrate vocalizations: mechanisms and communicative functions'.
Collapse
Affiliation(s)
- Andrey Anikin
- Division of Cognitive Science, Department of Philosophy, Lund University, Box 192, SE-221 00, Lund, Sweden
| | - David Reby
- ENES Bioacoustics Research Laboratory, CRNL, CNRS, Inserm, University of Saint-Etienne, Saint-Etienne42100, France
- Institute of Psychology, Institut universitaire de France, Paris, France
| | - Katarzyna Pisanski
- ENES Bioacoustics Research Laboratory, CRNL, CNRS, Inserm, University of Saint-Etienne, Saint-Etienne42100, France
- Institute of Psychology, University of Wrocław, Wrocław03-815, Poland
| |
Collapse
|
9
|
Corvin S, Massenet M, Hardy A, Patural H, Peyron R, Fauchon C, Mathevon N. Nonlinear acoustic phenomena affect the perception of pain in human baby cries. Philos Trans R Soc Lond B Biol Sci 2025; 380:20240023. [PMID: 40176515 PMCID: PMC11966150 DOI: 10.1098/rstb.2024.0023] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2024] [Revised: 10/22/2024] [Accepted: 11/28/2024] [Indexed: 04/04/2025] Open
Abstract
What makes the painful cries of human babies so difficult to ignore? Vocal traits known as 'nonlinear phenomena' are prime candidates. These acoustic irregularities are common in babies' cries and are typically associated with high levels of distress or pain. Despite the vital importance of cries for a baby's survival, how these nonlinear phenomena drive pain perception in adult listeners has not previously been systematically investigated. Here, by combining acoustic analyses of cries recorded in different contexts with playback experiments using natural and synthetic cries, we show that baby cries expressing acute pain are characterized by a pronounced presence of different nonlinear phenomena, and that these nonlinear phenomena drive pain evaluation by adult listeners. While adult listeners rated all cries presenting any of these nonlinear phenomena as expressing more pain, they were particularly sensitive to the presence of chaos. Our results thus show that nonlinear phenomena, especially chaos, encode pain information in baby cries and may be critically helpful for the development of vocal-based tools for monitoring babies' needs in the context of paediatric care.This article is part of the theme issue 'Nonlinear phenomena in vertebrate vocalizations: mechanisms and communicative functions'.
Collapse
Affiliation(s)
- Siloé Corvin
- ENES Bioacoustics Research Lab, CRNL, CNRS, Inserm, University of Saint-Etienne, Saint-Etienne42100, France
- NEUROPAIN Team, CRNL, CNRS, Inserm, University of Saint-Etienne, Saint-Etienne42270, France
| | - Mathilde Massenet
- ENES Bioacoustics Research Lab, CRNL, CNRS, Inserm, University of Saint-Etienne, Saint-Etienne42100, France
| | - Angélique Hardy
- NEUROPAIN Team, CRNL, CNRS, Inserm, University of Saint-Etienne, Saint-Etienne42270, France
- Neuro-Dol, Inserm, University of Clermont Auvergne, University Hospital of Clermont-Ferrand, Clermont-Ferrand63100, France
| | - Hugues Patural
- Neonatal and Pediatric Intensive Care Unit SAINBIOSE Laboratory, Inserm, University Hospital of Saint-Etienne, University of Saint-Etienne, Saint-Etienne42270, France
| | - Roland Peyron
- NEUROPAIN Team, CRNL, CNRS, Inserm, University of Saint-Etienne, Saint-Etienne42270, France
| | - Camille Fauchon
- NEUROPAIN Team, CRNL, CNRS, Inserm, University of Saint-Etienne, Saint-Etienne42270, France
- Neuro-Dol, Inserm, University of Clermont Auvergne, University Hospital of Clermont-Ferrand, Clermont-Ferrand63100, France
| | - Nicolas Mathevon
- ENES Bioacoustics Research Lab, CRNL, CNRS, Inserm, University of Saint-Etienne, Saint-Etienne42100, France
- Institut Universitaire de France, Paris75005, France
- Ecole Pratique des Hautes Etudes, CHArt lab, EPHE - PSL University, Paris, France
| |
Collapse
|
10
|
Herbst CT, Tokuda IT, Nishimura T, Ternström S, Ossio V, Levy M, Fitch WT, Dunn JC. 'Monkey yodels'-frequency jumps in New World monkey vocalizations greatly surpass human vocal register transitions. Philos Trans R Soc Lond B Biol Sci 2025; 380:20240005. [PMID: 40176522 PMCID: PMC11966155 DOI: 10.1098/rstb.2024.0005] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2024] [Revised: 11/04/2024] [Accepted: 11/04/2024] [Indexed: 04/04/2025] Open
Abstract
We investigated the causal basis of abrupt frequency jumps in a unique database of New World monkey vocalizations. We used a combination of acoustic and electroglottographic recordings in vivo, excised larynx investigations of vocal fold dynamics, and computational modelling. We particularly attended to the contribution of the vocal membranes: thin upward extensions of the vocal folds found in most primates but absent in humans. In three of the six investigated species, we observed two distinct modes of vocal fold vibration. The first, involving vocal fold vibration alone, produced low-frequency oscillations, and is analogous to that underlying human phonation. The second, incorporating the vocal membranes, resulted in much higher-frequency oscillation. Abrupt fundamental frequency shifts were observed in all three datasets. While these data are reminiscent of the rapid transitions in frequency observed in certain human singing styles (e.g. yodelling), the frequency jumps are considerably larger in the nonhuman primates studied. Our data suggest that peripheral modifications of vocal anatomy provide an important source of variability and complexity in the vocal repertoires of nonhuman primates. We further propose that the call repertoire is crucially related to a species' ability to vocalize with different laryngeal mechanisms, analogous to human vocal registers. This article is part of the theme issue 'Nonlinear phenomena in vertebrate vocalizations: mechanisms and communicative functions'.
Collapse
Affiliation(s)
- Christian T. Herbst
- Department of Behavioral and Cognitive Biology, University of Vienna, Vienna, Austria
- Center for the Evolutionary Origins of Human Behavior, Kyoto University, Inuyama, Japan
| | - Isao T. Tokuda
- Department of Mechanical Engineering, Ritsumeikan University, Kusatsu, Japan
| | - Takeshi Nishimura
- Graduate School of Human Sciences, The University of Osaka, Suita, Japan
| | - Sten Ternström
- Department of Speech, Music and Hearing, School of Electrical Engineering and Computer Science, KTH Royal Institute of Technology, Stockholm, Sweden
| | | | | | - W. Tecumseh Fitch
- Department of Behavioral and Cognitive Biology, University of Vienna, Vienna, Austria
| | - Jacob C. Dunn
- Department of Behavioral and Cognitive Biology, University of Vienna, Vienna, Austria
- Behavioural Ecology Research Group, Anglia Ruskin University, CambridgeCB1 1PT, UK
- ENES Bioacoustics Research Laboratory, University of Saint-Etienne, St-Etienne42023, France
| |
Collapse
|
11
|
Gélat P, van 't Wout E, Haqhenas R, Melbourne A, David AL, Mufti N, Henriques J, de Maisieres AT, Jauniaux E. Evaluation of fetal exposure to environmental noise using a computer-generated model. RESEARCH SQUARE 2024:rs.3.rs-5397645. [PMID: 39649159 PMCID: PMC11623772 DOI: 10.21203/rs.3.rs-5397645/v1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/10/2024]
Abstract
Acoustic noise can have profound effects on wellbeing, impacting the health of the pregnant mother and the development of the fetus. Mounting evidence suggests neural memory traces are formed by auditory learning in utero. A better understanding of the fetal auditory environment is therefore critical to avoid exposure to damaging noise levels. Using anatomical data from MRI scans (N = 3), we used a computational model to quantify the acoustic field inside the pregnant maternal abdomen. We obtained acoustic transfer characteristics across the human audio range and pressure maps in transverse planes passing through the uterus at 5 kHz, 10 kHz and 20 kHz, showcasing multiple scattering and modal patterns. Our calculations suggest that for all datasets, the sound transmitted in utero is attenuated by as little as 6 dB below 1 kHz, confirming results from animal studies that the maternal abdomen and pelvis do not shelter the fetus from external noise.
Collapse
Affiliation(s)
- Pierre Gélat
- Division of Surgery and Interventional Science, Faculty of Medical Sciences, University College London
| | - Elwin van 't Wout
- Institute for Mathematical and Computational Engineering, Pontificia Universidad Católica de Chile, Santiago, Chile
| | - Reza Haqhenas
- Department of Mechanical Engineering, Faculty of Engineering Sciences, University College London
| | - Andrew Melbourne
- Faculty of Life Sciences & Medicine, School of Biomedical Engineering & Imaging Sciences, Department of Surgical & Interventional Engineering, King's College London
- EGA Institute for Women's Health, Faculty of Population Health Sciences, University College London, London, UK
| | - Anna L David
- EGA Institute for Women's Health, Faculty of Population Health Sciences, University College London, London, UK
| | - Nada Mufti
- EGA Institute for Women's Health, Faculty of Population Health Sciences, University College London, London, UK
| | - Julian Henriques
- Department of Media, Communications and Cultural Studies, Goldsmiths, University of London, London, UK
| | | | - Eric Jauniaux
- EGA Institute for Women's Health, Faculty of Population Health Sciences, University College London, London, UK
| |
Collapse
|
12
|
Ponsonnet M, Coupé C, Pellegrino F, Garcia Arasco A, Pisanski K. Vowel signatures in emotional interjections and nonlinguistic vocalizations expressing pain, disgust, and joy across languagesa). THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 156:3118-3139. [PMID: 39531311 DOI: 10.1121/10.0032454] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/03/2024] [Accepted: 10/01/2024] [Indexed: 11/16/2024]
Abstract
In this comparative cross-linguistic study we test whether expressive interjections (words like ouch or yay) share similar vowel signatures across the world's languages, and whether these can be traced back to nonlinguistic vocalizations (like screams and cries) expressing the same emotions of pain, disgust, and joy. We analyze vowels in interjections from dictionaries of 131 languages (over 600 tokens) and compare these with nearly 500 vowels based on formant frequency measures from voice recordings of volitional nonlinguistic vocalizations. We show that across the globe, pain interjections feature a-like vowels and wide falling diphthongs ("ai" as in Ayyy! "aw" as in Ouch!), whereas disgust and joy interjections do not show robust vowel regularities that extend geographically. In nonlinguistic vocalizations, all emotions yield distinct vowel signatures: pain prompts open vowels such as [a], disgust schwa-like central vowels, and joy front vowels such as [i]. Our results show that pain is the only affective experience tested with a clear, robust vowel signature that is preserved between nonlinguistic vocalizations and interjections across languages. These results offer empirical evidence for iconicity in some expressive interjections. We consider potential mechanisms and origins, from evolutionary pressures and sound symbolism to colexification, proposing testable hypotheses for future research.
Collapse
Affiliation(s)
- Maïa Ponsonnet
- Dynamique Du Langage, CNRS et Université Lumière Lyon 2, Lyon, France
- School of Social Sciences, The University of Western Australia, Perth, Australia
| | - Christophe Coupé
- Department of Linguistics, The University of Hong Kong, Hong Kong SAR, China
| | | | | | - Katarzyna Pisanski
- Dynamique Du Langage, CNRS et Université Lumière Lyon 2, Lyon, France
- ENES Bioacoustics Research Laboratory, University Jean Monnet of Saint-Etienne, CRNL, CNRS, Saint-Etienne, France
- Institute of Psychology, University of Wrocław, Wrocław, Poland
| |
Collapse
|
13
|
Zhang Z. Principal dimensions of voice production and their role in vocal expression. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 156:278-283. [PMID: 38980102 PMCID: PMC11236430 DOI: 10.1121/10.0027913] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/08/2024] [Revised: 06/20/2024] [Accepted: 06/24/2024] [Indexed: 07/10/2024]
Abstract
How we produce and perceive voice is constrained by laryngeal physiology and biomechanics. Such constraints may present themselves as principal dimensions in the voice outcome space that are shared among speakers. This study attempts to identify such principal dimensions in the voice outcome space and the underlying laryngeal control mechanisms in a three-dimensional computational model of voice production. A large-scale voice simulation was performed with parametric variations in vocal fold geometry and stiffness, glottal gap, vocal tract shape, and subglottal pressure. Principal component analysis was applied to data combining both the physiological control parameters and voice outcome measures. The results showed three dominant dimensions accounting for at least 50% of the total variance. The first two dimensions describe respiratory-laryngeal coordination in controlling the energy balance between low- and high-frequency harmonics in the produced voice, and the third dimension describes control of the fundamental frequency. The dominance of these three dimensions suggests that voice changes along these principal dimensions are likely to be more consistently produced and perceived by most speakers than other voice changes, and thus are more likely to have emerged during evolution and be used to convey important personal information, such as emotion and larynx size.
Collapse
Affiliation(s)
- Zhaoyan Zhang
- Department of Head and Neck Surgery, University of California, Los Angeles, 31-24 Rehab Center, 1000 Veteran Avenue, Los Angeles, California 90095-1794, USA
| |
Collapse
|
14
|
Albouy P, Mehr SA, Hoyer RS, Ginzburg J, Du Y, Zatorre RJ. Spectro-temporal acoustical markers differentiate speech from song across cultures. Nat Commun 2024; 15:4835. [PMID: 38844457 PMCID: PMC11156671 DOI: 10.1038/s41467-024-49040-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2023] [Accepted: 05/21/2024] [Indexed: 06/09/2024] Open
Abstract
Humans produce two forms of cognitively complex vocalizations: speech and song. It is debated whether these differ based primarily on culturally specific, learned features, or if acoustical features can reliably distinguish them. We study the spectro-temporal modulation patterns of vocalizations produced by 369 people living in 21 urban, rural, and small-scale societies across six continents. Specific ranges of spectral and temporal modulations, overlapping within categories and across societies, significantly differentiate speech from song. Machine-learning classification shows that this effect is cross-culturally robust, vocalizations being reliably classified solely from their spectro-temporal features across all 21 societies. Listeners unfamiliar with the cultures classify these vocalizations using similar spectro-temporal cues as the machine learning algorithm. Finally, spectro-temporal features are better able to discriminate song from speech than a broad range of other acoustical variables, suggesting that spectro-temporal modulation-a key feature of auditory neuronal tuning-accounts for a fundamental difference between these categories.
Collapse
Affiliation(s)
- Philippe Albouy
- CERVO Brain Research Centre, School of Psychology, Laval University, Québec City, QC, Canada.
- International Laboratory for Brain, Music and Sound Research (BRAMS), Montreal, QC, Canada.
- Centre for Research in Brain, Language and Music and Centre for Interdisciplinary Research in Music, Media, and Technology, Montréal, QC, Canada.
| | - Samuel A Mehr
- International Laboratory for Brain, Music and Sound Research (BRAMS), Montreal, QC, Canada
- School of Psychology, University of Auckland, Auckland, 1010, New Zealand
- Child Study Center, Yale University, New Haven, CT, 06511, USA
| | - Roxane S Hoyer
- CERVO Brain Research Centre, School of Psychology, Laval University, Québec City, QC, Canada
| | - Jérémie Ginzburg
- CERVO Brain Research Centre, School of Psychology, Laval University, Québec City, QC, Canada
- Lyon Neuroscience Research Center, CNRS, UMR5292, INSERM, U1028 - Université Claude Bernard Lyon 1, F-69000, Lyon, France
- Cognitive Neuroscience Unit, Montreal Neurological Institute, McGill University, Montreal, QC, Canada
| | - Yi Du
- Institute of Psychology, Chinese Academy of Sciences, Beijing, China
| | - Robert J Zatorre
- International Laboratory for Brain, Music and Sound Research (BRAMS), Montreal, QC, Canada.
- Centre for Research in Brain, Language and Music and Centre for Interdisciplinary Research in Music, Media, and Technology, Montréal, QC, Canada.
- Cognitive Neuroscience Unit, Montreal Neurological Institute, McGill University, Montreal, QC, Canada.
| |
Collapse
|
15
|
Ozaki Y, Tierney A, Pfordresher PQ, McBride JM, Benetos E, Proutskova P, Chiba G, Liu F, Jacoby N, Purdy SC, Opondo P, Fitch WT, Hegde S, Rocamora M, Thorne R, Nweke F, Sadaphal DP, Sadaphal PM, Hadavi S, Fujii S, Choo S, Naruse M, Ehara U, Sy L, Parselelo ML, Anglada-Tort M, Hansen NC, Haiduk F, Færøvik U, Magalhães V, Krzyżanowski W, Shcherbakova O, Hereld D, Barbosa BS, Varella MAC, van Tongeren M, Dessiatnitchenko P, Zar SZ, El Kahla I, Muslu O, Troy J, Lomsadze T, Kurdova D, Tsope C, Fredriksson D, Arabadjiev A, Sarbah JP, Arhine A, Meachair TÓ, Silva-Zurita J, Soto-Silva I, Millalonco NEM, Ambrazevičius R, Loui P, Ravignani A, Jadoul Y, Larrouy-Maestri P, Bruder C, Teyxokawa TP, Kuikuro U, Natsitsabui R, Sagarzazu NB, Raviv L, Zeng M, Varnosfaderani SD, Gómez-Cañón JS, Kolff K, der Nederlanden CVB, Chhatwal M, David RM, Setiawan IPG, Lekakul G, Borsan VN, Nguqu N, Savage PE. Globally, songs and instrumental melodies are slower and higher and use more stable pitches than speech: A Registered Report. SCIENCE ADVANCES 2024; 10:eadm9797. [PMID: 38748798 PMCID: PMC11095461 DOI: 10.1126/sciadv.adm9797] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/15/2023] [Accepted: 04/19/2024] [Indexed: 05/19/2024]
Abstract
Both music and language are found in all known human societies, yet no studies have compared similarities and differences between song, speech, and instrumental music on a global scale. In this Registered Report, we analyzed two global datasets: (i) 300 annotated audio recordings representing matched sets of traditional songs, recited lyrics, conversational speech, and instrumental melodies from our 75 coauthors speaking 55 languages; and (ii) 418 previously published adult-directed song and speech recordings from 209 individuals speaking 16 languages. Of our six preregistered predictions, five were strongly supported: Relative to speech, songs use (i) higher pitch, (ii) slower temporal rate, and (iii) more stable pitches, while both songs and speech used similar (iv) pitch interval size and (v) timbral brightness. Exploratory analyses suggest that features vary along a "musi-linguistic" continuum when including instrumental melodies and recited lyrics. Our study provides strong empirical evidence of cross-cultural regularities in music and speech.
Collapse
Affiliation(s)
- Yuto Ozaki
- Graduate School of Media and Governance, Keio University, Fujisawa, Kanagawa, Japan
| | - Adam Tierney
- Department of Psychological Sciences, Birkbeck, University of London, London, UK
| | - Peter Q. Pfordresher
- Department of Psychology, University at Buffalo, State University of New York, Buffalo, NY, USA
| | - John M. McBride
- Center for Algorithmic and Robotized Synthesis, Institute for Basic Science, Ulsan, South Korea
| | - Emmanouil Benetos
- School of Electronic Engineering and Computer Science, Queen Mary University of London, London, UK
| | - Polina Proutskova
- School of Electronic Engineering and Computer Science, Queen Mary University of London, London, UK
| | - Gakuto Chiba
- Graduate School of Media and Governance, Keio University, Fujisawa, Kanagawa, Japan
| | - Fang Liu
- School of Psychology and Clinical Language Sciences, University of Reading, Reading, UK
| | - Nori Jacoby
- Computational Auditory Perception Group, Max-Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany
| | - Suzanne C. Purdy
- School of Psychology, University of Auckland, Auckland, New Zealand
- Centre for Brain Research and Eisdell Moore Centre for Hearing and Balance Research, University of Auckland, Auckland, New Zealand
| | - Patricia Opondo
- School of Arts, Music Discipline, University of KwaZulu Natal, Durban, South Africa
| | - W. Tecumseh Fitch
- Department of Behavioral and Cognitive Biology, University of Vienna, Vienna, Austria
| | - Shantala Hegde
- Music Cognition Lab, Department of Clinical Psychology, National Institute of Mental Health and Neuro Sciences, Bangalore, Karnataka, India
| | - Martín Rocamora
- Universidad de la República, Montevideo, Uruguay
- Music Technology Group, Universitat Pompeu Fabra, Barcelona, Spain
| | - Rob Thorne
- School of Music, Victoria University of Wellington, Wellington, New Zealand
| | - Florence Nweke
- Department of Creative Arts, University of Lagos, Lagos, Nigeria
- Department of Music, Mountain Top University, Ogun, Nigeria
| | - Dhwani P. Sadaphal
- Department of Behavioral and Cognitive Biology, University of Vienna, Vienna, Austria
| | | | - Shafagh Hadavi
- Graduate School of Media and Governance, Keio University, Fujisawa, Kanagawa, Japan
| | - Shinya Fujii
- Faculty of Environment and Information Studies, Keio University, Fujisawa, Kanagawa, Japan
| | - Sangbuem Choo
- Graduate School of Media and Governance, Keio University, Fujisawa, Kanagawa, Japan
| | - Marin Naruse
- Faculty of Policy Management, Keio University, Fujisawa, Kanagawa, Japan
| | | | - Latyr Sy
- Independent researcher, Tokyo, Japan
- Independent researcher, Dakar, Sénégal
| | - Mark Lenini Parselelo
- Memorial University of Newfoundland, St. John’s, NL, Canada
- Department of Music and Dance, Kenyatta University, Nairobi, Kenya
| | | | - Niels Chr. Hansen
- Aarhus Institute of Advanced Studies, Aarhus University, Aarhus, Denmark
- Centre of Excellence in Music, Mind, Body and Brain, University of Jyväskylä, Jyväskylä, Finland
- Interacting Minds Centre, School of Culture and Society, Aarhus University, Aarhus, Denmark
- Royal Academy of Music Aarhus/Aalborg, Aarhus, Denmark
| | - Felix Haiduk
- Department of Behavioral and Cognitive Biology, University of Vienna, Vienna, Austria
- Department of General Psychology, University of Padua, Padua, Italy
| | - Ulvhild Færøvik
- Institute of Biological and Medical Psychology, Department of Psychology, University of Bergen, Bergen, Norway
| | - Violeta Magalhães
- Centre of Linguistics of the University of Porto (CLUP), Porto, Portugal
- Faculty of Arts and Humanities of the University of Porto (FLUP), Porto, Portugal
- School of Education of the Polytechnic of Porto (ESE IPP), Porto, Portugal
| | - Wojciech Krzyżanowski
- Adam Mickiewicz University, Faculty of Art Studies, Musicology Institute, Poznań, Poland
| | | | - Diana Hereld
- Department of Psychiatry, UCLA Semel Institute for Neuroscience and Human Behavior, Los Angeles, CA, USA
| | | | | | | | | | - Su Zar Zar
- Headmistress, The Royal Music Academy, Yangon, Myanmar
| | - Iyadh El Kahla
- Department of Cultural Policy, University of Hildesheim, Hildesheim, Germany
| | - Olcay Muslu
- Centre for the Study of Higher Education, University of Kent, Canterbury, UK
- MIRAS, Centre for Cultural Sustainability, Istanbul, Turkey
| | - Jakelin Troy
- Director, Indigenous Research, Office of the Deputy Vice-Chancellor (Research); Department of Linguistics, Faculty of Arts and Social Sciences, The University of Sydney, Camperdown, NSW, Australia
| | - Teona Lomsadze
- International Research Center for Traditional Polyphony of the Tbilisi State Conservatoire, Tbilisi, Georgia
- Georgian Studies Fellow, University of Oxford, Oxford, UK
| | - Dilyana Kurdova
- South-West University Neofit Rilski, Blagoevgrad, Bulgaria
- Phoenix Perpeticum Foundation, Sofia, Bulgaria
| | | | | | - Aleksandar Arabadjiev
- Department of Folk Music Research and Ethnomusicology, University of Music and Performing Arts–MDW, Wien, Austria
| | | | - Adwoa Arhine
- Department of Music, University of Ghana, Accra, Ghana
| | - Tadhg Ó Meachair
- Department of Ethnomusicology and Folklore, Indiana University, Bloomington, IN, USA
| | - Javier Silva-Zurita
- Department of Humanities and Arts, University of Los Lagos, Osorno, Chile
- Millennium Nucleus on Musical and Sound Cultures (CMUS NCS 2022-16), Santiago, Chile
| | - Ignacio Soto-Silva
- Department of Humanities and Arts, University of Los Lagos, Osorno, Chile
- Millennium Nucleus on Musical and Sound Cultures (CMUS NCS 2022-16), Santiago, Chile
| | | | | | - Psyche Loui
- Music, Imaging and Neural Dynamics Lab, Northeastern University, Boston, MA, USA
| | - Andrea Ravignani
- Department of Human Neurosciences, Sapienza University of Rome, Rome, Italy
- Max Planck Institute for Psycholinguistics, Nijmegen, Netherlands
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University, Aarhus, Denmark & The Royal Academy of Music Aarhus/Aalborg, Aarhus, Denmark
| | - Yannick Jadoul
- Department of Human Neurosciences, Sapienza University of Rome, Rome, Italy
- Max Planck Institute for Psycholinguistics, Nijmegen, Netherlands
| | - Pauline Larrouy-Maestri
- Music Department, Max-Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany
- Max Planck—NYU Center for Language, Music, and Emotion (CLaME), New York, NY, USA
| | - Camila Bruder
- Music Department, Max-Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany
| | - Tutushamum Puri Teyxokawa
- Txemim Puri Project–Puri Language Research, Vitalization and Teaching/Recording and Preservation of Puri History and Culture, Rio de Janeiro, Brasil
| | | | | | | | - Limor Raviv
- Max Planck Institute for Psycholinguistics, Nijmegen, Netherlands
- cSCAN, University of Glasgow, Glasgow, UK
| | - Minyu Zeng
- Graduate School of Media and Governance, Keio University, Fujisawa, Kanagawa, Japan
- Rhode Island School of Design, Providence, RI, USA
| | - Shahaboddin Dabaghi Varnosfaderani
- Institute for English and American Studies (IEAS), Goethe University of Frankfurt am Main, Frankfurt am Main, Germany
- Cognitive and Developmental Psychology Unit, Centre, for Cognitive Science, University of Kaiserslautern-Landau (RPTU), Kaiserslautern, Germany
| | | | - Kayla Kolff
- Institute of Cognitive Science, University of Osnabrück, Osnabrück, Germany
| | | | - Meyha Chhatwal
- Department of Psychology, University of Toronto Mississauga, Mississauga, ON, Canada
| | - Ryan Mark David
- Department of Psychology, University of Toronto Mississauga, Mississauga, ON, Canada
| | | | - Great Lekakul
- Faculty of Fine Arts, Chiang Mai University, Chiang Mai, Thailand
| | - Vanessa Nina Borsan
- Graduate School of Media and Governance, Keio University, Fujisawa, Kanagawa, Japan
- Université de Lille, CNRS, Centrale Lille, UMR 9189 CRIStAL, F-59000 Lille, France
| | - Nozuko Nguqu
- School of Arts, Music Discipline, University of KwaZulu Natal, Durban, South Africa
| | - Patrick E. Savage
- School of Psychology, University of Auckland, Auckland, New Zealand
- Faculty of Environment and Information Studies, Keio University, Fujisawa, Kanagawa, Japan
| |
Collapse
|
16
|
Bruder C, Poeppel D, Larrouy-Maestri P. Perceptual (but not acoustic) features predict singing voice preferences. Sci Rep 2024; 14:8977. [PMID: 38637516 PMCID: PMC11026466 DOI: 10.1038/s41598-024-58924-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2023] [Accepted: 04/03/2024] [Indexed: 04/20/2024] Open
Abstract
Why do we prefer some singers to others? We investigated how much singing voice preferences can be traced back to objective features of the stimuli. To do so, we asked participants to rate short excerpts of singing performances in terms of how much they liked them as well as in terms of 10 perceptual attributes (e.g.: pitch accuracy, tempo, breathiness). We modeled liking ratings based on these perceptual ratings, as well as based on acoustic features and low-level features derived from Music Information Retrieval (MIR). Mean liking ratings for each stimulus were highly correlated between Experiments 1 (online, US-based participants) and 2 (in the lab, German participants), suggesting a role for attributes of the stimuli in grounding average preferences. We show that acoustic and MIR features barely explain any variance in liking ratings; in contrast, perceptual features of the voices achieved around 43% of prediction. Inter-rater agreement in liking and perceptual ratings was low, indicating substantial (and unsurprising) individual differences in participants' preferences and perception of the stimuli. Our results indicate that singing voice preferences are not grounded in acoustic attributes of the voices per se, but in how these features are perceptually interpreted by listeners.
Collapse
Affiliation(s)
- Camila Bruder
- Max Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany.
| | - David Poeppel
- New York University, New York, NY, USA
- Ernst Strüngmann Institute for Neuroscience, Frankfurt, Germany
- Max Planck-NYU Center for Language, Music, and Emotion (CLaME), New York, USA
| | - Pauline Larrouy-Maestri
- Max Planck Institute for Empirical Aesthetics, Frankfurt am Main, Germany
- Max Planck-NYU Center for Language, Music, and Emotion (CLaME), New York, USA
| |
Collapse
|