1
|
Zinsser LA, Marek A, Stone NI. Terminology of women's embodied experience of labouring and birthing sounds-a qualitative interview study. Midwifery 2025; 146:104422. [PMID: 40273543 DOI: 10.1016/j.midw.2025.104422] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/26/2024] [Revised: 03/14/2025] [Accepted: 04/09/2025] [Indexed: 04/26/2025]
Abstract
BACKGROUND Human beings have a diverse repertoire of linguistic and non-linguistic sounds used for self-expression. During labour and birth, women produce a range of sounds described in the scientific literature as moaning, primal sounds, roaring, screaming, singing, and vocalising. However, research on women's experiences of vocalisation during childbirth remains limited. AIM To explore women's experiences of the sounds they produce during labour and birth and deepen our understanding of these sounds. METHODS Semi-structured interviews with 18 women were conducted within the first six weeks postpartum. The interviews were transcribed verbatim and analysed using reflexive thematic analysis following the approach of Braun and Clarke, with MAXQDA used to organise and manage the data. FINDINGS Two themes emerged from the data: 'embodied sound and memory' and 'giving the sound a name'. The first theme explored women's embodied memory of sounds, including how they recalled these sounds during the interviews. The second theme explored in depth their subjective experiences and descriptions of the various sounds they produced. CONCLUSION The terminology used to describe the non-linguistic sounds produced by labouring and birthing women reflects a nuanced approach to conveying sound as a physical experience. Sounds such as breathing, humming, moaning, vocalising, and screaming, as well as the sounds made moments before birth, seem to be remembered as embodied, physical sensations rather than solely through cognitive reflection.
Collapse
Affiliation(s)
- Laura A Zinsser
- Hannover Medical School, Midwifery Research and Education Unit, Carl-Neuberg-Str. 1, 30625 Hannover, Germany.
| | - Annette Marek
- Hannover Medical School, Department of Speech and Language Therapy, Carl-Neuberg-Str. 1, 30625 Hannover, Germany.
| | - Nancy I Stone
- Hannover Medical School, Midwifery Research and Education Unit, Carl-Neuberg-Str. 1, 30625 Hannover, Germany.
| |
Collapse
|
2
|
Lameira AR, Caneco B, Kershenbaum A, Santamaría-Bonfil G, Call J. Generative vocal plasticity in chimpanzees. iScience 2025; 28:112381. [PMID: 40322082 PMCID: PMC12049825 DOI: 10.1016/j.isci.2025.112381] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/04/2024] [Revised: 01/15/2025] [Accepted: 04/04/2025] [Indexed: 05/08/2025] Open
Abstract
Modern theory posits that human-ape differences in voice command account for speech evolution. However, comparison has been indirect and conjectural based on vocal learning taxa far related from Hominids, instead of direct and quantitative based on great ape calls that, like all speech sounds, are local-specific and non-universal to the species. Moreover, the null hypothesis that the great ape voice command is purely reflexive has never been directly tested. Here, we show that in controlled, constant experimental settings, captive chimpanzees exhibit high-dimensional dexterity over voice activation and modulation in two atypical vowel-like calls. Subjects made unrestricted, multidimensional, and distinct voice changes within and between individuals, inducing parameter changes up to 10,000%, rejecting null hypothesis' predictions. Forecasting models indicated unmitigated voice novelty, altogether demonstrating emancipated and vast real-time voice control. Findings show that, contrary to traditional assumptions, speech and song evolution likely hinged on prolific voice command already available in ancestral ape-like ancestors.
Collapse
Affiliation(s)
| | | | - Arik Kershenbaum
- Department of Zoology, University of Cambridge, Cambridge, UK
- Girton College, University of Cambridge, Cambridge, UK
| | | | - Josep Call
- Department of Psychology and Neuroscience, University of St Andrews, St Andrews, UK
| |
Collapse
|
3
|
Fitch WT. Applying nonlinear dynamics to the voice: a historical perspective. Philos Trans R Soc Lond B Biol Sci 2025; 380:20240024. [PMID: 40176512 PMCID: PMC11966167 DOI: 10.1098/rstb.2024.0024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2024] [Revised: 08/29/2024] [Accepted: 09/04/2024] [Indexed: 04/04/2025] Open
Abstract
The recognition that nonlinear phenomena, including subharmonics, bifurcations and deterministic chaos, are present in human and animal vocalizations is a relatively recent one. I give a brief history of this revolution in our understanding of the voice, based on interviews with some of the key players and personal experience. Most of the key concepts and mathematical principles of nonlinear dynamics were already well worked out in the early 1980s. In the early 1990s, physicist Hanspeter Herzel and colleagues in Berlin recognized that these principles are applicable to the human voice, initially to baby cries. The physics and physiology underlying many of these nonlinear phenomena had remained mysterious up until then. This insight was later generalized to animal vocalizations. Nonlinear phenomena play a relatively peripheral role in most human vocal communication but are a common feature of many animal vocalizations. The broad recognition of the existence of nonlinear vocalizations, and the quantitative study of their production and perception, has now fuelled important and exciting advances in our understanding of animal communication. I concentrate on how the core concepts came into focus, and on their initial application to an ever-wider circle of call types and species, and end with a brief prospectus for the future.This article is part of the theme issue 'Nonlinear phenomena in vertebrate vocalizations: mechanisms and communicative functions'.
Collapse
Affiliation(s)
- W. Tecumseh Fitch
- Department of Behavioral and Cognitive Biology, University of Vienna, Vienna, Austria
| |
Collapse
|
4
|
Daunay V, Reby D, Bryant GA, Pisanski K. Production and perception of volitional laughter across social contexts. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2025; 157:2774-2789. [PMID: 40227885 DOI: 10.1121/10.0036388] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/11/2024] [Accepted: 03/23/2025] [Indexed: 04/16/2025]
Abstract
Human nonverbal vocalizations such as laughter communicate emotion, motivation, and intent during social interactions. While differences between spontaneous and volitional laughs have been described, little is known about the communicative functions of volitional (voluntary) laughter-a complex signal used across diverse social contexts. Here, we examined whether the acoustic structure of volitional laughter encodes social contextual information recognizable by humans and computers. We asked men and women to produce volitional laughs in eight distinct social contexts ranging from positive (e.g., watching a comedy) to negative valence (e.g., embarrassment). Human listeners and machine classification algorithms accurately identified most laughter contexts above chance. However, confusion often arose within valence categories, and could be largely explained by shared acoustics. Although some acoustic features varied across social contexts, including fundamental frequency (perceived as voice pitch) and energy parameters (entropy variance, loudness, spectral centroid, and cepstral peak prominence), which also predicted listeners' recognition of laughter contexts, laughs evoked across different social contexts still often overlapped in acoustic and perceptual space. Thus, we show that volitional laughter can convey some reliable information about social context, but much of this is tied to valence, suggesting that volitional laughter is a graded rather than discrete vocal signal.
Collapse
Affiliation(s)
- Virgile Daunay
- ENES Bioacoustics Research Laboratory, CRNL Center for Research in Neuroscience in Lyon, University of Saint-Étienne, 42023 Saint-Étienne, France
- DDL Dynamics of Language Lab, CNRS French National Centre for Scientific Research, University of Lyon 2, 69363 Lyon, France
| | - David Reby
- ENES Bioacoustics Research Laboratory, CRNL Center for Research in Neuroscience in Lyon, University of Saint-Étienne, 42023 Saint-Étienne, France
| | - Gregory A Bryant
- Department of Communication, Center for Behavior, Evolution, and Culture, University of California, Los Angeles, California 90095, USA
| | - Katarzyna Pisanski
- ENES Bioacoustics Research Laboratory, CRNL Center for Research in Neuroscience in Lyon, University of Saint-Étienne, 42023 Saint-Étienne, France
- DDL Dynamics of Language Lab, CNRS French National Centre for Scientific Research, University of Lyon 2, 69363 Lyon, France
| |
Collapse
|
5
|
Ponsonnet M, Coupé C, Pellegrino F, Garcia Arasco A, Pisanski K. Vowel signatures in emotional interjections and nonlinguistic vocalizations expressing pain, disgust, and joy across languagesa). THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 156:3118-3139. [PMID: 39531311 DOI: 10.1121/10.0032454] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/03/2024] [Accepted: 10/01/2024] [Indexed: 11/16/2024]
Abstract
In this comparative cross-linguistic study we test whether expressive interjections (words like ouch or yay) share similar vowel signatures across the world's languages, and whether these can be traced back to nonlinguistic vocalizations (like screams and cries) expressing the same emotions of pain, disgust, and joy. We analyze vowels in interjections from dictionaries of 131 languages (over 600 tokens) and compare these with nearly 500 vowels based on formant frequency measures from voice recordings of volitional nonlinguistic vocalizations. We show that across the globe, pain interjections feature a-like vowels and wide falling diphthongs ("ai" as in Ayyy! "aw" as in Ouch!), whereas disgust and joy interjections do not show robust vowel regularities that extend geographically. In nonlinguistic vocalizations, all emotions yield distinct vowel signatures: pain prompts open vowels such as [a], disgust schwa-like central vowels, and joy front vowels such as [i]. Our results show that pain is the only affective experience tested with a clear, robust vowel signature that is preserved between nonlinguistic vocalizations and interjections across languages. These results offer empirical evidence for iconicity in some expressive interjections. We consider potential mechanisms and origins, from evolutionary pressures and sound symbolism to colexification, proposing testable hypotheses for future research.
Collapse
Affiliation(s)
- Maïa Ponsonnet
- Dynamique Du Langage, CNRS et Université Lumière Lyon 2, Lyon, France
- School of Social Sciences, The University of Western Australia, Perth, Australia
| | - Christophe Coupé
- Department of Linguistics, The University of Hong Kong, Hong Kong SAR, China
| | | | | | - Katarzyna Pisanski
- Dynamique Du Langage, CNRS et Université Lumière Lyon 2, Lyon, France
- ENES Bioacoustics Research Laboratory, University Jean Monnet of Saint-Etienne, CRNL, CNRS, Saint-Etienne, France
- Institute of Psychology, University of Wrocław, Wrocław, Poland
| |
Collapse
|
6
|
Pisanski K, Reby D, Oleszkiewicz A. Humans need auditory experience to produce typical volitional nonverbal vocalizations. COMMUNICATIONS PSYCHOLOGY 2024; 2:65. [PMID: 39242947 PMCID: PMC11332021 DOI: 10.1038/s44271-024-00104-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/03/2023] [Accepted: 05/16/2024] [Indexed: 09/09/2024]
Abstract
Human nonverbal vocalizations such as screams and cries often reflect their evolved functions. Although the universality of these putatively primordial vocal signals and their phylogenetic roots in animal calls suggest a strong reflexive foundation, many of the emotional vocalizations that we humans produce are under our voluntary control. This suggests that, like speech, volitional vocalizations may require auditory input to develop typically. Here, we acoustically analyzed hundreds of volitional vocalizations produced by profoundly deaf adults and typically-hearing controls. We show that deaf adults produce unconventional and homogenous vocalizations of aggression and pain that are unusually high-pitched, unarticulated, and with extremely few harsh-sounding nonlinear phenomena compared to controls. In contrast, fear vocalizations of deaf adults are relatively acoustically typical. In four lab experiments involving a range of perception tasks with 444 participants, listeners were less accurate in identifying the intended emotions of vocalizations produced by deaf vocalizers than by controls, perceived their vocalizations as less authentic, and reliably detected deafness. Vocalizations of congenitally deaf adults with zero auditory experience were most atypical, suggesting additive effects of auditory deprivation. Vocal learning in humans may thus be required not only for speech, but also to acquire the full repertoire of volitional non-linguistic vocalizations.
Collapse
Affiliation(s)
- Katarzyna Pisanski
- ENES Bioacoustics Research Laboratory, CRNL Center for Research in Neuroscience in Lyon, University of Saint-Étienne, 42023, Saint-Étienne, France.
- CNRS French National Centre for Scientific Research, DDL Dynamics of Language Lab, University of Lyon 2, 69007, Lyon, France.
- Institute of Psychology, University of Wrocław, 50-527, Wrocław, Poland.
| | - David Reby
- ENES Bioacoustics Research Laboratory, CRNL Center for Research in Neuroscience in Lyon, University of Saint-Étienne, 42023, Saint-Étienne, France
- Institut Universitaire de France, Paris, France
| | - Anna Oleszkiewicz
- Institute of Psychology, University of Wrocław, 50-527, Wrocław, Poland.
- Department of Otorhinolaryngology, Smell and Taste Clinic, Carl Gustav Carus Medical School, Technische Universitaet Dresden, 01307, Dresden, Germany.
| |
Collapse
|
7
|
Yano-Nashimoto S, Truzzi A, Shinozuka K, Murayama AY, Kurachi T, Moriya-Ito K, Tokuno H, Miyazawa E, Esposito G, Okano H, Nakamura K, Saito A, Kuroda KO. Anxious about rejection, avoidant of neglect: Infant marmosets tune their attachment based on individual caregiver's parenting style. Commun Biol 2024; 7:212. [PMID: 38378797 PMCID: PMC10879543 DOI: 10.1038/s42003-024-05875-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2023] [Accepted: 01/30/2024] [Indexed: 02/22/2024] Open
Abstract
Children's secure attachment with their primary caregivers is crucial for physical, cognitive, and emotional maturation. Yet, the causal links between specific parenting behaviors and infant attachment patterns are not fully understood. Here we report infant attachment in New World monkeys common marmosets, characterized by shared infant care among parents and older siblings and complex vocal communications. By integrating natural variations in parenting styles and subsecond-scale microanalyses of dyadic vocal and physical interactions, we demonstrate that marmoset infants signal their needs through context-dependent call use and selective approaches toward familiar caregivers. The infant attachment behaviors are tuned to each caregiver's parenting style; infants use negative calls when carried by rejecting caregivers and selectively avoid neglectful and rejecting caregivers. Family-deprived infants fail to develop such adaptive uses of attachment behaviors. With these similarities with humans, marmosets offer a promising model for investigating the biological mechanisms of attachment security.
Collapse
Affiliation(s)
- Saori Yano-Nashimoto
- Laboratory for Affiliative Social Behavior, RIKEN Center for Brain Science, Wako, Japan
- Laboratory of Physiology, Department of Basic Veterinary Sciences, Graduate School of Veterinary Medicine, Hokkaido University, Sapporo, Japan
| | - Anna Truzzi
- Laboratory for Affiliative Social Behavior, RIKEN Center for Brain Science, Wako, Japan
- Trinity College Institute of Neuroscience, School of Psychology, Trinity College Dublin, Dublin, Ireland
- Department of Psychology and Cognitive Science, University of Trento, Rovereto, TN, Italy
| | - Kazutaka Shinozuka
- Laboratory for Affiliative Social Behavior, RIKEN Center for Brain Science, Wako, Japan
- Planning, Review and Research Institute for Social insurance and Medical program, Chiyoda-ku, Japan
| | - Ayako Y Murayama
- Laboratory for Affiliative Social Behavior, RIKEN Center for Brain Science, Wako, Japan
- Department of Physiology, Keio University School of Medicine, Shinjuku-ku, Japan
- Laboratory for Marmoset Neural Architecture, RIKEN Center for Brain Science, Wako, Japan
- Neural Circuit Unit, Okinawa Institute Science and Technology Graduate University, Onna, Japan
| | - Takuma Kurachi
- Laboratory for Affiliative Social Behavior, RIKEN Center for Brain Science, Wako, Japan
- Department of Agriculture, Tokyo University of Agriculture and Technology, Fuchu, Japan
| | - Keiko Moriya-Ito
- Department of Brain & Neurosciences, Tokyo Metropolitan Institute of Medical Science, Setagaya-ku, Japan
| | - Hironobu Tokuno
- Department of Brain & Neurosciences, Tokyo Metropolitan Institute of Medical Science, Setagaya-ku, Japan
| | - Eri Miyazawa
- Laboratory for Affiliative Social Behavior, RIKEN Center for Brain Science, Wako, Japan
| | - Gianluca Esposito
- Laboratory for Affiliative Social Behavior, RIKEN Center for Brain Science, Wako, Japan
- Department of Psychology and Cognitive Science, University of Trento, Rovereto, TN, Italy
| | - Hideyuki Okano
- Department of Physiology, Keio University School of Medicine, Shinjuku-ku, Japan
- Laboratory for Marmoset Neural Architecture, RIKEN Center for Brain Science, Wako, Japan
| | - Katsuki Nakamura
- Center for the Evolutionary Origins of Human Behavior, Kyoto University, Inuyama, Japan
| | - Atsuko Saito
- Laboratory for Affiliative Social Behavior, RIKEN Center for Brain Science, Wako, Japan.
- Department of Psychology, Sophia University, Chiyoda-ku, Japan.
| | - Kumi O Kuroda
- Laboratory for Affiliative Social Behavior, RIKEN Center for Brain Science, Wako, Japan.
- Kuroda Laboratory, School of Life Science and Technology, Tokyo Institute of Technology, Yokohama, Japan.
- Laboratory for Circuit and Behavioral Physiology, RIKEN Center for Brain Science, Wako, Japan.
| |
Collapse
|
8
|
Anikin A, Canessa-Pollard V, Pisanski K, Massenet M, Reby D. Beyond speech: Exploring diversity in the human voice. iScience 2023; 26:108204. [PMID: 37908309 PMCID: PMC10613903 DOI: 10.1016/j.isci.2023.108204] [Citation(s) in RCA: 16] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2023] [Revised: 07/20/2023] [Accepted: 10/11/2023] [Indexed: 11/02/2023] Open
Abstract
Humans have evolved voluntary control over vocal production for speaking and singing, while preserving the phylogenetically older system of spontaneous nonverbal vocalizations such as laughs and screams. To test for systematic acoustic differences between these vocal domains, we analyzed a broad, cross-cultural corpus representing over 2 h of speech, singing, and nonverbal vocalizations. We show that, while speech is relatively low-pitched and tonal with mostly regular phonation, singing and especially nonverbal vocalizations vary enormously in pitch and often display harsh-sounding, irregular phonation owing to nonlinear phenomena. The evolution of complex supralaryngeal articulatory spectro-temporal modulation has been critical for speech, yet has not significantly constrained laryngeal source modulation. In contrast, articulation is very limited in nonverbal vocalizations, which predominantly contain minimally articulated open vowels and rapid temporal modulation in the roughness range. We infer that vocal source modulation works best for conveying affect, while vocal filter modulation mainly facilitates semantic communication.
Collapse
Affiliation(s)
- Andrey Anikin
- Division of Cognitive Science, Lund University, Lund, Sweden
- ENES Bioacoustics Research Lab, CRNL, University of Saint-Etienne, CNRS, Inserm, 23 rue Michelon, 42023 Saint-Etienne, France
| | - Valentina Canessa-Pollard
- ENES Bioacoustics Research Lab, CRNL, University of Saint-Etienne, CNRS, Inserm, 23 rue Michelon, 42023 Saint-Etienne, France
- Psychology, Institute of Psychology, Business and Human Sciences, University of Chichester, Chichester, West Sussex PO19 6PE, UK
| | - Katarzyna Pisanski
- ENES Bioacoustics Research Lab, CRNL, University of Saint-Etienne, CNRS, Inserm, 23 rue Michelon, 42023 Saint-Etienne, France
- CNRS French National Centre for Scientific Research, DDL Dynamics of Language Lab, University of Lyon 2, 69007 Lyon, France
- Institute of Psychology, University of Wrocław, Dawida 1, 50-527 Wrocław, Poland
| | - Mathilde Massenet
- ENES Bioacoustics Research Lab, CRNL, University of Saint-Etienne, CNRS, Inserm, 23 rue Michelon, 42023 Saint-Etienne, France
| | - David Reby
- ENES Bioacoustics Research Lab, CRNL, University of Saint-Etienne, CNRS, Inserm, 23 rue Michelon, 42023 Saint-Etienne, France
| |
Collapse
|
9
|
Ceravolo L, Debracque C, Pool E, Gruber T, Grandjean D. Frontal mechanisms underlying primate calls recognition by humans. Cereb Cortex Commun 2023; 4:tgad019. [PMID: 38025828 PMCID: PMC10661312 DOI: 10.1093/texcom/tgad019] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2023] [Revised: 10/18/2023] [Accepted: 10/25/2023] [Indexed: 12/01/2023] Open
Abstract
Introduction The ability to process verbal language seems unique to humans and relies not only on semantics but on other forms of communication such as affective vocalizations, that we share with other primate species-particularly great apes (Hominidae). Methods To better understand these processes at the behavioral and brain level, we asked human participants to categorize vocalizations of four primate species including human, great apes (chimpanzee and bonobo), and monkey (rhesus macaque) during MRI acquisition. Results Classification was above chance level for all species but bonobo vocalizations. Imaging analyses were computed using a participant-specific, trial-by-trial fitted probability categorization value in a model-based style of data analysis. Model-based analyses revealed the implication of the bilateral orbitofrontal cortex and inferior frontal gyrus pars triangularis (IFGtri) respectively correlating and anti-correlating with the fitted probability of accurate species classification. Further conjunction analyses revealed enhanced activity in a sub-area of the left IFGtri specifically for the accurate classification of chimpanzee calls compared to human voices. Discussion Our data-that are controlled for acoustic variability between species-therefore reveal distinct frontal mechanisms that shed light on how the human brain evolved to process vocal signals.
Collapse
Affiliation(s)
- Leonardo Ceravolo
- Neuroscience of Emotions and Affective Dynamics lab, Department of Psychology and Educational Sciences, University of Geneva, Unimail building, Boulevard Pont-d’Arve 40CH-1205, Geneva, Switzerland
- Swiss Center for Affective Sciences, University of Geneva, Campus Biotech building, Chemin des Mines 9CH-1202, Geneva, Switzerland
| | - Coralie Debracque
- Neuroscience of Emotions and Affective Dynamics lab, Department of Psychology and Educational Sciences, University of Geneva, Unimail building, Boulevard Pont-d’Arve 40CH-1205, Geneva, Switzerland
- Swiss Center for Affective Sciences, University of Geneva, Campus Biotech building, Chemin des Mines 9CH-1202, Geneva, Switzerland
| | - Eva Pool
- Swiss Center for Affective Sciences, University of Geneva, Campus Biotech building, Chemin des Mines 9CH-1202, Geneva, Switzerland
- E3 Lab, Department of Psychology and Educational Sciences, University of Geneva, Unimail building, Boulevard Pont-d’Arve 40CH-1205, Geneva, Switzerland
| | - Thibaud Gruber
- Neuroscience of Emotions and Affective Dynamics lab, Department of Psychology and Educational Sciences, University of Geneva, Unimail building, Boulevard Pont-d’Arve 40CH-1205, Geneva, Switzerland
- Swiss Center for Affective Sciences, University of Geneva, Campus Biotech building, Chemin des Mines 9CH-1202, Geneva, Switzerland
- eccePAN lab, Department of Psychology and Educational Sciences, University of Geneva, Campus Biotech building, Chemin des Mines 9CH-1202, Geneva, Switzerland
| | - Didier Grandjean
- Neuroscience of Emotions and Affective Dynamics lab, Department of Psychology and Educational Sciences, University of Geneva, Unimail building, Boulevard Pont-d’Arve 40CH-1205, Geneva, Switzerland
- Swiss Center for Affective Sciences, University of Geneva, Campus Biotech building, Chemin des Mines 9CH-1202, Geneva, Switzerland
| |
Collapse
|
10
|
Jafari A, Dureux A, Zanini A, Menon RS, Gilbert KM, Everling S. A vocalization-processing network in marmosets. Cell Rep 2023; 42:112526. [PMID: 37195863 DOI: 10.1016/j.celrep.2023.112526] [Citation(s) in RCA: 16] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2023] [Revised: 03/31/2023] [Accepted: 05/02/2023] [Indexed: 05/19/2023] Open
Abstract
Vocalizations play an important role in the daily life of primates and likely form the basis of human language. Functional imaging studies have demonstrated that listening to voices activates a fronto-temporal voice perception network in human participants. Here, we acquired whole-brain ultrahigh-field (9.4 T) fMRI in awake marmosets (Callithrix jacchus) and demonstrate that these small, highly vocal New World primates possess a similar fronto-temporal network, including subcortical regions, that is activated by the presentation of conspecific vocalizations. The findings suggest that the human voice perception network has evolved from an ancestral vocalization-processing network that predates the separation of New and Old World primates.
Collapse
Affiliation(s)
- Azadeh Jafari
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, Canada
| | - Audrey Dureux
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, Canada
| | - Alessandro Zanini
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, Canada
| | - Ravi S Menon
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, Canada
| | - Kyle M Gilbert
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, Canada
| | - Stefan Everling
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, Canada; Department of Physiology and Pharmacology, University of Western Ontario, London, ON, Canada.
| |
Collapse
|
11
|
Loosening the leash: The unique emotional canvas of human screams. Behav Brain Sci 2023; 46:e10. [PMID: 36799052 DOI: 10.1017/s0140525x22000851] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/18/2023]
Abstract
We use screams to explore ideas presented in the target article. Evolving first in animals as a response to predation, screams reveal more complex social use in nonhuman primates and, in humans, uniquely, are associated with a much greater variety of emotional contexts including fear, anger, surprise, and happiness. This expansion, and the potential for manipulation, promotes listener social vigilance.
Collapse
|
12
|
Donhauser PW, Klein D. Audio-Tokens: A toolbox for rating, sorting and comparing audio samples in the browser. Behav Res Methods 2023; 55:508-515. [PMID: 35297013 PMCID: PMC10027774 DOI: 10.3758/s13428-022-01803-w] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 01/19/2022] [Indexed: 12/30/2022]
Abstract
Here we describe a JavaScript toolbox to perform online rating studies with auditory material. The main feature of the toolbox is that audio samples are associated with visual tokens on the screen that control audio playback and can be manipulated depending on the type of rating. This allows the collection of single- and multidimensional feature ratings, as well as categorical and similarity ratings. The toolbox ( github.com/pwdonh/audio_tokens ) can be used via a plugin for the widely used jsPsych, as well as using plain JavaScript for custom applications. We expect the toolbox to be useful in psychological research on speech and music perception, as well as for the curation and annotation of datasets in machine learning.
Collapse
Affiliation(s)
- Peter W Donhauser
- Cognitive Neuroscience Unit, Montreal Neurological Institute, McGill University, Montreal, QC, H3A 2B4, Canada.
- Ernst Strüngmann Institute for Neuroscience in Cooperation with Max Planck Society, 60528, Frankfurt am Main, Germany.
| | - Denise Klein
- Cognitive Neuroscience Unit, Montreal Neurological Institute, McGill University, Montreal, QC, H3A 2B4, Canada.
- Centre for Research on Brain, Language and Music, McGill University, Montreal, QC, H3G 2A8, Canada.
| |
Collapse
|
13
|
Schwartz JW, Gouzoules H. Humans read emotional arousal in monkey vocalizations: evidence for evolutionary continuities in communication. PeerJ 2022; 10:e14471. [PMID: 36518288 PMCID: PMC9744152 DOI: 10.7717/peerj.14471] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2022] [Accepted: 11/06/2022] [Indexed: 12/05/2022] Open
Abstract
Humans and other mammalian species communicate emotions in ways that reflect evolutionary conservation and continuity, an observation first made by Darwin. One approach to testing this hypothesis has been to assess the capacity to perceive the emotional content of the vocalizations of other species. Using a binary forced choice task, we tested perception of the emotional intensity represented in coos and screams of infant and juvenile female rhesus macaques (Macaca mulatta) by 113 human listeners without, and 12 listeners with, experience (as researchers or care technicians) with this species. Each stimulus pair contained one high- and one low-arousal vocalization, as measured at the time of recording by stress hormone levels for coos and the degree of intensity of aggression for screams. For coos as well as screams, both inexperienced and experienced participants accurately identified the high-arousal vocalization at significantly above-chance rates. Experience was associated with significantly greater accuracy with scream stimuli but not coo stimuli, and with a tendency to indicate screams as reflecting greater emotional intensity than coos. Neither measures of empathy, human emotion recognition, nor attitudes toward animal welfare showed any relationship with responses. Participants were sensitive to the fundamental frequency, noisiness, and duration of vocalizations; some of these tendencies likely facilitated accurate perceptions, perhaps due to evolutionary homologies in the physiology of arousal and vocal production between humans and macaques. Overall, our findings support a view of evolutionary continuity in emotional vocal communication. We discuss hypotheses about how distinctive dimensions of human nonverbal communication, like the expansion of scream usage across a range of contexts, might influence perceptions of other species' vocalizations.
Collapse
Affiliation(s)
- Jay W. Schwartz
- Department of Psychology, Emory University, Atlanta, GA, United States,Psychological Sciences Department, Western Oregon University, Monmouth, OR, United States
| | - Harold Gouzoules
- Department of Psychology, Emory University, Atlanta, GA, United States
| |
Collapse
|
14
|
Anikin A, Reby D. Ingressive phonation conveys arousal in human nonverbal vocalizations. BIOACOUSTICS 2022. [DOI: 10.1080/09524622.2022.2039295] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
Affiliation(s)
- Andrey Anikin
- Division of Cognitive Science, Lund University, Lund, Sweden
- Enes Sensory Neuro-Ethology Lab, Crnl, Jean Monnet University of Saint Étienne, St-Étienne, France
| | - David Reby
- Enes Sensory Neuro-Ethology Lab, Crnl, Jean Monnet University of Saint Étienne, St-Étienne, France
| |
Collapse
|
15
|
Pisanski K, Bryant GA, Cornec C, Anikin A, Reby D. Form follows function in human nonverbal vocalisations. ETHOL ECOL EVOL 2022. [DOI: 10.1080/03949370.2022.2026482] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Affiliation(s)
- Katarzyna Pisanski
- ENES Sensory Neuro-Ethology Lab, CRNL, Jean Monnet University of Saint Étienne, UMR 5293, St-Étienne 42023, France
- CNRS French National Centre for Scientific Research, DDL Dynamics of Language Lab, University of Lyon 2, Lyon 69007, France
| | - Gregory A. Bryant
- Department of Communication, Center for Behavior, Evolution, and Culture, University of California, Los Angeles, California, USA
| | - Clément Cornec
- ENES Sensory Neuro-Ethology Lab, CRNL, Jean Monnet University of Saint Étienne, UMR 5293, St-Étienne 42023, France
| | - Andrey Anikin
- ENES Sensory Neuro-Ethology Lab, CRNL, Jean Monnet University of Saint Étienne, UMR 5293, St-Étienne 42023, France
- Division of Cognitive Science, Lund University, Lund 22100, Sweden
| | - David Reby
- ENES Sensory Neuro-Ethology Lab, CRNL, Jean Monnet University of Saint Étienne, UMR 5293, St-Étienne 42023, France
| |
Collapse
|
16
|
Icht M, Wiznitser Ressis-Tal H, Lotan M. Can the Vocal Expression of Intellectually Disabled Individuals Be Used as a Pain Indicator? Initial Findings Supporting a Possible Novice Assessment Method. Front Psychol 2021; 12:655202. [PMID: 34366973 PMCID: PMC8339267 DOI: 10.3389/fpsyg.2021.655202] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2021] [Accepted: 06/22/2021] [Indexed: 11/23/2022] Open
Abstract
Pain is difficult to assess in non-verbal populations such as individuals with intellectual and developmental disability (IDD). Due to scarce research in this area, pain assessment for individuals with IDD is still lacking, leading to maltreatment. To improve medical care for individuals with IDD, immediate, reliable, easy to use pain detection methods should be developed. The goal of this preliminary study was to examine the sensitivity of acoustic features of vocal expressions in identifying pain for adults with IDD, assessing their feasibility as a pain detection indicator for those individuals. Such unique pain related vocal characteristics may be used to develop objective pain detection means. Adults with severe-profound IDD level (N = 9) were recorded in daily activities associated with pain (during diaper changes), or without pain (at rest). Spontaneous vocal expressions were acoustically analyzed to assess several voice characteristics. Analyzing the data revealed that pain related vocal expressions were characterized by significantly higher number of pulses and higher shimmer values relative to no-pain vocal expressions. Pain related productions were also characterized by longer duration, higher jitter and Cepstral Peak Prominence values, lower Harmonic-Noise Ratio, lower difference between the amplitude of the 1st and 2nd harmonic (corrected for vocal tract influence; H1H2c), and higher mean and standard deviation of voice fundamental frequency relative to no-pain related vocal productions, yet these findings were not statistically significant, possibly due to the small and heterogeneous sample. These initial results may prompt further research to explore the possibility to use pain related vocal output as an objective and easily identifiable indicator of pain in this population.
Collapse
Affiliation(s)
- Michal Icht
- Department of Communication Disorders, Ariel University, Ariel, Israel
| | | | - Meir Lotan
- Department of Physiotherapy, Ariel University, Ariel, Israel
| |
Collapse
|
17
|
Morfi V, Lachlan RF, Stowell D. Deep perceptual embeddings for unlabelled animal sound events. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 150:2. [PMID: 34340499 DOI: 10.1121/10.0005475] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/29/2020] [Accepted: 06/07/2021] [Indexed: 06/13/2023]
Abstract
Evaluating sound similarity is a fundamental building block in acoustic perception and computational analysis. Traditional data-driven analyses of perceptual similarity are based on heuristics or simplified linear models, and are thus limited. Deep learning embeddings, often using triplet networks, have been useful in many fields. However, such networks are usually trained using large class-labelled datasets. Such labels are not always feasible to acquire. We explore data-driven neural embeddings for sound event representation when class labels are absent, instead utilising proxies of perceptual similarity judgements. Ultimately, our target is to create a perceptual embedding space that reflects animals' perception of sound. We create deep perceptual embeddings for bird sounds using triplet models. In order to deal with the challenging nature of triplet loss training with the lack of class-labelled data, we utilise multidimensional scaling (MDS) pretraining, attention pooling, and a triplet mining scheme. We also evaluate the advantage of triplet learning compared to learning a neural embedding from a model trained on MDS alone. Using computational proxies of similarity judgements, we demonstrate the feasibility of the method to develop perceptual models for a wide range of data based on behavioural judgements, helping us understand how animals perceive sounds.
Collapse
Affiliation(s)
- Veronica Morfi
- Machine Listening Lab, Centre for Digital Music (C4DM), Department of Electronic Engineering, Queen Mary University of London, London, United Kingdom
| | - Robert F Lachlan
- Department of Psychology, Royal Holloway University of London, London, United Kingdom
| | - Dan Stowell
- Machine Listening Lab, Centre for Digital Music (C4DM), Department of Electronic Engineering, Queen Mary University of London, London, United Kingdom
| |
Collapse
|
18
|
Frühholz S, Dietziker J, Staib M, Trost W. Neurocognitive processing efficiency for discriminating human non-alarm rather than alarm scream calls. PLoS Biol 2021; 19:e3000751. [PMID: 33848299 PMCID: PMC8043411 DOI: 10.1371/journal.pbio.3000751] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2020] [Accepted: 02/15/2021] [Indexed: 11/19/2022] Open
Abstract
Across many species, scream calls signal the affective significance of events to other agents. Scream calls were often thought to be of generic alarming and fearful nature, to signal potential threats, with instantaneous, involuntary, and accurate recognition by perceivers. However, scream calls are more diverse in their affective signaling nature than being limited to fearfully alarming a threat, and thus the broader sociobiological relevance of various scream types is unclear. Here we used 4 different psychoacoustic, perceptual decision-making, and neuroimaging experiments in humans to demonstrate the existence of at least 6 psychoacoustically distinctive types of scream calls of both alarming and non-alarming nature, rather than there being only screams caused by fear or aggression. Second, based on perceptual and processing sensitivity measures for decision-making during scream recognition, we found that alarm screams (with some exceptions) were overall discriminated the worst, were responded to the slowest, and were associated with a lower perceptual sensitivity for their recognition compared with non-alarm screams. Third, the neural processing of alarm compared with non-alarm screams during an implicit processing task elicited only minimal neural signal and connectivity in perceivers, contrary to the frequent assumption of a threat processing bias of the primate neural system. These findings show that scream calls are more diverse in their signaling and communicative nature in humans than previously assumed, and, in contrast to a commonly observed threat processing bias in perceptual discriminations and neural processes, we found that especially non-alarm screams, and positive screams in particular, seem to have higher efficiency in speeded discriminations and the implicit neural processing of various scream types in humans.
Collapse
Affiliation(s)
- Sascha Frühholz
- Cognitive and Affective Neuroscience Unit, University of Zurich, Zurich, Switzerland
- Neuroscience Center Zurich, University of Zurich and ETH Zurich, Zurich, Switzerland
- Department of Psychology, University of Oslo, Oslo, Norway
- Center for the Interdisciplinary Study of Language Evolution, University of Zurich, Zurich, Switzerland
- * E-mail:
| | - Joris Dietziker
- Cognitive and Affective Neuroscience Unit, University of Zurich, Zurich, Switzerland
| | - Matthias Staib
- Cognitive and Affective Neuroscience Unit, University of Zurich, Zurich, Switzerland
| | - Wiebke Trost
- Cognitive and Affective Neuroscience Unit, University of Zurich, Zurich, Switzerland
| |
Collapse
|
19
|
Engelberg JWM, Schwartz JW, Gouzoules H. The emotional canvas of human screams: patterns and acoustic cues in the perceptual categorization of a basic call type. PeerJ 2021; 9:e10990. [PMID: 33854835 PMCID: PMC7953872 DOI: 10.7717/peerj.10990] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2020] [Accepted: 02/01/2021] [Indexed: 11/20/2022] Open
Abstract
Screams occur across taxonomically widespread species, typically in antipredator situations, and are strikingly similar acoustically, but in nonhuman primates, they have taken on acoustically varied forms in association with more contextually complex functions related to agonistic recruitment. Humans scream in an even broader range of contexts, but the extent to which acoustic variation allows listeners to perceive different emotional meanings remains unknown. We investigated how listeners responded to 30 contextually diverse human screams on six different emotion prompts as well as how selected acoustic cues predicted these responses. We found that acoustic variation in screams was associated with the perception of different emotions from these calls. Emotion ratings generally fell along two dimensions: one contrasting perceived anger, frustration, and pain with surprise and happiness, roughly associated with call duration and roughness, and one related to perceived fear, associated with call fundamental frequency. Listeners were more likely to rate screams highly in emotion prompts matching the source context, suggesting that some screams conveyed information about emotional context, but it is noteworthy that the analysis of screams from happiness contexts (n = 11 screams) revealed that they more often yielded higher ratings of fear. We discuss the implications of these findings for the role and evolution of nonlinguistic vocalizations in human communication, including consideration of how the expanded diversity in calls such as human screams might represent a derived function of language.
Collapse
Affiliation(s)
| | - Jay W. Schwartz
- Department of Psychology, Emory University, Atlanta, GA, USA
- Psychological Sciences Department, Western Oregon University, Monmouth, OR, USA
| | | |
Collapse
|
20
|
Fishbein AR, Prior NH, Brown JA, Ball GF, Dooling RJ. Discrimination of natural acoustic variation in vocal signals. Sci Rep 2021; 11:916. [PMID: 33441711 PMCID: PMC7807010 DOI: 10.1038/s41598-020-79641-z] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2020] [Accepted: 12/09/2020] [Indexed: 01/29/2023] Open
Abstract
Studies of acoustic communication often focus on the categories and units of vocalizations, but subtle variation also occurs in how these signals are uttered. In human speech, it is not only phonemes and words that carry information but also the timbre, intonation, and stress of how speech sounds are delivered (often referred to as "paralinguistic content"). In non-human animals, variation across utterances of vocal signals also carries behaviorally relevant information across taxa. However, the discriminability of these cues has been rarely tested in a psychophysical paradigm. Here, we focus on acoustic communication in the zebra finch (Taeniopygia guttata), a songbird species in which the male produces a single stereotyped motif repeatedly in song bouts. These motif renditions, like the song repetitions of many birds, sound very similar to the casual human listener. In this study, we show that zebra finches can easily discriminate between the renditions, even at the level of single song syllables, much as humans can discriminate renditions of speech sounds. These results support the notion that sensitivity to fine acoustic details may be a primary channel of information in zebra finch song, as well as a shared, foundational property of vocal communication systems across species.
Collapse
Affiliation(s)
- Adam R. Fishbein
- grid.164295.d0000 0001 0941 7177Department of Psychology, University of Maryland, Biology-Psychology Bldg., 4094 Campus Dr., College Park, MD 20742 USA ,grid.164295.d0000 0001 0941 7177Neuroscience and Cognitive Science Program, University of Maryland, College Park, MD USA
| | - Nora H. Prior
- grid.164295.d0000 0001 0941 7177Department of Psychology, University of Maryland, Biology-Psychology Bldg., 4094 Campus Dr., College Park, MD 20742 USA ,grid.164295.d0000 0001 0941 7177Neuroscience and Cognitive Science Program, University of Maryland, College Park, MD USA
| | - Jane A. Brown
- grid.164295.d0000 0001 0941 7177Department of Psychology, University of Maryland, Biology-Psychology Bldg., 4094 Campus Dr., College Park, MD 20742 USA
| | - Gregory F. Ball
- grid.164295.d0000 0001 0941 7177Department of Psychology, University of Maryland, Biology-Psychology Bldg., 4094 Campus Dr., College Park, MD 20742 USA ,grid.164295.d0000 0001 0941 7177Neuroscience and Cognitive Science Program, University of Maryland, College Park, MD USA
| | - Robert J. Dooling
- grid.164295.d0000 0001 0941 7177Department of Psychology, University of Maryland, Biology-Psychology Bldg., 4094 Campus Dr., College Park, MD 20742 USA ,grid.164295.d0000 0001 0941 7177Neuroscience and Cognitive Science Program, University of Maryland, College Park, MD USA
| |
Collapse
|
21
|
Anikin A, Pisanski K, Reby D. Do nonlinear vocal phenomena signal negative valence or high emotion intensity? ROYAL SOCIETY OPEN SCIENCE 2020; 7:201306. [PMID: 33489278 PMCID: PMC7813245 DOI: 10.1098/rsos.201306] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/23/2020] [Accepted: 11/05/2020] [Indexed: 05/06/2023]
Abstract
Nonlinear vocal phenomena (NLPs) are commonly reported in animal calls and, increasingly, in human vocalizations. These perceptually harsh and chaotic voice features function to attract attention and convey urgency, but they may also signal aversive states. To test whether NLPs enhance the perception of negative affect or only signal high arousal, we added subharmonics, sidebands or deterministic chaos to 48 synthetic human nonverbal vocalizations of ambiguous valence: gasps of fright/surprise, moans of pain/pleasure, roars of frustration/achievement and screams of fear/delight. In playback experiments (N = 900 listeners), we compared their perceived valence and emotion intensity in positive or negative contexts or in the absence of any contextual cues. Primarily, NLPs increased the perceived aversiveness of vocalizations regardless of context. To a smaller extent, they also increased the perceived emotion intensity, particularly when the context was negative or absent. However, NLPs also enhanced the perceived intensity of roars of achievement, indicating that their effects can generalize to positive emotions. In sum, a harsh voice with NLPs strongly tips the balance towards negative emotions when a vocalization is ambiguous, but with sufficiently informative contextual cues, NLPs may be re-evaluated as expressions of intense positive affect, underlining the importance of context in nonverbal communication.
Collapse
Affiliation(s)
- Andrey Anikin
- Division of Cognitive Science, Lund University, Lund, Sweden
- Equipe de Neuro-Ethologie Sensorielle (ENES) / Centre de Recherche en Neurosciences de Lyon (CRNL), University of Lyon/Saint-Etienne, CNRS UMR5292, INSERM UMR_S 1028, Saint-Etienne, France
- Author for correspondence: Andrey Anikin e-mail:
| | - Katarzyna Pisanski
- Equipe de Neuro-Ethologie Sensorielle (ENES) / Centre de Recherche en Neurosciences de Lyon (CRNL), University of Lyon/Saint-Etienne, CNRS UMR5292, INSERM UMR_S 1028, Saint-Etienne, France
| | - David Reby
- Equipe de Neuro-Ethologie Sensorielle (ENES) / Centre de Recherche en Neurosciences de Lyon (CRNL), University of Lyon/Saint-Etienne, CNRS UMR5292, INSERM UMR_S 1028, Saint-Etienne, France
| |
Collapse
|
22
|
Nonverbal auditory communication - Evidence for integrated neural systems for voice signal production and perception. Prog Neurobiol 2020; 199:101948. [PMID: 33189782 DOI: 10.1016/j.pneurobio.2020.101948] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2020] [Revised: 10/12/2020] [Accepted: 11/04/2020] [Indexed: 12/24/2022]
Abstract
While humans have developed a sophisticated and unique system of verbal auditory communication, they also share a more common and evolutionarily important nonverbal channel of voice signaling with many other mammalian and vertebrate species. This nonverbal communication is mediated and modulated by the acoustic properties of a voice signal, and is a powerful - yet often neglected - means of sending and perceiving socially relevant information. From the viewpoint of dyadic (involving a sender and a signal receiver) voice signal communication, we discuss the integrated neural dynamics in primate nonverbal voice signal production and perception. Most previous neurobiological models of voice communication modelled these neural dynamics from the limited perspective of either voice production or perception, largely disregarding the neural and cognitive commonalities of both functions. Taking a dyadic perspective on nonverbal communication, however, it turns out that the neural systems for voice production and perception are surprisingly similar. Based on the interdependence of both production and perception functions in communication, we first propose a re-grouping of the neural mechanisms of communication into auditory, limbic, and paramotor systems, with special consideration for a subsidiary basal-ganglia-centered system. Second, we propose that the similarity in the neural systems involved in voice signal production and perception is the result of the co-evolution of nonverbal voice production and perception systems promoted by their strong interdependence in dyadic interactions.
Collapse
|
23
|
Beaurenaut M, Tokarski E, Dezecache G, Grèzes J. The 'Threat of Scream' paradigm: a tool for studying sustained physiological and subjective anxiety. Sci Rep 2020; 10:12496. [PMID: 32719491 PMCID: PMC7385655 DOI: 10.1038/s41598-020-68889-0] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2019] [Accepted: 07/02/2020] [Indexed: 12/22/2022] Open
Abstract
Progress in understanding the emergence of pathological anxiety depends on the availability of paradigms effective in inducing anxiety in a simple, consistent and sustained manner. The Threat-of-Shock paradigm has typically been used to elicit anxiety, but poses ethical issues when testing vulnerable populations. Moreover, it is not clear from past studies whether anxiety can be sustained in experiments of longer durations. Here, we present empirical support for an alternative approach, the ‘Threat-of-Scream’ paradigm, in which shocks are replaced by screams. In two studies, participants were repeatedly exposed to blocks in which they were at risk of hearing aversive screams at any time vs. blocks in which they were safe from screams. Contrary to previous ‘Threat-of-Scream’ studies, we ensured that our screams were neither harmful nor intolerable by presenting them at low intensity. We found higher subjective reports of anxiety, higher skin conductance levels, and a positive correlation between the two measures, in threat compared to safe blocks. These results were reproducible and we found no significant change over time. The unpredictable delivery of low intensity screams could become an essential part of a psychology toolkit, particularly when investigating the impact of anxiety in a diversity of cognitive functions and populations.
Collapse
Affiliation(s)
- Morgan Beaurenaut
- Laboratoire de Neurosciences Cognitives et Computationnelles, ENS, PSL Research University, INSERM, Département d'études Cognitives, Paris, France.
| | - Elliot Tokarski
- Laboratoire de Neurosciences Cognitives et Computationnelles, ENS, PSL Research University, INSERM, Département d'études Cognitives, Paris, France
| | - Guillaume Dezecache
- Department of Experimental Psychology, Division of Psychology and Language Sciences, University College London, London, UK.,Université Clermont Auvergne, CNRS, LAPSCO, Clermont-Ferrand, France
| | - Julie Grèzes
- Laboratoire de Neurosciences Cognitives et Computationnelles, ENS, PSL Research University, INSERM, Département d'études Cognitives, Paris, France.
| |
Collapse
|
24
|
Abstract
In this article, I outline a relational-developmental conception of emotion that situates emotional activity within a broader conception of persons as holistic, relational beings. In this model, emotions consist of felt forms of engagement with the world. As felt aspects of ongoing action, uninhibited emotional experiences are not private states that are inaccessible to other people; instead, they are revealed directly through their bodily expressions. As multicomponent processes, emotional experiences exhibit both continuity and dramatic change in development. Building on these ideas, I describe an intersubjective methodology for studying developmental changes in the structure of emotional experience. I illustrate the approach with an analysis of developmental changes in the structure of anger from birth to adulthood.
Collapse
|
25
|
Clink DJ, Ahmad AH, Klinck H. Brevity is not a universal in animal communication: evidence for compression depends on the unit of analysis in small ape vocalizations. ROYAL SOCIETY OPEN SCIENCE 2020; 7:200151. [PMID: 32431905 PMCID: PMC7211885 DOI: 10.1098/rsos.200151] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/25/2020] [Accepted: 03/10/2020] [Indexed: 05/04/2023]
Abstract
Evidence for compression, or minimization of code length, has been found across biological systems from genomes to human language and music. Two linguistic laws-Menzerath's Law (which states that longer sequences consist of shorter constituents) and Zipf's Law of abbreviation (a negative relationship between signal length and frequency of use)-are predictions of compression. It has been proposed that compression is a universal in animal communication, but there have been mixed results, particularly in reference to Zipf's Law of abbreviation. Like songbirds, male gibbons (Hylobates muelleri) engage in long solo bouts with unique combinations of notes which combine into phrases. We found strong support for Menzerath's Law as the longer a phrase, the shorter the notes. To identify phrase types, we used state-of-the-art affinity propagation clustering, and were able to predict phrase types using support vector machines with a mean accuracy of 74%. Based on unsupervised phrase type classification, we did not find support for Zipf's Law of abbreviation. Our results indicate that adherence to linguistic laws in male gibbon solos depends on the unit of analysis. We conclude that principles of compression are applicable outside of human language, but may act differently across levels of organization in biological systems.
Collapse
Affiliation(s)
- Dena J. Clink
- Center for Conservation Bioacoustics, Cornell Laboratory of Ornithology, Cornell University, Ithaca, NY, USA
| | - Abdul Hamid Ahmad
- Faculty of Sustainable Agriculture, Universiti Malaysia Sabah, Sandakan Campus, Sabah, Malaysia
| | - Holger Klinck
- Center for Conservation Bioacoustics, Cornell Laboratory of Ornithology, Cornell University, Ithaca, NY, USA
| |
Collapse
|
26
|
Pisanski K, Raine J, Reby D. Individual differences in human voice pitch are preserved from speech to screams, roars and pain cries. ROYAL SOCIETY OPEN SCIENCE 2020; 7:191642. [PMID: 32257325 PMCID: PMC7062086 DOI: 10.1098/rsos.191642] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/18/2019] [Accepted: 01/21/2020] [Indexed: 05/19/2023]
Abstract
Fundamental frequency (F0, perceived as voice pitch) predicts sex and age, hormonal status, mating success and a range of social traits, and thus functions as an important biosocial marker in modal speech. Yet, the role of F0 in human nonverbal vocalizations remains unclear, and given considerable variability in F0 across call types, it is not known whether F0 cues to vocalizer attributes are shared across speech and nonverbal vocalizations. Here, using a corpus of vocal sounds from 51 men and women, we examined whether individual differences in F0 are retained across neutral speech, valenced speech and nonverbal vocalizations (screams, roars and pain cries). Acoustic analyses revealed substantial variability in F0 across vocal types, with mean F0 increasing as much as 10-fold in screams compared to speech in the same individual. Despite these extreme pitch differences, sexual dimorphism was preserved within call types and, critically, inter-individual differences in F0 correlated across vocal types (r = 0.36-0.80) with stronger relationships between vocal types of the same valence (e.g. 38% of the variance in roar F0 was predicted by aggressive speech F0). Our results indicate that biologically and socially relevant indexical cues in the human voice are preserved in simulated valenced speech and vocalizations, including vocalizations characterized by extreme F0 modulation, suggesting that voice pitch may function as a reliable individual and biosocial marker across disparate communication contexts.
Collapse
Affiliation(s)
- Katarzyna Pisanski
- Equipe de Neuro-Ethologie Sensorielle ENES/CRNL, University of Lyon/Saint-Etienne, CNRS UMR5292, INSERM UMR_S 1028, Saint-Etienne, France
- Author for correspondence: Katarzyna Pisanski e-mail:
| | - Jordan Raine
- Mammal Vocal Communication and Cognition Research Group, School of Psychology, University of Sussex, Brighton, UK
| | - David Reby
- Equipe de Neuro-Ethologie Sensorielle ENES/CRNL, University of Lyon/Saint-Etienne, CNRS UMR5292, INSERM UMR_S 1028, Saint-Etienne, France
- Mammal Vocal Communication and Cognition Research Group, School of Psychology, University of Sussex, Brighton, UK
| |
Collapse
|
27
|
Anikin A. A Moan of Pleasure Should Be Breathy: The Effect of Voice Quality on the Meaning of Human Nonverbal Vocalizations. PHONETICA 2020; 77:327-349. [PMID: 31962309 PMCID: PMC7592904 DOI: 10.1159/000504855] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/17/2019] [Accepted: 11/15/2019] [Indexed: 05/19/2023]
Abstract
Prosodic features, such as intonation and voice intensity, have a well-documented role in communicating emotion, but less is known about the role of laryngeal voice quality in speech and particularly in nonverbal vocalizations such as laughs and moans. Potentially, however, variations in voice quality between tense and breathy may convey rich information about the speaker's physiological and affective state. In this study breathiness was manipulated in synthetic human nonverbal vocalizations by adjusting the relative strength of upper harmonics and aspiration noise. In experiment 1 (28 prototypes × 3 manipulations = 84 sounds), otherwise identical vocalizations with tense versus breathy voice quality were associated with higher arousal (general alertness), higher dominance, and lower valence (unpleasant states). Ratings on discrete emotions in experiment 2 (56 × 3 = 168 sounds) confirmed that breathiness was reliably associated with positive emotions, particularly in ambiguous vocalizations (gasps and moans). The spectral centroid did not fully account for the effect of manipulation, confirming that the perceived change in voice quality was more specific than a general shift in timbral brightness. Breathiness is thus involved in communicating emotion with nonverbal vocalizations, possibly due to changes in low-level auditory salience and perceived vocal effort.
Collapse
|
28
|
Was That a Scream? Listener Agreement and Major Distinguishing Acoustic Features. JOURNAL OF NONVERBAL BEHAVIOR 2019. [DOI: 10.1007/s10919-019-00325-y] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/30/2022]
|
29
|
Demartsev V, Gordon N, Barocas A, Bar-Ziv E, Ilany T, Goll Y, Ilany A, Geffen E. The "Law of Brevity" in animal communication: Sex-specific signaling optimization is determined by call amplitude rather than duration. Evol Lett 2019; 3:623-634. [PMID: 31867122 PMCID: PMC6906988 DOI: 10.1002/evl3.147] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2019] [Revised: 10/10/2019] [Accepted: 10/22/2019] [Indexed: 11/25/2022] Open
Abstract
The efficiency of informational transfer is one of the key aspects of any communication system. The informational coding economy of human languages is often demonstrated by their almost universal fit to Zipf's “Law of Brevity,” expressing negative relationship between word length and its usage frequency. Animal vocal systems, however, provided mixed results in their adherence to this relationship, potentially due to conflicting evolutionary pressures related to differences in signaling range and communicational needs. To examine this potential parallel between human and animal vocal communication, and also to explore how divergent, sex‐specific, communicational settings affect signaling efficiency within a species, we examined the complete vocal repertoire of rock hyraxes (Procavia capensis). As male and female hyraxes differ in their sociality levels and male hyraxes vocal repertoire is dominated by sexual advertisement songs, we hypothesized that sex‐specific vocal repertoires could be subjected to different signaling optimization pressures. Our results show that the sexes differ in repertoire size, call usage, and adherence to coding efficiency principles. Interestingly, the classic call length/call usage relationship is not consistently found in rock hyraxes. Rather, a negative relationship between call amplitude and call usage is found, suggesting that the efficiency of the vocal repertoire is driven by call amplitude rather than duration. We hypothesize that, in contrast to human speech that is mainly intended for short distance, the need for frequent long‐range signaling shapes an animal's vocal repertoire efficiency according to the cost of call amplitude rather than call length. However, call duration may be a secondary factor affecting signaling efficiency, in cases where amplitude is under specific selection pressures, such as sexual selection.
Collapse
Affiliation(s)
- Vlad Demartsev
- Department of Biology University of Konstanz Konstanz 78464 Germany.,School of Zoology Tel Aviv University Tel Aviv 6997801 Israel
| | - Naomi Gordon
- School of Zoology Tel Aviv University Tel Aviv 6997801 Israel
| | - Adi Barocas
- San Diego Zoo's Institute for Conservation Research Escondido California 92027.,Wildlife Conservation Research Unit Department of Zoology University of Oxford Abingdon OX13 5QL United Kingdom
| | - Einat Bar-Ziv
- Mitrani Department of Desert Ecology Ben-Gurion University Midreshet Ben-Gurion 8499000 Israel
| | | | - Yael Goll
- School of Zoology Tel Aviv University Tel Aviv 6997801 Israel
| | - Amiyaal Ilany
- Faculty of Life Sciences Bar-Ilan University Ramat-Gan 5290002 Israel
| | - Eli Geffen
- School of Zoology Tel Aviv University Tel Aviv 6997801 Israel
| |
Collapse
|
30
|
Engelberg JWM, Schwartz JW, Gouzoules H. Do human screams permit individual recognition? PeerJ 2019; 7:e7087. [PMID: 31275746 PMCID: PMC6596410 DOI: 10.7717/peerj.7087] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2018] [Accepted: 05/07/2019] [Indexed: 11/20/2022] Open
Abstract
The recognition of individuals through vocalizations is a highly adaptive ability in the social behavior of many species, including humans. However, the extent to which nonlinguistic vocalizations such as screams permit individual recognition in humans remains unclear. Using a same-different vocalizer discrimination task, we investigated participants' ability to correctly identify whether pairs of screams were produced by the same person or two different people, a critical prerequisite to individual recognition. Despite prior theory-based contentions that screams are not acoustically well-suited to conveying identity cues, listeners discriminated individuals at above-chance levels by their screams, including both acoustically modified and unmodified exemplars. We found that vocalizer gender explained some variation in participants' discrimination abilities and response times, but participant attributes (gender, experience, empathy) did not. Our findings are consistent with abundant evidence from nonhuman primates, suggesting that both human and nonhuman screams convey cues to caller identity, thus supporting the thesis of evolutionary continuity in at least some aspects of scream function across primate species.
Collapse
Affiliation(s)
| | - Jay W Schwartz
- Department of Psychology, Emory University, Atlanta, GA, USA
| | | |
Collapse
|
31
|
Abstract
Voice synthesis is a useful method for investigating the communicative role of different acoustic features. Although many text-to-speech systems are available, researchers of human nonverbal vocalizations and bioacousticians may profit from a dedicated simple tool for synthesizing and manipulating natural-sounding vocalizations. Soundgen ( https://CRAN.R-project.org/package=soundgen ) is an open-source R package that synthesizes nonverbal vocalizations based on meaningful acoustic parameters, which can be specified from the command line or in an interactive app. This tool was validated by comparing the perceived emotion, valence, arousal, and authenticity of 60 recorded human nonverbal vocalizations (screams, moans, laughs, and so on) and their approximate synthetic reproductions. Each synthetic sound was created by manually specifying only a small number of high-level control parameters, such as syllable length and a few anchors for the intonation contour. Nevertheless, the valence and arousal ratings of synthetic sounds were similar to those of the original recordings, and the authenticity ratings were comparable, maintaining parity with the originals for less complex vocalizations. Manipulating the precise acoustic characteristics of synthetic sounds may shed light on the salient predictors of emotion in the human voice. More generally, soundgen may prove useful for any studies that require precise control over the acoustic features of nonspeech sounds, including research on animal vocalizations and auditory perception.
Collapse
Affiliation(s)
- Andrey Anikin
- Division of Cognitive Science, Department of Philosophy, Lund University, Box 192, SE-221 00, Lund, Sweden.
| |
Collapse
|
32
|
Anikin A. The perceptual effects of manipulating nonlinear phenomena in synthetic nonverbal vocalizations. BIOACOUSTICS 2019. [DOI: 10.1080/09524622.2019.1581839] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
Affiliation(s)
- Andrey Anikin
- Division of Cognitive Science, Department of Philosophy, Lund University, Lund, Sweden
| |
Collapse
|
33
|
Engelberg JW, Gouzoules H. The credibility of acted screams: Implications for emotional communication research. Q J Exp Psychol (Hove) 2018; 72:1889-1902. [PMID: 30514163 DOI: 10.1177/1747021818816307] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Researchers have long relied on acted material to study emotional expression and perception in humans. It has been suggested, however, that certain aspects of natural expressions are difficult or impossible to produce voluntarily outside of their associated emotional contexts, and that acted expressions tend to be overly intense caricatures. From an evolutionary perspective, listeners' abilities to distinguish acted from natural expressions likely depend on the type of expression in question, the costs entailed in its production, and elements of receiver psychology. Here, we investigated these issues as they relate to human screams. We also examined whether listeners' abilities to distinguish acted from natural screams might vary as a function of individual differences in emotional processing and empathy. Using a forced-choice categorization task, we found that listeners could not distinguish acted from natural exemplars, suggesting that actors can produce dramatisations of screams resembling natural vocalisations. Intensity ratings did not differ between acted and natural screams, nor did individual differences in emotional processing significantly predict performance. Scream duration predicted both the probability that an exemplar was categorised as acted and the probability that participants classified that scream accurately. These findings are discussed with respect to potential evolutionary implications and their practical relevance to future research using acted screams.
Collapse
|
34
|
Stomp M, Leroux M, Cellier M, Henry S, Lemasson A, Hausberger M. An unexpected acoustic indicator of positive emotions in horses. PLoS One 2018; 13:e0197898. [PMID: 29995876 PMCID: PMC6040684 DOI: 10.1371/journal.pone.0197898] [Citation(s) in RCA: 31] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2018] [Accepted: 05/10/2018] [Indexed: 11/23/2022] Open
Abstract
Indicators of positive emotions are still scarce and many proposed behavioural markers have proven ambiguous. Studies established a link between acoustic signals and emitter’s internal state, but few related to positive emotions and still fewer considered non-vocal sounds. One of them, the snort, is shared by several perrisodactyls and has been associated to positive contexts in these species. We hypothesized that this could be also the case in horses. In this species, there is a clear need for a thorough description of non-vocal acoustic signals (snorts, snores or blows are often used interchangeably) but overall this sound produced by nostrils during expiration has up to now been mostly considered as having a hygienic function. However, observations revealed that snorts were produced more in some individuals than in others, without relationship with air conditions. We observed 48 horses living in two “extreme” conditions: restricted conditions (single stall, low roughage diet) and naturalistic conditions (stable groups in pasture). The immediate place (e.g. stall/pasture) and the behavioural/postural (behaviour performed/ears positions) contexts of snort production were observed. We additionally performed an evaluation of the welfare state, using validated behavioural (e.g. stereotypies) and postural (e.g. overall ears positions) welfare indicators. The results show that 1) snort production was significantly associated with situations known to be positive for horses (e.g. feeding in pasture) and with a positive internal state (ears in forward or sidewards positions), 2) the riding school horses produced twice as many snorts when in pasture than in stall, 3) the naturalistic population emitted significantly more snorts than riding school ones in comparable contexts, 4) the frequency of snorts was negatively correlated with the composite total chronic stress score (TCSS, reflecting compromised welfare based on the horse’s rank on the different indicators): the lower the TCSS, the higher the snort rate. Snorts therefore appear as reliable indicators of positive emotions.
Collapse
Affiliation(s)
- Mathilde Stomp
- Université de Rennes 1, CNRS, UMR 6552 CNRS Ethologie Animale et Humaine, Université de Caen-Normandie, Station Biologique de Paimpont, Paimpont, France
- * E-mail:
| | - Maël Leroux
- Université de Rennes 1, CNRS, UMR 6552 CNRS Ethologie Animale et Humaine, Université de Caen-Normandie, Station Biologique de Paimpont, Paimpont, France
| | - Marjorie Cellier
- Université de Rennes 1, CNRS, UMR 6552 CNRS Ethologie Animale et Humaine, Université de Caen-Normandie, Station Biologique de Paimpont, Paimpont, France
| | - Séverine Henry
- Université de Rennes 1, CNRS, UMR 6552 CNRS Ethologie Animale et Humaine, Université de Caen-Normandie, Station Biologique de Paimpont, Paimpont, France
| | - Alban Lemasson
- Université de Rennes 1, CNRS, UMR 6552 CNRS Ethologie Animale et Humaine, Université de Caen-Normandie, Station Biologique de Paimpont, Paimpont, France
| | - Martine Hausberger
- CNRS, UMR 6552 Ethologie animale et humaine, Université de Rennes 1, Université de Caen-Normandie, Rennes, France
| |
Collapse
|