1
|
Holler J. Facial clues to conversational intentions. Trends Cogn Sci 2025:S1364-6613(25)00079-8. [PMID: 40345945 DOI: 10.1016/j.tics.2025.03.006] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/17/2025] [Revised: 03/27/2025] [Accepted: 03/31/2025] [Indexed: 05/11/2025]
Abstract
It has long been known that we use words to perform speech acts foundational to everyday conversation, such as requesting, informing, proposing, or complaining. However, the natural environment of human language is face-to-face interaction where we use words and an abundance of visual signals to communicate. The multimodal nature of human language is increasingly recognised in the language and cognitive sciences. In line with this turn of the tide, findings demonstrate that facial signals significantly contribute to communicating intentions and that they may facilitate pragmatically appropriate responding in the fast-paced environment of conversation. In light of this, the notion of speech acts no longer seems appropriate, highlighting the need for a modality-neutral conception, such as social action.
Collapse
Affiliation(s)
- Judith Holler
- Donders Institute for Brain, Cognition & Behaviour, Radboud University Max Planck Institute for Psycholinguistics, Nijmegen, Netherlands.
| |
Collapse
|
2
|
Hömke P, Levinson SC, Emmendorfer AK, Holler J. Eyebrow movements as signals of communicative problems in human face-to-face interaction. ROYAL SOCIETY OPEN SCIENCE 2025; 12:241632. [PMID: 40078914 PMCID: PMC11896710 DOI: 10.1098/rsos.241632] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/02/2024] [Accepted: 12/03/2024] [Indexed: 03/14/2025]
Abstract
Repair is a core building block of human communication, allowing us to address problems of understanding in conversation. Past research has uncovered the basic mechanisms by which interactants signal and solve such problems. However, the focus has been on verbal interaction, neglecting the fact that human communication is inherently multimodal. Here, we focus on a visual signal particularly prevalent in signalling problems of understanding: eyebrow furrows and raises. We present, first, a corpus study showing that differences in eyebrow actions (furrows versus raises) were systematically associated with differences in the format of verbal repair initiations. Second, we present a follow-up study using an avatar that allowed us to test the causal consequences of addressee eyebrow movements, zooming into the effect of eyebrow furrows as signals of trouble in understanding in particular. The results revealed that addressees' eyebrow furrows have a striking effect on speakers' speech, leading speakers to produce answers to questions several seconds longer than when not perceiving addressee eyebrow furrows while speaking. Together, the findings demonstrate that eyebrow movements play a communicative role in initiating repair during conversation rather than being merely epiphenomenal and that their occurrence can critically influence linguistic behaviour. Thus, eyebrow movements should be considered core coordination devices in human conversational interaction.
Collapse
Affiliation(s)
- Paul Hömke
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands
| | | | - Alexandra K. Emmendorfer
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands
| | - Judith Holler
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands
| |
Collapse
|
3
|
Trujillo JP, Dyer RMK, Holler J. Dyadic differences in empathy scores are associated with kinematic similarity during conversational question-answer pairs. DISCOURSE PROCESSES 2025; 62:195-213. [PMID: 40303465 PMCID: PMC12039893 DOI: 10.1080/0163853x.2025.2467605] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/02/2025]
Abstract
During conversation, speakers coordinate and synergize their behaviors at multiple levels, and in different ways. The extent to which individuals converge or diverge in their behaviors during interaction may relate to interpersonal differences relevant to social interaction, such as empathy as measured by the empathy quotient (EQ). An association between interpersonal difference in empathy and interpersonal entrainment could help to throw light on how interlocutor characteristics influence interpersonal entrainment. We investigated this possibility in a corpus of unconstrained conversation between dyads. We used dynamic time warping to quantify entrainment between interlocutors of head motion, hand motion, and maximum speech f0 during question-response sequences. We additionally calculated interlocutor differences in EQ scores. We found that, for both head and hand motion, greater difference in EQ was associated with higher entrainment. Thus, we consider that people who are dissimilar in EQ may need to "ground" their interaction with low-level movement entrainment. There was no significant relationship between f0 entrainment and EQ score differences.
Collapse
Affiliation(s)
- James P. Trujillo
- Donders Institute for Brain, Cognition & Behaviour, Radboud University, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- Institute for Logic, Language & Computation, University of Amsterdam, Amsterdam, The Netherlands
| | - Rebecca M. K. Dyer
- Institute for Logic, Language & Computation, University of Amsterdam, Amsterdam, The Netherlands
- Centre for Human Developmental Science, Cardiff University, Cardiff, UK
| | - Judith Holler
- Donders Institute for Brain, Cognition & Behaviour, Radboud University, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| |
Collapse
|
4
|
Emmendorfer AK, Holler J. Facial signals shape predictions about the nature of upcoming conversational responses. Sci Rep 2025; 15:1381. [PMID: 39779723 PMCID: PMC11711643 DOI: 10.1038/s41598-025-85192-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2024] [Accepted: 01/01/2025] [Indexed: 01/11/2025] Open
Abstract
Increasing evidence suggests that interlocutors use visual communicative signals to form predictions about unfolding utterances, but there is little data on the predictive potential of facial signals in conversation. In an online experiment with virtual agents, we examine whether facial signals produced by an addressee may allow speakers to anticipate the response to a question before it is given. Participants (n = 80) viewed videos of short conversation fragments between two virtual humans. Each fragment ended with the Questioner asking a question, followed by a pause during which the Responder looked either straight at the Questioner (baseline), or averted their gaze, or accompanied the straight gaze with one of the following facial signals: brow raise, brow frown, nose wrinkle, smile, squint, mouth corner pulled back (dimpler). Participants then indicated on a 6-point scale whether they expected a "yes" or "no" response. Analyses revealed that all signals received different ratings relative to the baseline: brow raises, dimplers, and smiles were associated with more positive responses, gaze aversions, brow frowns, nose wrinkles, and squints with more negative responses. Qur findings show that interlocutors may form strong associations between facial signals and upcoming responses to questions, highlighting their predictive potential in face-to-face conversation.
Collapse
Affiliation(s)
- Alexandra K Emmendorfer
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Wundtlaan 1, 6525 XD Nijmegen, Nijmegen, The Netherlands.
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands.
| | - Judith Holler
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Wundtlaan 1, 6525 XD Nijmegen, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| |
Collapse
|
5
|
Rasing NB, van de Geest-Buit W, Chan OYA, Mul K, Lanser A, Erasmus CE, Groothuis JT, Holler J, Ingels KJAO, Post B, Siemann I, Voermans NC. Psychosocial functioning in patients with altered facial expression: a scoping review in five neurological diseases. Disabil Rehabil 2024; 46:3772-3791. [PMID: 37752723 DOI: 10.1080/09638288.2023.2259310] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/19/2022] [Revised: 09/06/2023] [Accepted: 09/10/2023] [Indexed: 09/28/2023]
Abstract
PURPOSE To perform a scoping review to investigate the psychosocial impact of having an altered facial expression in five neurological diseases. METHODS A systematic literature search was performed. Studies were on Bell's palsy, facioscapulohumeral muscular dystrophy (FSHD), Moebius syndrome, myotonic dystrophy type 1, or Parkinson's disease patients; had a focus on altered facial expression; and had any form of psychosocial outcome measure. Data extraction focused on psychosocial outcomes. RESULTS Bell's palsy, myotonic dystrophy type 1, and Parkinson's disease patients more often experienced some degree of psychosocial distress than healthy controls. In FSHD, facial weakness negatively influenced communication and was experienced as a burden. The psychosocial distress applied especially to women (Bell's palsy and Parkinson's disease), and patients with more severely altered facial expression (Bell's palsy), but not for Moebius syndrome patients. Furthermore, Parkinson's disease patients with more pronounced hypomimia were perceived more negatively by observers. Various strategies were reported to compensate for altered facial expression. CONCLUSIONS This review showed that patients with altered facial expression in four of five included neurological diseases had reduced psychosocial functioning. Future research recommendations include studies on observers' judgements of patients during social interactions and on the effectiveness of compensation strategies in enhancing psychosocial functioning.
Collapse
Affiliation(s)
- Nathaniël B Rasing
- Department of Neurology, Donders Institute for Brain, Cognition and Behaviour, Radboud University Medical Center, Nijmegen, The Netherlands
| | - Willianne van de Geest-Buit
- Department of Neurology, Donders Institute for Brain, Cognition and Behaviour, Radboud University Medical Center, Nijmegen, The Netherlands
| | - On Ying A Chan
- Medical Library, Radboud University, Nijmegen, The Netherlands
| | - Karlien Mul
- Department of Neurology, Donders Institute for Brain, Cognition and Behaviour, Radboud University Medical Center, Nijmegen, The Netherlands
| | - Anke Lanser
- Patient Representative and Chairman FSHD Advocacy Group, Patient Organization for Muscular Disease Spierziekten Nederland, Baarn, The Netherlands
| | - Corrie E Erasmus
- Department of Pediatric Neurology, Radboud University Medical Center, Nijmegen, The Netherlands
| | - Jan T Groothuis
- Department of Rehabilitation, Donders Institute for Brain, Cognition and Behaviour, Radboud University Medical Center, Nijmegen, The Netherlands
| | - Judith Holler
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, and Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - Koen J A O Ingels
- Department of Otorhinolaryngology, Radboud University Medical Center, Nijmegen, The Netherlands
| | - Bart Post
- Department of Neurology, Donders Institute for Brain, Cognition and Behaviour, Radboud University Medical Center, Nijmegen, The Netherlands
| | - Ietske Siemann
- Department of Medical Psychology, Radboud Institute for Health Sciences, Radboud University Medical Center, Nijmegen, The Netherlands
| | - Nicol C Voermans
- Department of Neurology, Donders Institute for Brain, Cognition and Behaviour, Radboud University Medical Center, Nijmegen, The Netherlands
| |
Collapse
|
6
|
de Vries C, Andries F, Meissl K. Mocking enactments: a case study of multimodal stance-stacking. Front Psychol 2024; 15:1379593. [PMID: 38629031 PMCID: PMC11019027 DOI: 10.3389/fpsyg.2024.1379593] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2024] [Accepted: 02/26/2024] [Indexed: 04/19/2024] Open
Abstract
Although research into multimodal stance-taking has gained momentum over the past years, the multimodal construction of so-called stacked stances has not yet received systematic attention in the literature. Mocking enactments are a prime example of such complex social actions as they are layered both interactionally and stance-related, and they rely significantly on the use of bodily visual resources, depicting rather than describing events and stances. Using Du Bois' Stance Triangle as a framework, this study investigates mocking enactments as a case study to unravel the multimodal aspects of layered stance expressions. Drawing on three data sets-music instruction in Dutch, German, and English, spontaneous face-to-face interactions among friends in Dutch, and narrations on past events in Flemish Sign Language (VGT)-this study provides a qualitative exploration of mocking enactments across different communicative settings, languages, and modalities. The study achieves three main objectives: (1) illuminating how enactments are used for mocking, (2) identifying the layers of stance-taking at play, and (3) examining the multimodal construction of mocking enactments. Our analysis reveals various different uses of enactments for mocking. Aside from enacting the target of the mockery, participants can include other characters and viewpoints, highlighting the breadth of the phenomenon under scrutiny. Second, we uncover the layered construction of stance on all axes of the Stance Triangle (evaluation, positioning, and alignment). Third, we find that mocking enactments are embedded in highly evaluative contexts, indexed by the use of bodily visual resources. Interestingly, not all mocking enactments include a multimodally exaggerated depiction, but instead, some merely allude to an absurd hypothetical scenario. Our findings contribute to the growing body of literature on multimodal stance-taking, by showing how a nuanced interpretation of the Stance Triangle can offer a useful framework for analyzing layered stance acts.
Collapse
Affiliation(s)
- Clarissa de Vries
- Department of Linguistics, Faculty of Arts, KU Leuven, Leuven, Belgium
| | - Fien Andries
- Department of Linguistics, Faculty of Arts, KU Leuven, Antwerp, Belgium
| | - Katharina Meissl
- Department of Linguistics, Faculty of Arts, KU Leuven, Leuven, Belgium
| |
Collapse
|
7
|
Trujillo JP, Holler J. Conversational facial signals combine into compositional meanings that change the interpretation of speaker intentions. Sci Rep 2024; 14:2286. [PMID: 38280963 PMCID: PMC10821935 DOI: 10.1038/s41598-024-52589-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2023] [Accepted: 01/20/2024] [Indexed: 01/29/2024] Open
Abstract
Human language is extremely versatile, combining a limited set of signals in an unlimited number of ways. However, it is unknown whether conversational visual signals feed into the composite utterances with which speakers communicate their intentions. We assessed whether different combinations of visual signals lead to different intent interpretations of the same spoken utterance. Participants viewed a virtual avatar uttering spoken questions while producing single visual signals (i.e., head turn, head tilt, eyebrow raise) or combinations of these signals. After each video, participants classified the communicative intention behind the question. We found that composite utterances combining several visual signals conveyed different meaning compared to utterances accompanied by the single visual signals. However, responses to combinations of signals were more similar to the responses to related, rather than unrelated, individual signals, indicating a consistent influence of the individual visual signals on the whole. This study therefore provides first evidence for compositional, non-additive (i.e., Gestalt-like) perception of multimodal language.
Collapse
Affiliation(s)
- James P Trujillo
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands.
- Donders Institute for Brain, Cognition, and Behaviour, Nijmegen, The Netherlands.
| | - Judith Holler
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- Donders Institute for Brain, Cognition, and Behaviour, Nijmegen, The Netherlands
| |
Collapse
|
8
|
Ter Bekke M, Drijvers L, Holler J. Hand Gestures Have Predictive Potential During Conversation: An Investigation of the Timing of Gestures in Relation to Speech. Cogn Sci 2024; 48:e13407. [PMID: 38279899 DOI: 10.1111/cogs.13407] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2022] [Revised: 07/09/2023] [Accepted: 01/10/2024] [Indexed: 01/29/2024]
Abstract
During face-to-face conversation, transitions between speaker turns are incredibly fast. These fast turn exchanges seem to involve next speakers predicting upcoming semantic information, such that next turn planning can begin before a current turn is complete. Given that face-to-face conversation also involves the use of communicative bodily signals, an important question is how bodily signals such as co-speech hand gestures play into these processes of prediction and fast responding. In this corpus study, we found that hand gestures that depict or refer to semantic information started before the corresponding information in speech, which held both for the onset of the gesture as a whole, as well as the onset of the stroke (the most meaningful part of the gesture). This early timing potentially allows listeners to use the gestural information to predict the corresponding semantic information to be conveyed in speech. Moreover, we provided further evidence that questions with gestures got faster responses than questions without gestures. However, we found no evidence for the idea that how much a gesture precedes its lexical affiliate (i.e., its predictive potential) relates to how fast responses were given. The findings presented here highlight the importance of the temporal relation between speech and gesture and help to illuminate the potential mechanisms underpinning multimodal language processing during face-to-face conversation.
Collapse
Affiliation(s)
- Marlijn Ter Bekke
- Donders Institute for Brain, Cognition and Behaviour, Radboud University
- Max Planck Institute for Psycholinguistics
| | - Linda Drijvers
- Donders Institute for Brain, Cognition and Behaviour, Radboud University
- Max Planck Institute for Psycholinguistics
| | - Judith Holler
- Donders Institute for Brain, Cognition and Behaviour, Radboud University
- Max Planck Institute for Psycholinguistics
| |
Collapse
|
9
|
Nota N, Trujillo JP, Jacobs V, Holler J. Facilitating question identification through natural intensity eyebrow movements in virtual avatars. Sci Rep 2023; 13:21295. [PMID: 38042876 PMCID: PMC10693605 DOI: 10.1038/s41598-023-48586-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2022] [Accepted: 11/28/2023] [Indexed: 12/04/2023] Open
Abstract
In conversation, recognizing social actions (similar to 'speech acts') early is important to quickly understand the speaker's intended message and to provide a fast response. Fast turns are typical for fundamental social actions like questions, since a long gap can indicate a dispreferred response. In multimodal face-to-face interaction, visual signals may contribute to this fast dynamic. The face is an important source of visual signalling, and previous research found that prevalent facial signals such as eyebrow movements facilitate the rapid recognition of questions. We aimed to investigate whether early eyebrow movements with natural movement intensities facilitate question identification, and whether specific intensities are more helpful in detecting questions. Participants were instructed to view videos of avatars where the presence of eyebrow movements (eyebrow frown or raise vs. no eyebrow movement) was manipulated, and to indicate whether the utterance in the video was a question or statement. Results showed higher accuracies for questions with eyebrow frowns, and faster response times for questions with eyebrow frowns and eyebrow raises. No additional effect was observed for the specific movement intensity. This suggests that eyebrow movements that are representative of naturalistic multimodal behaviour facilitate question recognition.
Collapse
Affiliation(s)
- Naomi Nota
- Donders Institute for Brain, Cognition, and Behaviour, Nijmegen, The Netherlands.
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands.
| | - James P Trujillo
- Donders Institute for Brain, Cognition, and Behaviour, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - Vere Jacobs
- Faculty of Arts, Radboud University, Nijmegen, The Netherlands
| | - Judith Holler
- Donders Institute for Brain, Cognition, and Behaviour, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| |
Collapse
|
10
|
Nota N, Trujillo JP, Holler J. Conversational Eyebrow Frowns Facilitate Question Identification: An Online Study Using Virtual Avatars. Cogn Sci 2023; 47:e13392. [PMID: 38058215 DOI: 10.1111/cogs.13392] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2022] [Revised: 10/11/2023] [Accepted: 11/27/2023] [Indexed: 12/08/2023]
Abstract
Conversation is a time-pressured environment. Recognizing a social action (the ''speech act,'' such as a question requesting information) early is crucial in conversation to quickly understand the intended message and plan a timely response. Fast turns between interlocutors are especially relevant for responses to questions since a long gap may be meaningful by itself. Human language is multimodal, involving speech as well as visual signals from the body, including the face. But little is known about how conversational facial signals contribute to the communication of social actions. Some of the most prominent facial signals in conversation are eyebrow movements. Previous studies found links between eyebrow movements and questions, suggesting that these facial signals could contribute to the rapid recognition of questions. Therefore, we aimed to investigate whether early eyebrow movements (eyebrow frown or raise vs. no eyebrow movement) facilitate question identification. Participants were instructed to view videos of avatars where the presence of eyebrow movements accompanying questions was manipulated. Their task was to indicate whether the utterance was a question or a statement as accurately and quickly as possible. Data were collected using the online testing platform Gorilla. Results showed higher accuracies and faster response times for questions with eyebrow frowns, suggesting a facilitative role of eyebrow frowns for question identification. This means that facial signals can critically contribute to the communication of social actions in conversation by signaling social action-specific visual information and providing visual cues to speakers' intentions.
Collapse
Affiliation(s)
- Naomi Nota
- Donders Institute for Brain, Cognition, and Behaviour, Nijmegen
- Max Planck Institute for Psycholinguistics, Nijmegen
| | - James P Trujillo
- Donders Institute for Brain, Cognition, and Behaviour, Nijmegen
- Max Planck Institute for Psycholinguistics, Nijmegen
| | - Judith Holler
- Donders Institute for Brain, Cognition, and Behaviour, Nijmegen
- Max Planck Institute for Psycholinguistics, Nijmegen
| |
Collapse
|
11
|
Trujillo JP, Holler J. Interactionally Embedded Gestalt Principles of Multimodal Human Communication. PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2023; 18:1136-1159. [PMID: 36634318 PMCID: PMC10475215 DOI: 10.1177/17456916221141422] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/14/2023]
Abstract
Natural human interaction requires us to produce and process many different signals, including speech, hand and head gestures, and facial expressions. These communicative signals, which occur in a variety of temporal relations with each other (e.g., parallel or temporally misaligned), must be rapidly processed as a coherent message by the receiver. In this contribution, we introduce the notion of interactionally embedded, affordance-driven gestalt perception as a framework that can explain how this rapid processing of multimodal signals is achieved as efficiently as it is. We discuss empirical evidence showing how basic principles of gestalt perception can explain some aspects of unimodal phenomena such as verbal language processing and visual scene perception but require additional features to explain multimodal human communication. We propose a framework in which high-level gestalt predictions are continuously updated by incoming sensory input, such as unfolding speech and visual signals. We outline the constituent processes that shape high-level gestalt perception and their role in perceiving relevance and prägnanz. Finally, we provide testable predictions that arise from this multimodal interactionally embedded gestalt-perception framework. This review and framework therefore provide a theoretically motivated account of how we may understand the highly complex, multimodal behaviors inherent in natural social interaction.
Collapse
Affiliation(s)
- James P. Trujillo
- Donders Institute for Brain, Cognition, and Behaviour, Nijmegen, the Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, the Netherlands
| | - Judith Holler
- Donders Institute for Brain, Cognition, and Behaviour, Nijmegen, the Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, the Netherlands
| |
Collapse
|
12
|
Nota N, Trujillo JP, Holler J. Specific facial signals associate with categories of social actions conveyed through questions. PLoS One 2023; 18:e0288104. [PMID: 37467253 DOI: 10.1371/journal.pone.0288104] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2022] [Accepted: 06/20/2023] [Indexed: 07/21/2023] Open
Abstract
The early recognition of fundamental social actions, like questions, is crucial for understanding the speaker's intended message and planning a timely response in conversation. Questions themselves may express more than one social action category (e.g., an information request "What time is it?", an invitation "Will you come to my party?" or a criticism "Are you crazy?"). Although human language use occurs predominantly in a multimodal context, prior research on social actions has mainly focused on the verbal modality. This study breaks new ground by investigating how conversational facial signals may map onto the expression of different types of social actions conveyed through questions. The distribution, timing, and temporal organization of facial signals across social actions was analysed in a rich corpus of naturalistic, dyadic face-to-face Dutch conversations. These social actions were: Information Requests, Understanding Checks, Self-Directed questions, Stance or Sentiment questions, Other-Initiated Repairs, Active Participation questions, questions for Structuring, Initiating or Maintaining Conversation, and Plans and Actions questions. This is the first study to reveal differences in distribution and timing of facial signals across different types of social actions. The findings raise the possibility that facial signals may facilitate social action recognition during language processing in multimodal face-to-face interaction.
Collapse
Affiliation(s)
- Naomi Nota
- Donders Institute for Brain, Cognition, and Behaviour, Radboud University, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - James P Trujillo
- Donders Institute for Brain, Cognition, and Behaviour, Radboud University, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - Judith Holler
- Donders Institute for Brain, Cognition, and Behaviour, Radboud University, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| |
Collapse
|
13
|
Binter J, Pešout O, Pieniak M, Martínez-Molina J, Noon EJ, Stefanczyk MM, Eder SJ. Predictors and motives for mask-wearing behavior and vaccination intention. Sci Rep 2023; 13:10293. [PMID: 37357247 DOI: 10.1038/s41598-023-37072-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2022] [Accepted: 06/15/2023] [Indexed: 06/27/2023] Open
Abstract
Containing a pandemic requires that individuals adhere to measures such as wearing face-masks and getting vaccinated. Therefore, identifying predictors and motives for both behaviors is of importance. Here, we study the decisions made by a cross-national sample in randomized hypothetical scenarios during the COVID-19 pandemic. Our results show that mask-wearing was predicted by empathic tendencies, germ aversion, and higher age, whilst belief in misinformation and presentation of an interaction partner as a family member lowered the safety standards. The main motives associated with taking the mask off included: rationalization, facilitating interaction, and comfort. Vaccination intention was positively predicted by empathy, and negatively predicted by belief in misinformation and higher costs of the vaccine. We found no effect of immunization status of the surrounding social group. The most common motive for vaccination was protection of oneself and others, whereas undecided and anti-vaccine groups reported doubts about the effectiveness and fear of side effects. Together, we identify social and psychological predictors and motives of mask-wearing behavior and vaccination intention. The results highlight the importance of social context for mask-wearing, easy access to vaccines, empathy, and trust in publicly distributed information.
Collapse
Affiliation(s)
- Jakub Binter
- Faculty of Social and Economic Studies, Jan Evangelista Purkyně University, Moskevská 54, 400 96, Ústí nad Labem, Czech Republic.
- Department of Philosophy and History of Science, Faculty of Science, Charles University, Prague, Czech Republic.
| | - Ondra Pešout
- Department of Psychology, Jan Evangelista Purkyně University, Ústí nad Labem, Czech Republic
| | - Michał Pieniak
- Institute of Psychology, University of Wroclaw, Wroclaw, Poland
| | | | - Edward J Noon
- Institute of Childhood and Education, Leeds Trinity University, Leeds, UK
| | | | - Stephanie J Eder
- Department of Neurosciences and Developmental Biology, University of Vienna, Djerrassiplatz 1, 1030, Vienna, Austria.
| |
Collapse
|
14
|
Establishing conversational engagement and being effective: The role of body movement in mediated communication. Acta Psychol (Amst) 2023; 233:103840. [PMID: 36681014 DOI: 10.1016/j.actpsy.2023.103840] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2022] [Revised: 10/28/2022] [Accepted: 01/16/2023] [Indexed: 01/21/2023] Open
Abstract
A model for investigating the effects of body movement on conversational effectiveness in computer-mediated communication (CMC) is developed based on theories of motor cognition and embodiment. Movement is relevant to a wide range of CMC settings, including remote interviews, court testimonials, instructing, medical consultation, and socializing. The present work allows for a consideration of different forms of motoric activation, including gesturing and full-body motion, in mediated conversational settings and the derivation of a range of testable hypothesis. Motor cognition and embodiment provide an account of how speaker and listener become subject to the consequences of the muscular activation patterns that come with body movement. While movement supports internal elaboration, thus helping the speaker in formulating messages, it also has direct effects on the listener through behavioral synchrony and motor contagion. The effects of movement in CMC environments depend on two general characteristics: the level of visibility of movement and the extent to which the technology facilitates or inhibits movement. Available channels, set-up of technology, and further customization therefore determine whether movement can fulfil its internal functions (relevant to cognitive-affective elaboration of what is being said by the speaker) and its external functions (relevant to what is being perceived by and activated within the listener). Several indicators of conversational effectiveness are identified that serve as outcome variables. This MCEE model is intended to help users, developers and service provides to make CMC more engaging and more meaningful.
Collapse
|
15
|
Benetti S, Ferrari A, Pavani F. Multimodal processing in face-to-face interactions: A bridging link between psycholinguistics and sensory neuroscience. Front Hum Neurosci 2023; 17:1108354. [PMID: 36816496 PMCID: PMC9932987 DOI: 10.3389/fnhum.2023.1108354] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2022] [Accepted: 01/11/2023] [Indexed: 02/05/2023] Open
Abstract
In face-to-face communication, humans are faced with multiple layers of discontinuous multimodal signals, such as head, face, hand gestures, speech and non-speech sounds, which need to be interpreted as coherent and unified communicative actions. This implies a fundamental computational challenge: optimally binding only signals belonging to the same communicative action while segregating signals that are not connected by the communicative content. How do we achieve such an extraordinary feat, reliably, and efficiently? To address this question, we need to further move the study of human communication beyond speech-centred perspectives and promote a multimodal approach combined with interdisciplinary cooperation. Accordingly, we seek to reconcile two explanatory frameworks recently proposed in psycholinguistics and sensory neuroscience into a neurocognitive model of multimodal face-to-face communication. First, we introduce a psycholinguistic framework that characterises face-to-face communication at three parallel processing levels: multiplex signals, multimodal gestalts and multilevel predictions. Second, we consider the recent proposal of a lateral neural visual pathway specifically dedicated to the dynamic aspects of social perception and reconceive it from a multimodal perspective ("lateral processing pathway"). Third, we reconcile the two frameworks into a neurocognitive model that proposes how multiplex signals, multimodal gestalts, and multilevel predictions may be implemented along the lateral processing pathway. Finally, we advocate a multimodal and multidisciplinary research approach, combining state-of-the-art imaging techniques, computational modelling and artificial intelligence for future empirical testing of our model.
Collapse
Affiliation(s)
- Stefania Benetti
- Centre for Mind/Brain Sciences, University of Trento, Trento, Italy,Interuniversity Research Centre “Cognition, Language, and Deafness”, CIRCLeS, Catania, Italy,*Correspondence: Stefania Benetti,
| | - Ambra Ferrari
- Max Planck Institute for Psycholinguistics, Donders Institute for Brain, Cognition, and Behaviour, Radboud University, Nijmegen, Netherlands
| | - Francesco Pavani
- Centre for Mind/Brain Sciences, University of Trento, Trento, Italy,Interuniversity Research Centre “Cognition, Language, and Deafness”, CIRCLeS, Catania, Italy
| |
Collapse
|
16
|
Holler J. Visual bodily signals as core devices for coordinating minds in interaction. Philos Trans R Soc Lond B Biol Sci 2022; 377:20210094. [PMID: 35876208 PMCID: PMC9310176 DOI: 10.1098/rstb.2021.0094] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2021] [Accepted: 01/21/2022] [Indexed: 12/11/2022] Open
Abstract
The view put forward here is that visual bodily signals play a core role in human communication and the coordination of minds. Critically, this role goes far beyond referential and propositional meaning. The human communication system that we consider to be the explanandum in the evolution of language thus is not spoken language. It is, instead, a deeply multimodal, multilayered, multifunctional system that developed-and survived-owing to the extraordinary flexibility and adaptability that it endows us with. Beyond their undisputed iconic power, visual bodily signals (manual and head gestures, facial expressions, gaze, torso movements) fundamentally contribute to key pragmatic processes in modern human communication. This contribution becomes particularly evident with a focus that includes non-iconic manual signals, non-manual signals and signal combinations. Such a focus also needs to consider meaning encoded not just via iconic mappings, since kinematic modulations and interaction-bound meaning are additional properties equipping the body with striking pragmatic capacities. Some of these capacities, or its precursors, may have already been present in the last common ancestor we share with the great apes and may qualify as early versions of the components constituting the hypothesized interaction engine. This article is part of the theme issue 'Revisiting the human 'interaction engine': comparative approaches to social action coordination'.
Collapse
Affiliation(s)
- Judith Holler
- Max-Planck-Institut für Psycholinguistik, Nijmegen, The Netherlands
- Donders Centre for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
| |
Collapse
|
17
|
Creaghe N, Kidd E. Symbolic play as a zone of proximal development: An analysis of informational exchange. SOCIAL DEVELOPMENT 2022. [DOI: 10.1111/sode.12592] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
Affiliation(s)
- Noëlie Creaghe
- Research School of Psychology The Australian National University Canberra Australia
- ARC Centre of Excellence for the Dynamics of Language Acton Australia
| | - Evan Kidd
- Research School of Psychology The Australian National University Canberra Australia
- ARC Centre of Excellence for the Dynamics of Language Acton Australia
- Max Planck Institute for Psycholinguistics Nijmegen The Netherlands
- Donders Institute for Brain, Cognition, and Behaviour Radboud University Nijmegen The Netherlands
| |
Collapse
|
18
|
Cuevas P, He Y, Steines M, Straube B. The Processing of Semantic Complexity and Cospeech Gestures in Schizophrenia: A Naturalistic, Multimodal fMRI Study. SCHIZOPHRENIA BULLETIN OPEN 2022; 3:sgac026. [PMID: 39144758 PMCID: PMC11205911 DOI: 10.1093/schizbullopen/sgac026] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 08/16/2024]
Abstract
Schizophrenia is marked by aberrant processing of complex speech and gesture, which may contribute functionally to its impaired social communication. To date, extant neuroscientific studies of schizophrenia have largely investigated dysfunctional speech and gesture in isolation, and no prior research has examined how the two communicative channels may interact in more natural contexts. Here, we tested if patients with schizophrenia show aberrant neural processing of semantically complex story segments, and if speech-associated gestures (co-speech gestures) might modulate this effect. In a functional MRI study, we presented to 34 participants (16 patients and 18 matched-controls) an ecologically-valid retelling of a continuous story, performed via speech and spontaneous gestures. We split the entire story into ten-word segments, and measured the semantic complexity for each segment with idea density, a linguistic measure that is commonly used clinically to evaluate aberrant language dysfunction at the semantic level. Per segment, the presence of numbers of gestures varied (n = 0, 1, +2). Our results suggest that, in comparison to controls, patients showed reduced activation for more complex segments in the bilateral middle frontal and inferior parietal regions. Importantly, this neural aberrance was normalized in segments presented with gestures. Thus, for the first time with a naturalistic multimodal stimulation paradigm, we show that gestures reduced group differences when processing a natural story, probably by facilitating the processing of semantically complex segments of the story in schizophrenia.
Collapse
Affiliation(s)
- Paulina Cuevas
- Translational Neuroimaging Lab Marburg, Department of Psychiatry and Psychotherapy, Philipps University Marburg, Marburg, Germany
- Center for Mind, Brain, and Behavior (CMBB), University of Marburg and Justus Liebig University, Giessen, Marburg, Germany
| | - Yifei He
- Translational Neuroimaging Lab Marburg, Department of Psychiatry and Psychotherapy, Philipps University Marburg, Marburg, Germany
- Center for Mind, Brain, and Behavior (CMBB), University of Marburg and Justus Liebig University, Giessen, Marburg, Germany
| | - Miriam Steines
- Translational Neuroimaging Lab Marburg, Department of Psychiatry and Psychotherapy, Philipps University Marburg, Marburg, Germany
- Center for Mind, Brain, and Behavior (CMBB), University of Marburg and Justus Liebig University, Giessen, Marburg, Germany
| | - Benjamin Straube
- Translational Neuroimaging Lab Marburg, Department of Psychiatry and Psychotherapy, Philipps University Marburg, Marburg, Germany
- Center for Mind, Brain, and Behavior (CMBB), University of Marburg and Justus Liebig University, Giessen, Marburg, Germany
| |
Collapse
|