1
|
Nomiya H, Shimokawa K, Namba S, Osumi M, Sato W. An Artificial Intelligence Model for Sensing Affective Valence and Arousal from Facial Images. SENSORS (BASEL, SWITZERLAND) 2025; 25:1188. [PMID: 40006417 PMCID: PMC11859956 DOI: 10.3390/s25041188] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/13/2025] [Revised: 02/09/2025] [Accepted: 02/13/2025] [Indexed: 02/27/2025]
Abstract
Artificial intelligence (AI) models can sense subjective affective states from facial images. Although recent psychological studies have indicated that dimensional affective states of valence and arousal are systematically associated with facial expressions, no AI models have been developed to estimate these affective states from facial images based on empirical data. We developed a recurrent neural network-based AI model to estimate subjective valence and arousal states from facial images. We trained our model using a database containing participant valence/arousal states and facial images. Leave-one-out cross-validation supported the validity of the model for predicting subjective valence and arousal states. We further validated the effectiveness of the model by analyzing a dataset containing participant valence/arousal ratings and facial videos. The model predicted second-by-second valence and arousal states, with prediction performance comparable to that of FaceReader, a commercial AI model that estimates dimensional affective states based on a different approach. We constructed a graphical user interface to show real-time affective valence and arousal states by analyzing facial video data. Our model is the first distributable AI model for sensing affective valence and arousal from facial images/videos to be developed based on an empirical database; we anticipate that it will have many practical uses, such as in mental health monitoring and marketing research.
Collapse
Affiliation(s)
- Hiroki Nomiya
- Faculty of Information and Human Sciences, Kyoto Institute of Technology, Kyoto 606-8585, Japan;
| | - Koh Shimokawa
- RIKEN, Psychological Process Research Team, Guardian Robot Project, Kyoto 619-0288, Japan;
| | - Shushi Namba
- Department of Psychology, Hiroshima University, Hiroshima 739-8524, Japan;
| | - Masaki Osumi
- KOHINATA Limited Liability Company, Osaka 556-0020, Japan;
| | - Wataru Sato
- RIKEN, Psychological Process Research Team, Guardian Robot Project, Kyoto 619-0288, Japan;
| |
Collapse
|
2
|
Zhang J, Sato W, Kawamura N, Shimokawa K, Tang B, Nakamura Y. Sensing emotional valence and arousal dynamics through automated facial action unit analysis. Sci Rep 2024; 14:19563. [PMID: 39174675 PMCID: PMC11341571 DOI: 10.1038/s41598-024-70563-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/07/2024] [Accepted: 08/19/2024] [Indexed: 08/24/2024] Open
Abstract
Information about the concordance between dynamic emotional experiences and objective signals is practically useful. Previous studies have shown that valence dynamics can be estimated by recording electrical activity from the muscles in the brows and cheeks. However, whether facial actions based on video data and analyzed without electrodes can be used for sensing emotion dynamics remains unknown. We investigated this issue by recording video of participants' faces and obtaining dynamic valence and arousal ratings while they observed emotional films. Action units (AUs) 04 (i.e., brow lowering) and 12 (i.e., lip-corner pulling), detected through an automated analysis of the video data, were negatively and positively correlated with dynamic ratings of subjective valence, respectively. Several other AUs were also correlated with dynamic valence or arousal ratings. Random forest regression modeling, interpreted using the SHapley Additive exPlanation tool, revealed non-linear associations between the AUs and dynamic ratings of valence or arousal. These results suggest that an automated analysis of facial expression video data can be used to estimate dynamic emotional states, which could be applied in various fields including mental health diagnosis, security monitoring, and education.
Collapse
Affiliation(s)
- Junyao Zhang
- Graduate School of Informatics, Kyoto University, Yoshida-Honmachi, Sakyo, Kyoto, 606-8507, Japan
| | - Wataru Sato
- Psychological Process Research Team, Guardian Robot Project, RIKEN, 2-2-2 Hikaridai, Seika-cho, Soraku-gun, Kyoto, 619-0288, Japan.
| | - Naoya Kawamura
- Graduate School of Informatics, Kyoto University, Yoshida-Honmachi, Sakyo, Kyoto, 606-8507, Japan
| | - Koh Shimokawa
- Psychological Process Research Team, Guardian Robot Project, RIKEN, 2-2-2 Hikaridai, Seika-cho, Soraku-gun, Kyoto, 619-0288, Japan
| | - Budu Tang
- Graduate School of Informatics, Kyoto University, Yoshida-Honmachi, Sakyo, Kyoto, 606-8507, Japan
| | - Yuichi Nakamura
- Academic Center for Computing and Media Studies, Kyoto University, Yoshida-Honmachi, Sakyo, Kyoto, 606-8507, Japan
| |
Collapse
|
3
|
Halamová J, Kanovský M, Brockington G, Strnádelová B. Automated facial expression analysis of participants self-criticising via the two-chair technique: exploring facial behavioral markers of self-criticism. Front Psychol 2023; 14:1138916. [PMID: 37179867 PMCID: PMC10166807 DOI: 10.3389/fpsyg.2023.1138916] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2023] [Accepted: 04/05/2023] [Indexed: 05/15/2023] Open
Abstract
Introduction As self-rating scales are prone to many measurement distortions, there is a growing call for more objective measures based on physiological or behavioural indicators. Self-criticism is one of the major transdiagnostic factor of all mental disorders therefore it is important to be able to distinguish what are the characteristic facial features of self-criticizing. To the best of our knowledge, there has been no automated facial emotion expression analysis of participants self-criticising via the two-chair technique. The aim of this study was to detect which action units of facial expressions were significantly more often present in participants performing self-criticism using the two-chair technique. The broader goal was to contribute to the scientific knowledge on objective behavioural descriptions of self-criticism and to provide an additional diagnostic means to the existing self-rating scales by exploring facial behavioral markers of self-criticism. Methods The non-clinical sample consisted of 80 participants (20 men and 60 women) aged 19 years to 57 years (M = 23.86; SD = 5.98). In the analysis we used iMotions's Affectiva AFFDEX module (Version 8.1) to classify the participants' actions units from the self-criticising videos. For the statistical analysis we used a multilevel model to account for the repeated-measures design. Results Based on the significant results the self-critical facial expression may therefore comprise the following action units: Dimpler, Lip Press, Eye Closure, Jaw Drop, and Outer Brow Raise, which are related to contempt, fear, and embarrassment or shame; and Eye Closure and Eye Widen (in rapid sequence Blink), which are a sign that highly negative stimuli are being emotionally processed. Discussion The research study need to be further analysed using clinical samples to compare the results.
Collapse
Affiliation(s)
- Júlia Halamová
- Faculty of Social and Economic Sciences, Institute of Applied Psychology, Comenius University Bratislava, Bratislava, Slovakia
| | - Martin Kanovský
- Faculty of Social and Economic Sciences, Institute of Social Anthropology, Comenius University Bratislava, Bratislava, Slovakia
| | | | - Bronislava Strnádelová
- Faculty of Social and Economic Sciences, Institute of Applied Psychology, Comenius University Bratislava, Bratislava, Slovakia
| |
Collapse
|
4
|
Hyniewska S, Dąbrowska J, Makowska I, Jankowiak-Siuda K, Rymarczyk K. The Borderline Bias in Explicit Emotion Interpretation. Front Psychol 2021; 12:733742. [PMID: 34975623 PMCID: PMC8715824 DOI: 10.3389/fpsyg.2021.733742] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2021] [Accepted: 10/26/2021] [Indexed: 11/28/2022] Open
Abstract
Atypical emotion interpretation has been widely reported in individuals with borderline personality disorder (iBPD); however, empirical studies reported mixed results so far. We suggest that discrepancies in observations of emotion interpretation by iBPD can be explained by biases related to their fear of rejection and abandonment, i.e., the three moral emotions of anger, disgust, and contempt. In this study, we hypothesized that iBPD would show a higher tendency to correctly interpret these three displays of social rejection and attribute more negative valence. A total of 28 inpatient iBPDs and 28 healthy controls were asked to judge static and dynamic facial expressions in terms of emotions, valence, and self-reported arousal evoked by the observed faces. Our results partially confirmed our expectations. The iBPD correctly interpreted the three unambiguous moral emotions. Contempt, a complex emotion with a difficulty in recognizing facial expressions, was recognized better by iBPD than by healthy controls. All negative emotions were judged more negatively by iBPD than by controls, but no difference was observed in the neutral or positive emotion. Alexithymia and anxiety trait and state levels were controlled in all analyses.
Collapse
Affiliation(s)
- Sylwia Hyniewska
- Department of Experimental Psychology, University College London, London, United Kingdom
- *Correspondence: Sylwia Hyniewska,
| | - Joanna Dąbrowska
- Psychiatric Clinic I, Institute of Psychiatry and Neurology, Warsaw, Poland
| | - Iwona Makowska
- Child and Adolescent Psychiatric Department, Medical University of Łódź, Łódź, Poland
| | - Kamila Jankowiak-Siuda
- Department of Biological Psychology, Behavioral Neuroscience Lab, SWPS University of Social Sciences and Humanities, Warsaw, Poland
| | - Krystyna Rymarczyk
- Department of Biological Psychology, Behavioral Neuroscience Lab, SWPS University of Social Sciences and Humanities, Warsaw, Poland
- Krystyna Rymarczyk,
| |
Collapse
|
5
|
Krumhuber EG, Hyniewska S, Orlowska A. Contextual effects on smile perception and recognition memory. CURRENT PSYCHOLOGY 2021. [DOI: 10.1007/s12144-021-01910-5] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
Abstract
AbstractMost past research has focused on the role played by social context information in emotion classification, such as whether a display is perceived as belonging to one emotion category or another. The current study aims to investigate whether the effect of context extends to the interpretation of emotion displays, i.e. smiles that could be judged either as posed or spontaneous readouts of underlying positive emotion. A between-subjects design (N = 93) was used to investigate the perception and recall of posed smiles, presented together with a happy or polite social context scenario. Results showed that smiles seen in a happy context were judged as more spontaneous than the same smiles presented in a polite context. Also, smiles were misremembered as having more of the physical attributes (i.e., Duchenne marker) associated with spontaneous enjoyment when they appeared in the happy than polite context condition. Together, these findings indicate that social context information is routinely encoded during emotion perception, thereby shaping the interpretation and recognition memory of facial expressions.
Collapse
|
6
|
Namba S, Kambara T. Semantics Based on the Physical Characteristics of Facial Expressions Used to Produce Japanese Vowels. Behav Sci (Basel) 2020; 10:E157. [PMID: 33066229 PMCID: PMC7602070 DOI: 10.3390/bs10100157] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2020] [Revised: 10/02/2020] [Accepted: 10/12/2020] [Indexed: 12/18/2022] Open
Abstract
Previous studies have reported that verbal sounds are associated-non-arbitrarily-with specific meanings (e.g., sound symbolism and onomatopoeia), including visual forms of information such as facial expressions; however, it remains unclear how mouth shapes used to utter each vowel create our semantic impressions. We asked 81 Japanese participants to evaluate mouth shapes associated with five Japanese vowels by using 10 five-item semantic differential scales. The results reveal that the physical characteristics of the facial expressions (mouth shapes) induced specific evaluations. For example, the mouth shape made to voice the vowel "a" was the one with the biggest, widest, and highest facial components compared to other mouth shapes, and people perceived words containing that vowel sound as bigger. The mouth shapes used to pronounce the vowel "i" were perceived as more likable than the other four vowels. These findings indicate that the mouth shapes producing vowels imply specific meanings. Our study provides clues about the meaning of verbal sounds and what the facial expressions in communication represent to the perceiver.
Collapse
Affiliation(s)
- Shushi Namba
- Psychological Process Team, BZP, Robotics Project, RIKEN, 2-2-2 Hikaridai, Seika-cho, Soraku-gun, Kyoto 6190288, Japan;
| | - Toshimune Kambara
- Department of Psychology, Graduate School of Education, Hiroshima University, 1-1-1 Kagamiyama, Higashi-Hiroshima, Hiroshima 7398524, Japan
| |
Collapse
|
7
|
Scherer KR, Ellgring H, Dieckmann A, Unfried M, Mortillaro M. Dynamic Facial Expression of Emotion and Observer Inference. Front Psychol 2019; 10:508. [PMID: 30941073 PMCID: PMC6434775 DOI: 10.3389/fpsyg.2019.00508] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2018] [Accepted: 02/20/2019] [Indexed: 11/13/2022] Open
Abstract
Research on facial emotion expression has mostly focused on emotion recognition, assuming that a small number of discrete emotions is elicited and expressed via prototypical facial muscle configurations as captured in still photographs. These are expected to be recognized by observers, presumably via template matching. In contrast, appraisal theories of emotion propose a more dynamic approach, suggesting that specific elements of facial expressions are directly produced by the result of certain appraisals and predicting the facial patterns to be expected for certain appraisal configurations. This approach has recently been extended to emotion perception, claiming that observers first infer individual appraisals and only then make categorical emotion judgments based on the estimated appraisal patterns, using inference rules. Here, we report two related studies to empirically investigate the facial action unit configurations that are used by actors to convey specific emotions in short affect bursts and to examine to what extent observers can infer a person's emotions from the predicted facial expression configurations. The results show that (1) professional actors use many of the predicted facial action unit patterns to enact systematically specified appraisal outcomes in a realistic scenario setting, and (2) naïve observers infer the respective emotions based on highly similar facial movement configurations with a degree of accuracy comparable to earlier research findings. Based on estimates of underlying appraisal criteria for the different emotions we conclude that the patterns of facial action units identified in this research correspond largely to prior predictions and encourage further research on appraisal-driven expression and inference.
Collapse
Affiliation(s)
- Klaus R Scherer
- Department of Psychology and Swiss Center for Affective Sciences, University of Geneva, Geneva, Switzerland
| | - Heiner Ellgring
- Department of Psychology, University of Würzburg, Würzburg, Germany
| | | | | | - Marcello Mortillaro
- Department of Psychology and Swiss Center for Affective Sciences, University of Geneva, Geneva, Switzerland
| |
Collapse
|