1
|
Reitelshöfer S, Merz N, Garcia G, Wei Y, Franke J. Making social robots adaptable and to some extent educable by a marketplace for the selection and adjustment of different interaction characters living inside a single robot. Front Robot AI 2025; 12:1534346. [PMID: 40264561 PMCID: PMC12012301 DOI: 10.3389/frobt.2025.1534346] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2024] [Accepted: 03/10/2025] [Indexed: 04/24/2025] Open
Abstract
The increasing integration of autonomous robotic systems across various industries necessitates adaptable social interaction capabilities. This paper presents a novel software architecture for socially adaptable robots, emphasizing simplicity, domain independence, and user influence on robotic behaviour. The architecture leverages a marketplace-based agent selection system to dynamically adapt social interaction patterns to diverse users and scenarios. Implemented using ROS2, the framework comprises four core components: scene analysis, a bidding platform, social agents, and a feedback service. A Validation through simulated experiments shows the architecture's feasibility and adaptability, with respect to varying feedback conditions and learning rates. This work lays the foundation for scalable, adaptable, and user-friendly robotic systems, addressing key challenges in industrial and social robotics. Future improvements include enhanced scene analysis, integration of machine learning techniques, and support for more complex behavioural scripts.
Collapse
Affiliation(s)
- Sebastian Reitelshöfer
- Institute for Factory Automation and Production Systems, Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen, Germany
| | | | | | | | | |
Collapse
|
2
|
Li M, Qin J, Li J, Liu Q, Shi Y, Kang Y. Game-Based Approximate Optimal Motion Planning for Safe Human-Swarm Interaction. IEEE TRANSACTIONS ON CYBERNETICS 2024; 54:5649-5660. [PMID: 38163300 DOI: 10.1109/tcyb.2023.3340659] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/03/2024]
Abstract
Safety as a fundamental requirement for human-swarm interaction has attracted a lot of attention in recent years. Most existing approaches solve a constrained optimization problem at each time step, which has a high real-time requirement. To deal with this challenge, this article formulates the safe human-swarm interaction problem as a Stackerberg-Nash game, in which the optimization is performed over the entire time domain. The leader robot is supposed to be in a dominant position, interacting directly with the human operator to realize trajectory tracking and responsible for guiding the swarm to avoid obstacles. The follower robots always take their best responses to leader's behavior with the purpose of achieving the desired formation. Following the bottom-up principle, we first design the best-response controllers, that is, Nash equilibrium strategies, for the followers. Then, a Lyapunov-like control barrier function-based safety controller and a learning-based formation tracking controller for the leader are designed to realize safe and robust cooperation. We show that the designed controllers can make the robotic swarms move in a desired geometric formation following the human command and modify their motion trajectories autonomously when the human command is unsafe. The effectiveness of the proposed approach is verified through simulation and experiments. The experiment results further show that safety can still be guaranteed even when there exists a dynamic obstacle.
Collapse
|
3
|
Maroto-Gómez M, Burguete-Alventosa J, Álvarez-Arias S, Malfaz M, Salichs MÁ. A Bio-Inspired Dopamine Model for Robots with Autonomous Decision-Making. Biomimetics (Basel) 2024; 9:504. [PMID: 39194483 DOI: 10.3390/biomimetics9080504] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/15/2024] [Revised: 08/18/2024] [Accepted: 08/19/2024] [Indexed: 08/29/2024] Open
Abstract
Decision-making systems allow artificial agents to adapt their behaviours, depending on the information they perceive from the environment and internal processes. Human beings possess unique decision-making capabilities, adapting to current situations and anticipating future challenges. Autonomous robots with adaptive and anticipatory decision-making emulating humans can bring robots with skills that users can understand more easily. Human decisions highly depend on dopamine, a brain substance that regulates motivation and reward, acknowledging positive and negative situations. Considering recent neuroscience studies about the dopamine role in the human brain and its influence on decision-making and motivated behaviour, this paper proposes a model based on how dopamine drives human motivation and decision-making. The model allows robots to behave autonomously in dynamic environments, learning the best action selection strategy and anticipating future rewards. The results show the model's performance in five scenarios, emphasising how dopamine levels vary depending on the robot's situation and stimuli perception. Moreover, we show the model's integration into the Mini social robot to provide insights into how dopamine levels drive motivated autonomous behaviour regulating biologically inspired internal processes emulated in the robot.
Collapse
Affiliation(s)
- Marcos Maroto-Gómez
- Department of Systems Engineering and Automation, University Carlos III of Madrid, Av. de la Universidad, 30, 28911 Leganes, Madrid, Spain
| | - Javier Burguete-Alventosa
- Department of Systems Engineering and Automation, University Carlos III of Madrid, Av. de la Universidad, 30, 28911 Leganes, Madrid, Spain
| | - Sofía Álvarez-Arias
- Department of Systems Engineering and Automation, University Carlos III of Madrid, Av. de la Universidad, 30, 28911 Leganes, Madrid, Spain
| | - María Malfaz
- Department of Systems Engineering and Automation, University Carlos III of Madrid, Av. de la Universidad, 30, 28911 Leganes, Madrid, Spain
| | - Miguel Ángel Salichs
- Department of Systems Engineering and Automation, University Carlos III of Madrid, Av. de la Universidad, 30, 28911 Leganes, Madrid, Spain
| |
Collapse
|
4
|
Fernández-Rodicio E, Castro-González Á, Gamboa-Montero JJ, Carrasco-Martínez S, Salichs MA. Creating Expressive Social Robots That Convey Symbolic and Spontaneous Communication. SENSORS (BASEL, SWITZERLAND) 2024; 24:3671. [PMID: 38894462 PMCID: PMC11175349 DOI: 10.3390/s24113671] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/04/2024] [Revised: 05/30/2024] [Accepted: 06/03/2024] [Indexed: 06/21/2024]
Abstract
Robots are becoming an increasingly important part of our society and have started to be used in tasks that require communicating with humans. Communication can be decoupled in two dimensions: symbolic (information aimed to achieve a particular goal) and spontaneous (displaying the speaker's emotional and motivational state) communication. Thus, to enhance human-robot interactions, the expressions that are used have to convey both dimensions. This paper presents a method for modelling a robot's expressiveness as a combination of these two dimensions, where each of them can be generated independently. This is the first contribution of our work. The second contribution is the development of an expressiveness architecture that uses predefined multimodal expressions to convey the symbolic dimension and integrates a series of modulation strategies for conveying the robot's mood and emotions. In order to validate the performance of the proposed architecture, the last contribution is a series of experiments that aim to study the effect that the addition of the spontaneous dimension of communication and its fusion with the symbolic dimension has on how people perceive a social robot. Our results show that the modulation strategies improve the users' perception and can convey a recognizable affective state.
Collapse
Affiliation(s)
- Enrique Fernández-Rodicio
- RoboticsLab, Department of Systems Engineering and Automation, Universidad Carlos III de Madrid, Av. de la Universidad 30, 28911 Madrid, Spain; (Á.C.-G.); (J.J.G.-M.); (S.C.-M.); (M.A.S.)
| | | | | | | | | |
Collapse
|
5
|
Irfan B, Kuoppamäki S, Skantze G. Recommendations for designing conversational companion robots with older adults through foundation models. Front Robot AI 2024; 11:1363713. [PMID: 38860032 PMCID: PMC11163135 DOI: 10.3389/frobt.2024.1363713] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2023] [Accepted: 05/07/2024] [Indexed: 06/12/2024] Open
Abstract
Companion robots are aimed to mitigate loneliness and social isolation among older adults by providing social and emotional support in their everyday lives. However, older adults' expectations of conversational companionship might substantially differ from what current technologies can achieve, as well as from other age groups like young adults. Thus, it is crucial to involve older adults in the development of conversational companion robots to ensure that these devices align with their unique expectations and experiences. The recent advancement in foundation models, such as large language models, has taken a significant stride toward fulfilling those expectations, in contrast to the prior literature that relied on humans controlling robots (i.e., Wizard of Oz) or limited rule-based architectures that are not feasible to apply in the daily lives of older adults. Consequently, we conducted a participatory design (co-design) study with 28 older adults, demonstrating a companion robot using a large language model (LLM), and design scenarios that represent situations from everyday life. The thematic analysis of the discussions around these scenarios shows that older adults expect a conversational companion robot to engage in conversation actively in isolation and passively in social settings, remember previous conversations and personalize, protect privacy and provide control over learned data, give information and daily reminders, foster social skills and connections, and express empathy and emotions. Based on these findings, this article provides actionable recommendations for designing conversational companion robots for older adults with foundation models, such as LLMs and vision-language models, which can also be applied to conversational robots in other domains.
Collapse
Affiliation(s)
- Bahar Irfan
- Division of Speech, Music and Hearing, KTH Royal Institute of Technology, Stockholm, Sweden
| | - Sanna Kuoppamäki
- Division of Health Informatics and Logistics, KTH Royal Institute of Technology, Stockholm, Sweden
| | - Gabriel Skantze
- Division of Speech, Music and Hearing, KTH Royal Institute of Technology, Stockholm, Sweden
| |
Collapse
|
6
|
Maroto-Gómez M, Castro-González Á, Malfaz M, Salichs MÁ. A biologically inspired decision-making system for the autonomous adaptive behavior of social robots. COMPLEX INTELL SYST 2023:1-19. [PMID: 37361968 PMCID: PMC10225289 DOI: 10.1007/s40747-023-01077-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2022] [Accepted: 04/17/2023] [Indexed: 06/28/2023]
Abstract
The decisions made by social robots while they fulfill their tasks have a strong influence on their performance. In these contexts, autonomous social robots must exhibit adaptive and social-based behavior to make appropriate decisions and operate correctly in complex and dynamic scenarios. This paper presents a Decision-Making System for social robots working on long-term interactions like cognitive stimulation or entertainment. The Decision-making System employs the robot's sensors, user information, and a biologically inspired module to replicate how human behavior emerges in the robot. Besides, the system personalizes the interaction to maintain the users' engagement while adapting to their features and preferences, overcoming possible interaction limitations. The system evaluation was in terms of usability, performance metrics, and user perceptions. We used the Mini social robot as the device where we integrated the architecture and carried out the experimentation. The usability evaluation consisted of 30 participants interacting with the autonomous robot in 30 min sessions. Then, 19 participants evaluated their perceptions of robot attributes of the Godspeed questionnaire by playing with the robot in 30 min sessions. The participants rated the Decision-making System with excellent usability (81.08 out of 100 points), perceiving the robot as intelligent (4.28 out of 5), animated (4.07 out of 5), and likable (4.16 out of 5). However, they also rated Mini as unsafe (security perceived as 3.15 out of 5), probably because users could not influence the robot's decisions.
Collapse
Affiliation(s)
- Marcos Maroto-Gómez
- Systems Engineering and Automation, University Carlos III of Madrid, Butarque 15, 28911 Leganés, Madrid Spain
| | - Álvaro Castro-González
- Systems Engineering and Automation, University Carlos III of Madrid, Butarque 15, 28911 Leganés, Madrid Spain
| | - María Malfaz
- Systems Engineering and Automation, University Carlos III of Madrid, Butarque 15, 28911 Leganés, Madrid Spain
| | - Miguel Ángel Salichs
- Systems Engineering and Automation, University Carlos III of Madrid, Butarque 15, 28911 Leganés, Madrid Spain
| |
Collapse
|
7
|
Zhuang H, Xia Y, Wang N, Li W, Dong L, Li B. Interactive method research of dual mode information coordination integration for astronaut gesture and eye movement signals based on hybrid model. SCIENCE CHINA. TECHNOLOGICAL SCIENCES 2023; 66:1717-1733. [PMID: 37288339 PMCID: PMC10182537 DOI: 10.1007/s11431-022-2368-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/27/2022] [Accepted: 03/02/2023] [Indexed: 06/09/2023]
Abstract
The lightweight human-robot interaction model with high real-time, high accuracy, and strong anti-interference capability can be better applied to future lunar surface exploration and construction work. Based on the feature information inputted from the monocular camera, the signal acquisition and processing fusion of the astronaut gesture and eye-movement modal interaction can be performed. Compared with the single mode, the human-robot interaction model of bimodal collaboration can achieve the issuance of complex interactive commands more efficiently. The optimization of the target detection model is executed by inserting attention into YOLOv4 and filtering image motion blur. The central coordinates of pupils are identified by the neural network to realize the human-robot interaction in the eye movement mode. The fusion between the astronaut gesture signal and eye movement signal is performed at the end of the collaborative model to achieve complex command interactions based on a lightweight model. The dataset used in the network training is enhanced and extended to simulate the realistic lunar space interaction environment. The human-robot interaction effects of complex commands in the single mode are compared with those of complex commands in the bimodal collaboration. The experimental results show that the concatenated interaction model of the astronaut gesture and eye movement signals can excavate the bimodal interaction signal better, discriminate the complex interaction commands more quickly, and has stronger signal anti-interference capability based on its stronger feature information mining ability. Compared with the command interaction realized by using the single gesture modal signal and the single eye movement modal signal, the interaction model of bimodal collaboration is shorter about 79% to 91% of the time under the single mode interaction. Regardless of the influence of any image interference item, the overall judgment accuracy of the proposed model can be maintained at about 83% to 97%. The effectiveness of the proposed method is verified.
Collapse
Affiliation(s)
- HongChao Zhuang
- School of Mechanical Engineering, Tianjin University of Technology and Education, Tianjin, 300222 China
| | - YiLu Xia
- School of Mechanical Engineering, Tianjin University of Technology and Education, Tianjin, 300222 China
| | - Ning Wang
- School of Information Technology Engineering, Tianjin University of Technology and Education, Tianjin, 300222 China
| | - WeiHua Li
- School of Automotive Engineering, Harbin Institute of Technology (Weihai), Weihai, 264209 China
| | - Lei Dong
- School of Mechanical Engineering, Tianjin University of Technology and Education, Tianjin, 300222 China
| | - Bo Li
- Tianjin Institute of Aerospace Mechanical and Electrical Equipment, Tianjin, 300458 China
- School of Mechatronics Engineering, Harbin Institute of Technology, Harbin, 150000 China
| |
Collapse
|
8
|
Maroto-Gómez M, Alonso-Martín F, Malfaz M, Castro-González Á, Castillo JC, Salichs MÁ. A Systematic Literature Review of Decision-Making and Control Systems for Autonomous and Social Robots. Int J Soc Robot 2023. [DOI: 10.1007/s12369-023-00977-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/29/2023]
Abstract
AbstractIn the last years, considerable research has been carried out to develop robots that can improve our quality of life during tedious and challenging tasks. In these contexts, robots operating without human supervision open many possibilities to assist people in their daily activities. When autonomous robots collaborate with humans, social skills are necessary for adequate communication and cooperation. Considering these facts, endowing autonomous and social robots with decision-making and control models is critical for appropriately fulfiling their initial goals. This manuscript presents a systematic review of the evolution of decision-making systems and control architectures for autonomous and social robots in the last three decades. These architectures have been incorporating new methods based on biologically inspired models and Machine Learning to enhance these systems’ possibilities to developed societies. The review explores the most novel advances in each application area, comparing their most essential features. Additionally, we describe the current challenges of software architecture devoted to action selection, an analysis not provided in similar reviews of behavioural models for autonomous and social robots. Finally, we present the future directions that these systems can take in the future.
Collapse
|
9
|
Long-Term Exercise Assistance: Group and One-on-One Interactions between a Social Robot and Seniors. ROBOTICS 2023. [DOI: 10.3390/robotics12010009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/11/2023] Open
Abstract
For older adults, regular exercises can provide both physical and mental benefits, increase their independence, and reduce the risks of diseases associated with aging. However, only a small portion of older adults regularly engage in physical activity. Therefore, it is important to promote exercise among older adults to help maintain overall health. In this paper, we present the first exploratory long-term human–robot interaction (HRI) study conducted at a local long-term care facility to investigate the benefits of one-on-one and group exercise interactions with an autonomous socially assistive robot and older adults. To provide targeted facilitation, our robot utilizes a unique emotion model that can adapt its assistive behaviors to users’ affect and track their progress towards exercise goals through repeated sessions using the Goal Attainment Scale (GAS), while also monitoring heart rate to prevent overexertion. Results of the study show that users had positive valence and high engagement towards the robot and were able to maintain their exercise performance throughout the study. Questionnaire results showed high robot acceptance for both types of interactions. However, users in the one-on-one sessions perceived the robot as more sociable and intelligent, and had more positive perception of the robot’s appearance and movements.
Collapse
|
10
|
Learning coordinated emotion representation between voice and face. APPL INTELL 2022. [DOI: 10.1007/s10489-022-04216-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
11
|
Emotion and Mood Blending in Embodied Artificial Agents: Expressing Affective States in the Mini Social Robot. Int J Soc Robot 2022. [DOI: 10.1007/s12369-022-00915-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/14/2022]
Abstract
AbstractRobots that are devised for assisting and interacting with humans are becoming fundamental in many applications, including in healthcare, education, and entertainment. For these robots, the capacity to exhibit affective states plays a crucial role in creating emotional bonding with the user. In this work, we present an affective architecture that grounds biological foundations to shape the affective state of the Mini social robot in terms of mood and emotion blending. The affective state depends upon the perception of stimuli in the environment, which influence how the robot behaves and affectively communicates with other peers. According to research in neuroscience, mood typically rules our affective state in the long run, while emotions do it in the short term, although both processes can overlap. Consequently, the model that is presented in this manuscript deals with emotion and mood blending towards expressing the robot’s internal state to the users. Thus, the primary novelty of our affective model is the expression of: (i) mood, (ii) punctual emotional reactions to stimuli, and (iii) the decay that mood and emotion undergo with time. The system evaluation explored whether users can correctly perceive the mood and emotions that the robot is expressing. In an online survey, users evaluated the robot’s expressions showing different moods and emotions. The results reveal that users could correctly perceive the robot’s mood and emotion. However, emotions were more easily recognized, probably because they are more intense affective states and mainly arise as a stimuli reaction. To conclude the manuscript, a case study shows how our model modulates Mini’s expressiveness depending on its affective state during a human-robot interaction scenario.
Collapse
|
12
|
Xu C, Wang M, Chi G, Liu Q. An inertial neural network approach for loco-manipulation trajectory tracking of mobile robot with redundant manipulator. Neural Netw 2022; 155:215-223. [DOI: 10.1016/j.neunet.2022.08.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/19/2022] [Revised: 07/20/2022] [Accepted: 08/11/2022] [Indexed: 10/31/2022]
|
13
|
Berry M, Lewin S, Brown S. Correlated expression of the body, face, and voice during character portrayal in actors. Sci Rep 2022; 12:8253. [PMID: 35585175 PMCID: PMC9117657 DOI: 10.1038/s41598-022-12184-7] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2022] [Accepted: 05/03/2022] [Indexed: 11/25/2022] Open
Abstract
Actors are required to engage in multimodal modulations of their body, face, and voice in order to create a holistic portrayal of a character during performance. We present here the first trimodal analysis, to our knowledge, of the process of character portrayal in professional actors. The actors portrayed a series of stock characters (e.g., king, bully) that were organized according to a predictive scheme based on the two orthogonal personality dimensions of assertiveness and cooperativeness. We used 3D motion capture technology to analyze the relative expansion/contraction of 6 body segments across the head, torso, arms, and hands. We compared this with previous results for these portrayals for 4 segments of facial expression and the vocal parameters of pitch and loudness. The results demonstrated significant cross-modal correlations for character assertiveness (but not cooperativeness), as manifested collectively in a straightening of the head and torso, expansion of the arms and hands, lowering of the jaw, and a rise in vocal pitch and loudness. These results demonstrate what communication theorists refer to as "multichannel reinforcement". We discuss this reinforcement in light of both acting theories and theories of human communication more generally.
Collapse
Affiliation(s)
- Matthew Berry
- Department of Psychology, Neuroscience & Behaviour, McMaster University, 1280 Main St. West, Hamilton, ON, L8S 4K1, Canada.
| | - Sarah Lewin
- Department of Psychology, Neuroscience & Behaviour, McMaster University, 1280 Main St. West, Hamilton, ON, L8S 4K1, Canada
| | - Steven Brown
- Department of Psychology, Neuroscience & Behaviour, McMaster University, 1280 Main St. West, Hamilton, ON, L8S 4K1, Canada
| |
Collapse
|
14
|
Quiroz M, Patiño R, Diaz-Amado J, Cardinale Y. Group Emotion Detection Based on Social Robot Perception. SENSORS (BASEL, SWITZERLAND) 2022; 22:3749. [PMID: 35632160 PMCID: PMC9145339 DOI: 10.3390/s22103749] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/01/2022] [Revised: 05/03/2022] [Accepted: 05/05/2022] [Indexed: 12/16/2022]
Abstract
Social robotics is an emerging area that is becoming present in social spaces, by introducing autonomous social robots. Social robots offer services, perform tasks, and interact with people in such social environments, demanding more efficient and complex Human-Robot Interaction (HRI) designs. A strategy to improve HRI is to provide robots with the capacity of detecting the emotions of the people around them to plan a trajectory, modify their behaviour, and generate an appropriate interaction with people based on the analysed information. However, in social environments in which it is common to find a group of persons, new approaches are needed in order to make robots able to recognise groups of people and the emotion of the groups, which can be also associated with a scene in which the group is participating. Some existing studies are focused on detecting group cohesion and the recognition of group emotions; nevertheless, these works do not focus on performing the recognition tasks from a robocentric perspective, considering the sensory capacity of robots. In this context, a system to recognise scenes in terms of groups of people, to then detect global (prevailing) emotions in a scene, is presented. The approach proposed to visualise and recognise emotions in typical HRI is based on the face size of people recognised by the robot during its navigation (face sizes decrease when the robot moves away from a group of people). On each frame of the video stream of the visual sensor, individual emotions are recognised based on the Visual Geometry Group (VGG) neural network pre-trained to recognise faces (VGGFace); then, to detect the emotion of the frame, individual emotions are aggregated with a fusion method, and consequently, to detect global (prevalent) emotion in the scene (group of people), the emotions of its constituent frames are also aggregated. Additionally, this work proposes a strategy to create datasets with images/videos in order to validate the estimation of emotions in scenes and personal emotions. Both datasets are generated in a simulated environment based on the Robot Operating System (ROS) from videos captured by robots through their sensory capabilities. Tests are performed in two simulated environments in ROS/Gazebo: a museum and a cafeteria. Results show that the accuracy in the detection of individual emotions is 99.79% and the detection of group emotion (scene emotion) in each frame is 90.84% and 89.78% in the cafeteria and the museum scenarios, respectively.
Collapse
Affiliation(s)
- Marco Quiroz
- Electrical and Electronics Engineering Department, School of Electronics and Telecommunications Engineering, Universidad Católica San Pablo, Arequipa 04001, Peru; (M.Q.); (R.P.); (J.D.-A.)
| | - Raquel Patiño
- Electrical and Electronics Engineering Department, School of Electronics and Telecommunications Engineering, Universidad Católica San Pablo, Arequipa 04001, Peru; (M.Q.); (R.P.); (J.D.-A.)
| | - José Diaz-Amado
- Electrical and Electronics Engineering Department, School of Electronics and Telecommunications Engineering, Universidad Católica San Pablo, Arequipa 04001, Peru; (M.Q.); (R.P.); (J.D.-A.)
- Instituto Federal da Bahia, Vitoria da Conquista 45078-300, Brazil
| | - Yudith Cardinale
- Electrical and Electronics Engineering Department, School of Electronics and Telecommunications Engineering, Universidad Católica San Pablo, Arequipa 04001, Peru; (M.Q.); (R.P.); (J.D.-A.)
- Higher School of Engineering, Science and Technology, Universidad Internacional de Valencia, 46002 Valencia, Spain
| |
Collapse
|
15
|
Chang Y, Sun L. EEG-Based Emotion Recognition for Modulating Social-Aware Robot Navigation. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2021; 2021:5709-5712. [PMID: 34892417 DOI: 10.1109/embc46164.2021.9630721] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Companion robots play an important role to accompany humans and provide emotional support, such as reducing human social isolation and loneliness. Based on recognizing human partner's mental states, a companion robot is able to dynamically adjust its behaviors, and make human-robot interaction smoother and natural. Human emotion has been recognized by many modalities like facial expression and voice. Neurophysiological signals have shown promising results in emotion recognition, since it is an innate signal of human brain which cannot be faked. In this paper, emotional state recognition using a neurophysiology method is studied to guide and modulate companion-robot navigation to enhance its social capabilities. Electroencephalogram (EEG), a type of neurophysiological signals, is used to recognize human emotional state, and then feed into a navigation path planning algorithm for controlling a companion robot's routes. Simulation results show that mobile robot presents navigation behaviors modulated by dynamic human emotional states.
Collapse
|
16
|
Cano S, González CS, Gil-Iranzo RM, Albiol-Pérez S. Affective Communication for Socially Assistive Robots (SARs) for Children with Autism Spectrum Disorder: A Systematic Review. SENSORS 2021; 21:s21155166. [PMID: 34372402 PMCID: PMC8347754 DOI: 10.3390/s21155166] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/21/2021] [Revised: 07/18/2021] [Accepted: 07/26/2021] [Indexed: 11/16/2022]
Abstract
Research on affective communication for socially assistive robots has been conducted to enable physical robots to perceive, express, and respond emotionally. However, the use of affective computing in social robots has been limited, especially when social robots are designed for children, and especially those with autism spectrum disorder (ASD). Social robots are based on cognitive-affective models, which allow them to communicate with people following social behaviors and rules. However, interactions between a child and a robot may change or be different compared to those with an adult or when the child has an emotional deficit. In this study, we systematically reviewed studies related to computational models of emotions for children with ASD. We used the Scopus, WoS, Springer, and IEEE-Xplore databases to answer different research questions related to the definition, interaction, and design of computational models supported by theoretical psychology approaches from 1997 to 2021. Our review found 46 articles; not all the studies considered children or those with ASD.
Collapse
Affiliation(s)
- Sandra Cano
- School of Computer Engineering, Pontificia Universidad Católica de Valparaíso, Valparaíso 2340000, Chile
- Correspondence:
| | - Carina S. González
- Department of Computer Engineering and Systems, University of La Laguna, 38204 La Laguna, Spain;
| | - Rosa María Gil-Iranzo
- Department of Computer Engineering and Industrial, University of Lleida, 25001 Lleida, Spain;
| | - Sergio Albiol-Pérez
- Aragón Health Research Institute (IIS Aragón), Universidad de Zaragoza, Cdad. Escolar, 4, 44003 Teruel, Spain;
| |
Collapse
|
17
|
Xie B, Sidulova M, Park CH. Robust Multimodal Emotion Recognition from Conversation with Transformer-Based Crossmodality Fusion. SENSORS (BASEL, SWITZERLAND) 2021; 21:4913. [PMID: 34300651 PMCID: PMC8309929 DOI: 10.3390/s21144913] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/01/2021] [Revised: 07/17/2021] [Accepted: 07/17/2021] [Indexed: 11/29/2022]
Abstract
Decades of scientific research have been conducted on developing and evaluating methods for automated emotion recognition. With exponentially growing technology, there is a wide range of emerging applications that require emotional state recognition of the user. This paper investigates a robust approach for multimodal emotion recognition during a conversation. Three separate models for audio, video and text modalities are structured and fine-tuned on the MELD. In this paper, a transformer-based crossmodality fusion with the EmbraceNet architecture is employed to estimate the emotion. The proposed multimodal network architecture can achieve up to 65% accuracy, which significantly surpasses any of the unimodal models. We provide multiple evaluation techniques applied to our work to show that our model is robust and can even outperform the state-of-the-art models on the MELD.
Collapse
Affiliation(s)
| | | | - Chung Hyuk Park
- Department of Biomedical Engineering, School of Engineering and Applied Science, George Washington University, Washington, DC 20052, USA; (B.X.); (M.S.)
| |
Collapse
|
18
|
Fuentetaja R, García-Olaya A, García J, González JC, Fernández F. An Automated Planning Model for HRI: Use Cases on Social Assistive Robotics. SENSORS 2020; 20:s20226520. [PMID: 33202674 PMCID: PMC7697113 DOI: 10.3390/s20226520] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/01/2020] [Revised: 11/05/2020] [Accepted: 11/11/2020] [Indexed: 11/16/2022]
Abstract
Using Automated Planning for the high level control of robotic architectures is becoming very popular thanks mainly to its capability to define the tasks to perform in a declarative way. However, classical planning tasks, even in its basic standard Planning Domain Definition Language (PDDL) format, are still very hard to formalize for non expert engineers when the use case to model is complex. Human Robot Interaction (HRI) is one of those complex environments. This manuscript describes the rationale followed to design a planning model able to control social autonomous robots interacting with humans. It is the result of the authors’ experience in modeling use cases for Social Assistive Robotics (SAR) in two areas related to healthcare: Comprehensive Geriatric Assessment (CGA) and non-contact rehabilitation therapies for patients with physical impairments. In this work a general definition of these two use cases in a unique planning domain is proposed, which favors the management and integration with the software robotic architecture, as well as the addition of new use cases. Results show that the model is able to capture all the relevant aspects of the Human-Robot interaction in those scenarios, allowing the robot to autonomously perform the tasks by using a standard planning-execution architecture.
Collapse
|
19
|
Abstract
To effectively communicate with people, social robots must be capable of detecting, interpreting, and responding to human affect during human–robot interactions (HRIs). In order to accurately detect user affect during HRIs, affect elicitation techniques need to be developed to create and train appropriate affect detection models. In this paper, we present such a novel affect elicitation and detection method for social robots in HRIs. Non-verbal emotional behaviors of the social robot were designed to elicit user affect, which was directly measured through electroencephalography (EEG) signals. HRI experiments with both younger and older adults were conducted to evaluate our affect elicitation technique and compare the two types of affect detection models we developed and trained utilizing multilayer perceptron neural networks (NNs) and support vector machines (SVMs). The results showed that; on average, the self-reported valence and arousal were consistent with the intended elicited affect. Furthermore, it was also noted that the EEG data obtained could be used to train affect detection models with the NN models achieving higher classification rates
Collapse
|