1
|
Karwowski J, Szynkiewicz W, Niewiadomska-Szynkiewicz E. Bridging Requirements, Planning, and Evaluation: A Review of Social Robot Navigation. SENSORS (BASEL, SWITZERLAND) 2024; 24:2794. [PMID: 38732900 PMCID: PMC11086376 DOI: 10.3390/s24092794] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/25/2024] [Revised: 04/21/2024] [Accepted: 04/24/2024] [Indexed: 05/13/2024]
Abstract
Navigation lies at the core of social robotics, enabling robots to navigate and interact seamlessly in human environments. The primary focus of human-aware robot navigation is minimizing discomfort among surrounding humans. Our review explores user studies, examining factors that cause human discomfort, to perform the grounding of social robot navigation requirements and to form a taxonomy of elementary necessities that should be implemented by comprehensive algorithms. This survey also discusses human-aware navigation from an algorithmic perspective, reviewing the perception and motion planning methods integral to social navigation. Additionally, the review investigates different types of studies and tools facilitating the evaluation of social robot navigation approaches, namely datasets, simulators, and benchmarks. Our survey also identifies the main challenges of human-aware navigation, highlighting the essential future work perspectives. This work stands out from other review papers, as it not only investigates the variety of methods for implementing human awareness in robot control systems but also classifies the approaches according to the grounded requirements regarded in their objectives.
Collapse
Affiliation(s)
| | | | - Ewa Niewiadomska-Szynkiewicz
- Institute of Control and Computation Engineering, Warsaw University of Technology, 00-665 Warsaw, Poland; (J.K.); (W.S.)
| |
Collapse
|
2
|
Caccavale R, Finzi A. A Robotic Cognitive Control Framework for Collaborative Task Execution and Learning. Top Cogn Sci 2021; 14:327-343. [PMID: 34826350 DOI: 10.1111/tops.12587] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2021] [Revised: 10/20/2021] [Accepted: 10/21/2021] [Indexed: 11/30/2022]
Abstract
In social and service robotics, complex collaborative tasks are expected to be executed while interacting with humans in a natural and fluent manner. In this scenario, the robotic system is typically provided with structured tasks to be accomplished, but must also continuously adapt to human activities, commands, and interventions. We propose to tackle these issues by exploiting the concept of cognitive control, introduced in cognitive psychology and neuroscience to describe the executive mechanisms needed to support adaptive responses and complex goal-directed behaviors. Specifically, we rely on a supervisory attentional system to orchestrate the execution of hierarchically organized robotic behaviors. This paradigm seems particularly effective not only for flexible plan execution but also for human-robot interaction, because it directly provides attention mechanisms considered as pivotal for implicit, non-verbal human-human communication. Following this approach, we are currently developing a robotic cognitive control framework enabling collaborative task execution and incremental task learning. In this paper, we provide a uniform overview of the framework illustrating its main features and discussing the potential of the supervisory attentional system paradigm in different scenarios where humans and robots have to collaborate for learning and executing everyday activities.
Collapse
Affiliation(s)
- Riccardo Caccavale
- Dipartimento di Ingegneria Elettrica e Tecnologie dell'Informazione (DIETI), Università degli Studi di Napoli "Federico II"
| | - Alberto Finzi
- Dipartimento di Ingegneria Elettrica e Tecnologie dell'Informazione (DIETI), Università degli Studi di Napoli "Federico II"
| |
Collapse
|
3
|
D'Amelio A, Boccignone G. Gazing at Social Interactions Between Foraging and Decision Theory. Front Neurorobot 2021; 15:639999. [PMID: 33859558 PMCID: PMC8042312 DOI: 10.3389/fnbot.2021.639999] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2020] [Accepted: 03/09/2021] [Indexed: 11/30/2022] Open
Abstract
Finding the underlying principles of social attention in humans seems to be essential for the design of the interaction between natural and artificial agents. Here, we focus on the computational modeling of gaze dynamics as exhibited by humans when perceiving socially relevant multimodal information. The audio-visual landscape of social interactions is distilled into a number of multimodal patches that convey different social value, and we work under the general frame of foraging as a tradeoff between local patch exploitation and landscape exploration. We show that the spatio-temporal dynamics of gaze shifts can be parsimoniously described by Langevin-type stochastic differential equations triggering a decision equation over time. In particular, value-based patch choice and handling is reduced to a simple multi-alternative perceptual decision making that relies on a race-to-threshold between independent continuous-time perceptual evidence integrators, each integrator being associated with a patch.
Collapse
Affiliation(s)
- Alessandro D'Amelio
- PHuSe Lab, Department of Computer Science, Universitá degli Studi di Milano, Milan, Italy
| | - Giuseppe Boccignone
- PHuSe Lab, Department of Computer Science, Universitá degli Studi di Milano, Milan, Italy
| |
Collapse
|
4
|
Kumari R, Jeong JY, Lee BH, Choi KN, Choi K. Topic modelling and social network analysis of publications and patents in humanoid robot technology. J Inf Sci 2019. [DOI: 10.1177/0165551519887878] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
This article presents analysis of data from scientific articles and patents to identify the evolving trends and underlying topics in research on humanoid robots. We used topic modelling based on latent Dirichlet allocation analysis to identify underlying topics in sub-areas in the field. We also used social network analysis to measure the centrality indices of publication keywords to detect important and influential sub-areas and used co-occurrence analysis of keywords to visualise relationships among subfields. The research result is useful to identify evolving topics and areas of current focus in the field of humanoid technology. The results contribute to identify valuable research patterns from publications and to increase understanding of the hidden knowledge themes that are revealed by patents.
Collapse
Affiliation(s)
- Richa Kumari
- Department of Science and Technology Management Policy, University of Science and Technology, South Korea
| | - Jae Yun Jeong
- Department of Science and Technology Management Policy, University of Science and Technology, South Korea
| | - Byeong-Hee Lee
- Department of Science and Technology Management Policy, University of Science and Technology, South Korea; NTIS Center, Korea Institute of Science and Technology Information, South Korea
| | - Kwang-Nam Choi
- NTIS Center, Korea Institute of Science and Technology Information, South Korea
| | - Kiseok Choi
- Department of Science and Technology Management Policy, University of Science and Technology, South Korea; NTIS Center, Korea Institute of Science and Technology Information, South Korea
| |
Collapse
|
5
|
El Hafi L, Isobe S, Tabuchi Y, Katsumata Y, Nakamura H, Fukui T, Matsuo T, Garcia Ricardez GA, Yamamoto M, Taniguchi A, Hagiwara Y, Taniguchi T. System for augmented human–robot interaction through mixed reality and robot training by non-experts in customer service environments. Adv Robot 2019. [DOI: 10.1080/01691864.2019.1694068] [Citation(s) in RCA: 16] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
Affiliation(s)
- L. El Hafi
- College of Information Science and Engineering, Ritsumeikan University, Kusatsu, Japan
| | - S. Isobe
- College of Information Science and Engineering, Ritsumeikan University, Kusatsu, Japan
| | - Y. Tabuchi
- College of Information Science and Engineering, Ritsumeikan University, Kusatsu, Japan
| | - Y. Katsumata
- College of Information Science and Engineering, Ritsumeikan University, Kusatsu, Japan
| | - H. Nakamura
- College of Information Science and Engineering, Ritsumeikan University, Kusatsu, Japan
| | - T. Fukui
- College of Information Science and Engineering, Ritsumeikan University, Kusatsu, Japan
| | - T. Matsuo
- College of Information Science and Engineering, Ritsumeikan University, Kusatsu, Japan
| | - G. A. Garcia Ricardez
- Division of Information Science, Nara Institute of Science and Technology, Ikoma, Japan
| | - M. Yamamoto
- Business Innovation Division, Panasonic Corporation, Osaka, Japan
| | - A. Taniguchi
- College of Information Science and Engineering, Ritsumeikan University, Kusatsu, Japan
| | - Y. Hagiwara
- College of Information Science and Engineering, Ritsumeikan University, Kusatsu, Japan
| | - T. Taniguchi
- College of Information Science and Engineering, Ritsumeikan University, Kusatsu, Japan
| |
Collapse
|
6
|
Caccavale R, Finzi A. Learning attentional regulations for structured tasks execution in robotic cognitive control. Auton Robots 2019. [DOI: 10.1007/s10514-019-09876-x] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
7
|
|
8
|
Habibovic A, Andersson J, Malmsten Lundgren V, Klingegård M, Englund C, Larsson S. External Vehicle Interfaces for Communication with Other Road Users? LECTURE NOTES IN MOBILITY 2019. [DOI: 10.1007/978-3-319-94896-6_9] [Citation(s) in RCA: 16] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/07/2023]
|
9
|
Habibovic A, Lundgren VM, Andersson J, Klingegård M, Lagström T, Sirkka A, Fagerlönn J, Edgren C, Fredriksson R, Krupenia S, Saluäär D, Larsson P. Communicating Intent of Automated Vehicles to Pedestrians. Front Psychol 2018; 9:1336. [PMID: 30131737 PMCID: PMC6090516 DOI: 10.3389/fpsyg.2018.01336] [Citation(s) in RCA: 118] [Impact Index Per Article: 19.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2017] [Accepted: 07/12/2018] [Indexed: 11/25/2022] Open
Abstract
While traffic signals, signs, and road markings provide explicit guidelines for those operating in and around the roadways, some decisions, such as determinations of “who will go first,” are made by implicit negotiations between road users. In such situations, pedestrians are today often dependent on cues in drivers’ behavior such as eye contact, postures, and gestures. With the introduction of more automated functions and the transfer of control from the driver to the vehicle, pedestrians cannot rely on such non-verbal cues anymore. To study how the interaction between pedestrians and automated vehicles (AVs) might look like in the future, and how this might be affected if AVs were to communicate their intent to pedestrians, we designed an external vehicle interface called automated vehicle interaction principle (AVIP) that communicates vehicles’ mode and intent to pedestrians. The interaction was explored in two experiments using a Wizard of Oz approach to simulate automated driving. The first experiment was carried out at a zebra crossing and involved nine pedestrians. While it focused mainly on assessing the usability of the interface, it also revealed initial indications related to pedestrians’ emotions and perceived safety when encountering an AV with/without the interface. The second experiment was carried out in a parking lot and involved 24 pedestrians, which enabled a more detailed assessment of pedestrians’ perceived safety when encountering an AV, both with and without the interface. For comparison purposes, these pedestrians also encountered a conventional vehicle. After a short training course, the interface was deemed easy for the pedestrians to interpret. The pedestrians stated that they felt significantly less safe when they encountered the AV without the interface, compared to the conventional vehicle and the AV with the interface. This suggests that the interface could contribute to a positive experience and improved perceived safety in pedestrian encounters with AVs – something that might be important for general acceptance of AVs. As such, this topic should be further investigated in future studies involving a larger sample and more dynamic conditions.
Collapse
Affiliation(s)
| | | | | | | | | | - Anna Sirkka
- RISE Research Institutes of Sweden, Gothenburg, Sweden
| | | | | | | | | | | | | |
Collapse
|
10
|
|
11
|
Wiese E, Metta G, Wykowska A. Robots As Intentional Agents: Using Neuroscientific Methods to Make Robots Appear More Social. Front Psychol 2017; 8:1663. [PMID: 29046651 PMCID: PMC5632653 DOI: 10.3389/fpsyg.2017.01663] [Citation(s) in RCA: 100] [Impact Index Per Article: 14.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2017] [Accepted: 09/11/2017] [Indexed: 12/30/2022] Open
Abstract
Robots are increasingly envisaged as our future cohabitants. However, while considerable progress has been made in recent years in terms of their technological realization, the ability of robots to interact with humans in an intuitive and social way is still quite limited. An important challenge for social robotics is to determine how to design robots that can perceive the user's needs, feelings, and intentions, and adapt to users over a broad range of cognitive abilities. It is conceivable that if robots were able to adequately demonstrate these skills, humans would eventually accept them as social companions. We argue that the best way to achieve this is using a systematic experimental approach based on behavioral and physiological neuroscience methods such as motion/eye-tracking, electroencephalography, or functional near-infrared spectroscopy embedded in interactive human-robot paradigms. This approach requires understanding how humans interact with each other, how they perform tasks together and how they develop feelings of social connection over time, and using these insights to formulate design principles that make social robots attuned to the workings of the human brain. In this review, we put forward the argument that the likelihood of artificial agents being perceived as social companions can be increased by designing them in a way that they are perceived as intentional agents that activate areas in the human brain involved in social-cognitive processing. We first review literature related to social-cognitive processes and mechanisms involved in human-human interactions, and highlight the importance of perceiving others as intentional agents to activate these social brain areas. We then discuss how attribution of intentionality can positively affect human-robot interaction by (a) fostering feelings of social connection, empathy and prosociality, and by (b) enhancing performance on joint human-robot tasks. Lastly, we describe circumstances under which attribution of intentionality to robot agents might be disadvantageous, and discuss challenges associated with designing social robots that are inspired by neuroscientific principles.
Collapse
Affiliation(s)
- Eva Wiese
- Department of Psychology, George Mason University, Fairfax, VA, United States
| | | | | |
Collapse
|
12
|
Emotional metacontrol of attention: Top-down modulation of sensorimotor processes in a robotic visual search task. PLoS One 2017; 12:e0184960. [PMID: 28934291 PMCID: PMC5608313 DOI: 10.1371/journal.pone.0184960] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2017] [Accepted: 09/05/2017] [Indexed: 11/19/2022] Open
Abstract
Emotions play a significant role in internal regulatory processes. In this paper, we advocate four key ideas. First, novelty detection can be grounded in the sensorimotor experience and allow higher order appraisal. Second, cognitive processes, such as those involved in self-assessment, influence emotional states by eliciting affects like boredom and frustration. Third, emotional processes such as those triggered by self-assessment influence attentional processes. Last, close emotion-cognition interactions implement an efficient feedback loop for the purpose of top-down behavior regulation. The latter is what we call ‘Emotional Metacontrol’. We introduce a model based on artificial neural networks. This architecture is used to control a robotic system in a visual search task. The emotional metacontrol intervenes to bias the robot visual attention during active object recognition. Through a behavioral and statistical analysis, we show that this mechanism increases the robot performance and fosters the exploratory behavior to avoid deadlocks.
Collapse
|
13
|
A salient region detection model combining background distribution measure for indoor robots. PLoS One 2017; 12:e0180519. [PMID: 28742089 PMCID: PMC5524399 DOI: 10.1371/journal.pone.0180519] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2017] [Accepted: 06/17/2017] [Indexed: 11/19/2022] Open
Abstract
Vision system plays an important role in the field of indoor robot. Saliency detection methods, capturing regions that are perceived as important, are used to improve the performance of visual perception system. Most of state-of-the-art methods for saliency detection, performing outstandingly in natural images, cannot work in complicated indoor environment. Therefore, we propose a new method comprised of graph-based RGB-D segmentation, primary saliency measure, background distribution measure, and combination. Besides, region roundness is proposed to describe the compactness of a region to measure background distribution more robustly. To validate the proposed approach, eleven influential methods are compared on the DSD and ECSSD dataset. Moreover, we build a mobile robot platform for application in an actual environment, and design three different kinds of experimental constructions that are different viewpoints, illumination variations and partial occlusions. Experimental results demonstrate that our model outperforms existing methods and is useful for indoor mobile robots.
Collapse
|
14
|
Vignolo A, Noceti N, Rea F, Sciutti A, Odone F, Sandini G. Detecting Biological Motion for Human–Robot Interaction: A Link between Perception and Action. Front Robot AI 2017. [DOI: 10.3389/frobt.2017.00014] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/07/2023] Open
|
15
|
Abstract
Newly emerging robotics applications for domestic or entertainment purposes are slowly introducing autonomous robots into society at large. A critical capability of such robots is their ability to interact with humans, and in particular, untrained users. In this paper we explore the hypothesis that people will intuitively interact with robots in a natural social manner provided the robot can perceive, interpret, and appropriately respond with familiar human social cues. Two experiments are presented where naive human subjects interact with an anthropomorphic robot. We present evidence for mutual regulation and entrainment of the interaction, and we discuss how this benefits the interaction as a whole.
Collapse
Affiliation(s)
- Cynthia Breazeal
- MIT Media Lab 77 Massachusetts Ave NE18-5fl Cambridge, MA 02139, USA
| |
Collapse
|
16
|
Castellano G, Leite I, Paiva A. Detecting perceived quality of interaction with a robot using contextual features. Auton Robots 2016. [DOI: 10.1007/s10514-016-9592-y] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
17
|
Castellano G, Leite I, Pereira A, Martinho C, Paiva A, Mcowan PW. Context-Sensitive Affect Recognition for a Robotic Game Companion. ACM T INTERACT INTEL 2014. [DOI: 10.1145/2622615] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
Abstract
Social perception abilities are among the most important skills necessary for robots to engage humans in natural forms of interaction. Affect-sensitive robots are more likely to be able to establish and maintain believable interactions over extended periods of time. Nevertheless, the integration of affect recognition frameworks in real-time human-robot interaction scenarios is still underexplored. In this article, we propose and evaluate a context-sensitive affect recognition framework for a robotic game companion for children. The robot can automatically detect affective states experienced by children in an interactive chess game scenario. The affect recognition framework is based on the automatic extraction of task features and social interaction-based features. Vision-based indicators of the children’s nonverbal behaviour are merged with contextual features related to the game and the interaction and given as input to support vector machines to create a context-sensitive multimodal system for affect recognition. The affect recognition framework is fully integrated in an architecture for adaptive human-robot interaction. Experimental evaluation showed that children’s affect can be successfully predicted using a combination of behavioural and contextual data related to the game and the interaction with the robot. It was found that contextual data alone can be used to successfully predict a subset of affective dimensions, such as interest toward the robot. Experiments also showed that engagement with the robot can be predicted using information about the user’s valence, interest and anticipatory behaviour. These results provide evidence that social engagement can be modelled as a state consisting of affect and attention components in the context of the interaction.
Collapse
Affiliation(s)
| | | | - André Pereira
- INESC-ID and Instituto Superior Técnico, Technical University of Lisbon, Porto Salvo, Portugal
| | - Carlos Martinho
- INESC-ID and Instituto Superior Técnico, Technical University of Lisbon, Porto Salvo, Portugal
| | - Ana Paiva
- INESC-ID and Instituto Superior Técnico, Technical University of Lisbon, Porto Salvo, Portugal
| | | |
Collapse
|
18
|
Deficient gaze pattern during virtual multiparty conversation in patients with schizophrenia. Comput Biol Med 2014; 49:60-6. [DOI: 10.1016/j.compbiomed.2014.03.012] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2013] [Revised: 03/08/2014] [Accepted: 03/28/2014] [Indexed: 11/23/2022]
|
19
|
Yan H, Ang MH, Poo AN. A Survey on Perception Methods for Human–Robot Interaction in Social Robots. Int J Soc Robot 2013. [DOI: 10.1007/s12369-013-0199-6] [Citation(s) in RCA: 77] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
20
|
Yücel Z, Salah AA, Meriçli Ç, Meriçli T, Valenti R, Gevers T. Joint attention by gaze interpolation and saliency. IEEE TRANSACTIONS ON CYBERNETICS 2013; 43:829-842. [PMID: 23047879 DOI: 10.1109/tsmcb.2012.2216979] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
Joint attention, which is the ability of coordination of a common point of reference with the communicating party, emerges as a key factor in various interaction scenarios. This paper presents an image-based method for establishing joint attention between an experimenter and a robot. The precise analysis of the experimenter's eye region requires stability and high-resolution image acquisition, which is not always available. We investigate regression-based interpolation of the gaze direction from the head pose of the experimenter, which is easier to track. Gaussian process regression and neural networks are contrasted to interpolate the gaze direction. Then, we combine gaze interpolation with image-based saliency to improve the target point estimates and test three different saliency schemes. We demonstrate the proposed method on a human-robot interaction scenario. Cross-subject evaluations, as well as experiments under adverse conditions (such as dimmed or artificial illumination or motion blur), show that our method generalizes well and achieves rapid gaze estimation for establishing joint attention.
Collapse
Affiliation(s)
- Zeynep Yücel
- Intelligent Robotics and Communication Laboratories, Advanced Telecommunications Research Institute International, Kyoto 619-0288, Japan.
| | | | | | | | | | | |
Collapse
|
21
|
Ferreira JF, Lobo J, Bessière P, Castelo-Branco M, Dias J. A Bayesian framework for active artificial perception. IEEE TRANSACTIONS ON CYBERNETICS 2013; 43:699-711. [PMID: 23014760 DOI: 10.1109/tsmcb.2012.2214477] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
In this paper, we present a Bayesian framework for the active multimodal perception of 3-D structure and motion. The design of this framework finds its inspiration in the role of the dorsal perceptual pathway of the human brain. Its composing models build upon a common egocentric spatial configuration that is naturally fitting for the integration of readings from multiple sensors using a Bayesian approach. In the process, we will contribute with efficient and robust probabilistic solutions for cyclopean geometry-based stereovision and auditory perception based only on binaural cues, modeled using a consistent formalization that allows their hierarchical use as building blocks for the multimodal sensor fusion framework. We will explicitly or implicitly address the most important challenges of sensor fusion using this framework, for vision, audition, and vestibular sensing. Moreover, interaction and navigation require maximal awareness of spatial surroundings, which, in turn, is obtained through active attentional and behavioral exploration of the environment. The computational models described in this paper will support the construction of a simultaneously flexible and powerful robotic implementation of multimodal active perception to be used in real-world applications, such as human-machine interaction or mobile robot navigation.
Collapse
|
22
|
Ferreira JF, Tsiourti C, Dias J. Learning emergent behaviours for a hierarchical Bayesian framework for active robotic perception. Cogn Process 2012; 13 Suppl 1:S155-9. [PMID: 22806665 DOI: 10.1007/s10339-012-0481-9] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
In this research work, we contribute with a behaviour learning process for a hierarchical Bayesian framework for multimodal active perception, devised to be emergent, scalable and adaptive. This framework is composed by models built upon a common spatial configuration for encoding perception and action that is naturally fitting for the integration of readings from multiple sensors, using a Bayesian approach devised in previous work. The proposed learning process is shown to reproduce goal-dependent human-like active perception behaviours by learning model parameters (referred to as "attentional sets") for different free-viewing and active search tasks. Learning was performed by presenting several 3D audiovisual virtual scenarios using a head-mounted display, while logging the spatial distribution of fixations of the subject (in 2D, on left and right images, and in 3D space), data which are consequently used as the training set for the framework. As a consequence, the hierarchical Bayesian framework adequately implements high-level behaviour resulting from low-level interaction of simpler building blocks by using the attentional sets learned for each task, and is able to change these attentional sets "on the fly," allowing the implementation of goal-dependent behaviours (i.e., top-down influences).
Collapse
|
23
|
Omrčen D, Ude A. Redundancy Control of a Humanoid Head for Foveation and Three-Dimensional Object Tracking: A Virtual Mechanism Approach. Adv Robot 2012. [DOI: 10.1163/016918610x534303] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
Affiliation(s)
- Damir Omrčen
- a Jozef Stefan Institute, Jamova cesta 39, 1000 Ljubljana, Slovenia
| | - Aleš Ude
- b Jozef Stefan Institute, Jamova cesta 39, 1000 Ljubljana, Slovenia; ATR Computational Neuroscience Laboratories, 2-2-2 Hikaridai, Seika-cho, Sokaro-gun, Kyoto 619-0288, Japan
| |
Collapse
|
24
|
BREAZEAL CYNTHIA, BROOKS ANDREW, GRAY JESSE, HOFFMAN GUY, KIDD CORY, LEE HANS, LIEBERMAN JEFF, LOCKERD ANDREA, CHILONGO DAVID. TUTELAGE AND COLLABORATION FOR HUMANOID ROBOTS. INT J HUM ROBOT 2012. [DOI: 10.1142/s0219843604000150] [Citation(s) in RCA: 94] [Impact Index Per Article: 7.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
This paper presents an overview of our work towards building socially intelligent, cooperative humanoid robots that can work and learn in partnership with people. People understand each other in social terms, allowing them to engage others in a variety of complex social interactions including communication, social learning, and cooperation. We present our theoretical framework that is a novel combination of Joint Intention Theory and Situated Learning Theory and demonstrate how this framework can be applied to develop our sociable humanoid robot, Leonardo. We demonstrate the robot's ability to learn quickly and effectively from natural human instruction using gesture and dialog, and then cooperate to perform a learned task jointly with a person. Such issues must be addressed to enable many new and exciting applications for robots that require them to play a long-term role in people's daily lives.
Collapse
Affiliation(s)
- CYNTHIA BREAZEAL
- MIT Media Lab, Robotic Life Group, 77 Massachusetts Ave E15-468, Cambridge, MA 02139, USA
| | - ANDREW BROOKS
- MIT Media Lab, Robotic Life Group, 77 Massachusetts Ave E15-468, Cambridge, MA 02139, USA
| | - JESSE GRAY
- MIT Media Lab, Robotic Life Group, 77 Massachusetts Ave E15-468, Cambridge, MA 02139, USA
| | - GUY HOFFMAN
- MIT Media Lab, Robotic Life Group, 77 Massachusetts Ave E15-468, Cambridge, MA 02139, USA
| | - CORY KIDD
- MIT Media Lab, Robotic Life Group, 77 Massachusetts Ave E15-468, Cambridge, MA 02139, USA
| | - HANS LEE
- MIT Media Lab, Robotic Life Group, 77 Massachusetts Ave E15-468, Cambridge, MA 02139, USA
| | - JEFF LIEBERMAN
- MIT Media Lab, Robotic Life Group, 77 Massachusetts Ave E15-468, Cambridge, MA 02139, USA
| | - ANDREA LOCKERD
- MIT Media Lab, Robotic Life Group, 77 Massachusetts Ave E15-468, Cambridge, MA 02139, USA
| | - DAVID CHILONGO
- MIT Media Lab, Robotic Life Group, 77 Massachusetts Ave E15-468, Cambridge, MA 02139, USA
| |
Collapse
|
25
|
|
26
|
WON WOONGJAE, YEO JIYOUNG, BAN SANGWOO, LEE MINHO. BIOLOGICALLY MOTIVATED INCREMENTAL OBJECT PERCEPTION BASED ON SELECTIVE ATTENTION. INT J PATTERN RECOGN 2011. [DOI: 10.1142/s021800140700596x] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
In this paper, we propose an object selective attention and perception system, which was implemented by integrating a specific object preferable attention model with an incremental object perception model. The object oriented attention model can selectively pay attention to the candidates of an object in natural scenes based on a bottom-up selective attention model in conjunction with a top-down biased attention mechanism for a specific object. A generative model based on an incremental Bayesian parameter estimation is considered in order to perceive arbitrary objects in the attended areas. Combining an object oriented attention model with general object perception model, the developed system cannot only pay attention to a specific target object but can also memorize the characteristics of task nonspecific objects in an incremental manner. Experimental results show that the developed system generates good performance in successfully focusing on the target objects as well as incrementally perceiving objects in natural scenes.
Collapse
Affiliation(s)
- WOONG-JAE WON
- Department of Mechatronics Intelligent Vehicle System Research Team, Daegu Gyeongbuk Institute of Science and Technology, 711 Hosan-dong, Dalseo-gu, Daegu 704-230, Korea
| | - JIYOUNG YEO
- Department of Sensor Engineering, Kyungpook National University, 1370 Sankyuk-Dong, Puk-Gu, Taegu 702-701, Korea
| | - SANG-WOO BAN
- Department of Information and Communication Engineering, Dongguk University 707 Seokjang-Dong, Gyeongju, Gyeongbuk, 780-714, Korea
| | - MINHO LEE
- School of Electrical Engineering and Computer Science, Kyungpook National University, 1370 Sankyuk-Dong, Puk-Gu, Taegu 702-701, Korea
| |
Collapse
|
27
|
|
28
|
Yu Y, Mann GKI, Gosine RG. An Object-Based Visual Attention Model for Robotic Applications. ACTA ACUST UNITED AC 2010; 40:1398-412. [DOI: 10.1109/tsmcb.2009.2038895] [Citation(s) in RCA: 48] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Affiliation(s)
- Yuanlong Yu
- Faculty of Engineering and Applied Science, Memorial University of Newfoundland, St. John's, NL A1B 3X5, Canada.
| | | | | |
Collapse
|
29
|
Rajruangrabin J, Popa DO. Robot Head Motion Control with an Emphasis on Realism of Neck–Eye Coordination during Object Tracking. J INTELL ROBOT SYST 2010. [DOI: 10.1007/s10846-010-9468-x] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
30
|
Begum M, Karray F, Mann GKI, Gosine RG. A probabilistic model of overt visual attention for cognitive robots. IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS. PART B, CYBERNETICS : A PUBLICATION OF THE IEEE SYSTEMS, MAN, AND CYBERNETICS SOCIETY 2010; 40:1305-18. [PMID: 20089477 DOI: 10.1109/tsmcb.2009.2037511] [Citation(s) in RCA: 23] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Visual attention is one of the major requirements for a robot to serve as a cognitive companion for human. The robotic visual attention is mostly concerned with overt attention which accompanies head and eye movements of a robot. In this case, each movement of the camera head triggers a number of events, namely transformation of the camera and the image coordinate systems, change of content of the visual field, and partial appearance of the stimuli. All of these events contribute to the reduction in probability of meaningful identification of the next focus of attention. These events are specific to overt attention with head movement and, therefore, their effects are not addressed in the classical models of covert visual attention. This paper proposes a Bayesian model as a robot-centric solution for the overt visual attention problem. The proposed model, while taking inspiration from the primates visual attention mechanism, guides a robot to direct its camera toward behaviorally relevant and/or visually demanding stimuli. A particle filter implementation of this model addresses the challenges involved in overt attention with head movement. Experimental results demonstrate the performance of the proposed model.
Collapse
Affiliation(s)
- Momotaz Begum
- Department of Electrical and Computer Engineering, University of Waterloo, Waterloo, ON N2L 3G1, Canada.
| | | | | | | |
Collapse
|
31
|
Green SA, Billinghurst M, Chen X, Chase JG. Human-Robot Collaboration: A Literature Review and Augmented Reality Approach in Design. INT J ADV ROBOT SYST 2008. [DOI: 10.5772/5664] [Citation(s) in RCA: 153] [Impact Index Per Article: 9.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022] Open
Abstract
NASA's vision for space exploration stresses the cultivation of human-robotic systems. Similar systems are also envisaged for a variety of hazardous earthbound applications such as urban search and rescue. Recent research has pointed out that to reduce human workload, costs, fatigue driven error and risk, intelligent robotic systems will need to be a significant part of mission design. However, little attention has been paid to joint human-robot teams. Making human-robot collaboration natural and efficient is crucial. In particular, grounding, situational awareness, a common frame of reference and spatial referencing are vital in effective communication and collaboration. Augmented Reality (AR), the overlaying of computer graphics onto the real worldview, can provide the necessary means for a human-robotic system to fulfill these requirements for effective collaboration. This article reviews the field of human-robot interaction and augmented reality, investigates the potential avenues for creating natural human-robot collaboration through spatial dialogue utilizing AR and proposes a holistic architectural design for human-robot collaboration.
Collapse
Affiliation(s)
- Scott A. Green
- Department of Mechanical Engineering, University of Canterbury, Christchurch, New Zealand
- Human Interface Technology Laboratory, New Zealand (HITLab NZ), Christchurch, New Zealand
| | - Mark Billinghurst
- Human Interface Technology Laboratory, New Zealand (HITLab NZ), Christchurch, New Zealand
| | - XiaoQi Chen
- Department of Mechanical Engineering, University of Canterbury, Christchurch, New Zealand
| | - J. Geoffrey Chase
- Department of Mechanical Engineering, University of Canterbury, Christchurch, New Zealand
| |
Collapse
|
32
|
Maini ES, Manfredi L, Laschi C, Dario P. Bioinspired velocity control of fast gaze shifts on a robotic anthropomorphic head. Auton Robots 2007. [DOI: 10.1007/s10514-007-9078-z] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
33
|
Spexard T, Hanheide M, Sagerer G. Human-Oriented Interaction With an Anthropomorphic Robot. IEEE T ROBOT 2007. [DOI: 10.1109/tro.2007.904903] [Citation(s) in RCA: 50] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
34
|
Schaal S. The New Robotics-towards human-centered machines. HFSP JOURNAL 2007; 1:115-26. [PMID: 19404417 DOI: 10.2976/1.2748612] [Citation(s) in RCA: 53] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/26/2006] [Accepted: 05/21/2007] [Indexed: 11/19/2022]
Abstract
Research in robotics has moved away from its primary focus on industrial applications. The New Robotics is a vision that has been developed in past years by our own university and many other national and international research institutions and addresses how increasingly more human-like robots can live among us and take over tasks where our current society has shortcomings. Elder care, physical therapy, child education, search and rescue, and general assistance in daily life situations are some of the examples that will benefit from the New Robotics in the near future. With these goals in mind, research for the New Robotics has to embrace a broad interdisciplinary approach, ranging from traditional mathematical issues of robotics to novel issues in psychology, neuroscience, and ethics. This paper outlines some of the important research problems that will need to be resolved to make the New Robotics a reality.
Collapse
Affiliation(s)
- Stefan Schaal
- Computer Science and Neuroscience, University of Southern California, 3710 S. McClintock Avenue-RTH 401, Los Angeles, California 90089-2905 and ATR Computational Neuroscience Laboratories, 2-2 Hikaridai, Seika-cho, Soraku-gun, Kyoto 619-02
| |
Collapse
|
35
|
Breazeal C, Buchsbaum D, Gray J, Gatenby D, Blumberg B. Learning from and about others: towards using imitation to bootstrap the social understanding of others by robots. ARTIFICIAL LIFE 2005; 11:31-62. [PMID: 15811219 DOI: 10.1162/1064546053278955] [Citation(s) in RCA: 52] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/24/2023]
Abstract
We want to build robots capable of rich social interactions with humans, including natural communication and cooperation. This work explores how imitation as a social learning and teaching process may be applied to building socially intelligent robots, and summarizes our progress toward building a robot capable of learning how to imitate facial expressions from simple imitative games played with a human, using biologically inspired mechanisms. It is possible for the robot to bootstrap from this imitative ability to infer the affective reaction of the human with whom it interacts and then use this affective assessment to guide its subsequent behavior. Our approach is heavily influenced by the ways human infants learn to communicate with their caregivers and come to understand the actions and expressive behavior of others in intentional and motivational terms. Specifically, our approach is guided by the hypothesis that imitative interactions between infant and caregiver, starting with facial mimicry, are a significant stepping-stone to developing appropriate social behavior, to predicting others' actions, and ultimately to understanding people as social beings.
Collapse
Affiliation(s)
- Cynthia Breazeal
- MIT Media Lab, 77 Massachusetts Avenue, NE18, 5th floor, Cambridge, MA 02142, USA.
| | | | | | | | | |
Collapse
|
36
|
|