1
|
Priorelli M, Stoianov IP. Dynamic planning in hierarchical active inference. Neural Netw 2025; 185:107075. [PMID: 39817980 DOI: 10.1016/j.neunet.2024.107075] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2024] [Revised: 11/13/2024] [Accepted: 12/18/2024] [Indexed: 01/18/2025]
Abstract
By dynamic planning, we refer to the ability of the human brain to infer and impose motor trajectories related to cognitive decisions. A recent paradigm, active inference, brings fundamental insights into the adaptation of biological organisms, constantly striving to minimize prediction errors to restrict themselves to life-compatible states. Over the past years, many studies have shown how human and animal behaviors could be explained in terms of active inference - either as discrete decision-making or continuous motor control - inspiring innovative solutions in robotics and artificial intelligence. Still, the literature lacks a comprehensive outlook on effectively planning realistic actions in changing environments. Setting ourselves the goal of modeling complex tasks such as tool use, we delve into the topic of dynamic planning in active inference, keeping in mind two crucial aspects of biological behavior: the capacity to understand and exploit affordances for object manipulation, and to learn the hierarchical interactions between the self and the environment, including other agents. We start from a simple unit and gradually describe more advanced structures, comparing recently proposed design choices and providing basic examples. This study distances itself from traditional views centered on neural networks and reinforcement learning, and points toward a yet unexplored direction in active inference: hybrid representations in hierarchical models.
Collapse
Affiliation(s)
- Matteo Priorelli
- Institute of Cognitive Sciences and Technologies, National Research Council, Padova, Italy; Sapienza University of Rome, Rome, Italy
| | - Ivilin Peev Stoianov
- Institute of Cognitive Sciences and Technologies, National Research Council, Padova, Italy.
| |
Collapse
|
2
|
Van de Maele T, Verbelen T, Mazzaglia P, Ferraro S, Dhoedt B. Object-Centric Scene Representations Using Active Inference. Neural Comput 2024; 36:677-704. [PMID: 38457764 DOI: 10.1162/neco_a_01637] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Accepted: 10/17/2023] [Indexed: 03/10/2024]
Abstract
Representing a scene and its constituent objects from raw sensory data is a core ability for enabling robots to interact with their environment. In this letter, we propose a novel approach for scene understanding, leveraging an object-centric generative model that enables an agent to infer object category and pose in an allocentric reference frame using active inference, a neuro-inspired framework for action and perception. For evaluating the behavior of an active vision agent, we also propose a new benchmark where, given a target viewpoint of a particular object, the agent needs to find the best matching viewpoint given a workspace with randomly positioned objects in 3D. We demonstrate that our active inference agent is able to balance epistemic foraging and goal-driven behavior, and quantitatively outperforms both supervised and reinforcement learning baselines by more than a factor of two in terms of success rate.
Collapse
Affiliation(s)
| | - Tim Verbelen
- VERSES AI Research Lab, Los Angeles, CA 90016, U.S.A.
| | | | | | | |
Collapse
|
3
|
Ferraro S, Van de Maele T, Verbelen T, Dhoedt B. Symmetry and complexity in object-centric deep active inference models. Interface Focus 2023; 13:20220077. [PMID: 37065264 PMCID: PMC10102726 DOI: 10.1098/rsfs.2022.0077] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2022] [Accepted: 03/01/2023] [Indexed: 04/18/2023] Open
Abstract
Humans perceive and interact with hundreds of objects every day. In doing so, they need to employ mental models of these objects and often exploit symmetries in the object's shape and appearance in order to learn generalizable and transferable skills. Active inference is a first principles approach to understanding and modelling sentient agents. It states that agents entertain a generative model of their environment, and learn and act by minimizing an upper bound on their surprisal, i.e. their free energy. The free energy decomposes into an accuracy and complexity term, meaning that agents favour the least complex model that can accurately explain their sensory observations. In this paper, we investigate how inherent symmetries of particular objects also emerge as symmetries in the latent state space of the generative model learnt under deep active inference. In particular, we focus on object-centric representations, which are trained from pixels to predict novel object views as the agent moves its viewpoint. First, we investigate the relation between model complexity and symmetry exploitation in the state space. Second, we do a principal component analysis to demonstrate how the model encodes the principal axis of symmetry of the object in the latent space. Finally, we also demonstrate how more symmetrical representations can be exploited for better generalization in the context of manipulation.
Collapse
Affiliation(s)
- Stefano Ferraro
- IDLab, Department of Information Technology, Ghent University–imec, Ghent, Belgium
| | - Toon Van de Maele
- IDLab, Department of Information Technology, Ghent University–imec, Ghent, Belgium
| | - Tim Verbelen
- IDLab, Department of Information Technology, Ghent University–imec, Ghent, Belgium
| | - Bart Dhoedt
- IDLab, Department of Information Technology, Ghent University–imec, Ghent, Belgium
| |
Collapse
|
4
|
Sun H, Zhu F, Li Y, Zhao P, Kong Y, Wang J, Wan Y, Fu S. Viewpoint planning with transition management for active object recognition. Front Neurorobot 2023; 17:1093132. [PMID: 36910268 PMCID: PMC9998679 DOI: 10.3389/fnbot.2023.1093132] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2022] [Accepted: 02/08/2023] [Indexed: 03/14/2023] Open
Abstract
Active object recognition (AOR) provides a paradigm where an agent can capture additional evidence by purposefully changing its viewpoint to improve the quality of recognition. One of the most concerned problems in AOR is viewpoint planning (VP) which refers to developing a policy to determine the next viewpoints of the agent. A research trend is to solve the VP problem with reinforcement learning, namely to use the viewpoint transitions explored by the agent to train the VP policy. However, most research discards the trained transitions, which may lead to an inefficient use of the explored transitions. To solve this challenge, we present a novel VP method with transition management based on reinforcement learning, which can reuse the explored viewpoint transitions. To be specific, a learning framework of the VP policy is first established via the deterministic policy gradient theory, which provides an opportunity to reuse the explored transitions. Then, we design a scheme of viewpoint transition management that can store the explored transitions and decide which transitions are used for the policy learning. Finally, within the framework, we develop an algorithm based on twin delayed deep deterministic policy gradient and the designed scheme to train the VP policy. Experiments on the public and challenging dataset GERMS show the effectiveness of our method in comparison with several competing approaches.
Collapse
Affiliation(s)
- Haibo Sun
- Faculty of Robot Science and Engineering, Northeastern University, Shenyang, China.,Key Laboratory of Opto-Electronic Information Processing, Chinese Academy of Sciences, Shenyang, China.,Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang, China.,Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang, China
| | - Feng Zhu
- Key Laboratory of Opto-Electronic Information Processing, Chinese Academy of Sciences, Shenyang, China.,Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang, China.,Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang, China
| | - Yangyang Li
- Key Laboratory of Opto-Electronic Information Processing, Chinese Academy of Sciences, Shenyang, China.,Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang, China.,Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang, China.,University of Chinese Academy of Sciences, Beijing, China
| | - Pengfei Zhao
- Key Laboratory of Opto-Electronic Information Processing, Chinese Academy of Sciences, Shenyang, China.,Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang, China.,Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang, China.,University of Chinese Academy of Sciences, Beijing, China
| | - Yanzi Kong
- Key Laboratory of Opto-Electronic Information Processing, Chinese Academy of Sciences, Shenyang, China.,Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang, China.,Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang, China.,University of Chinese Academy of Sciences, Beijing, China
| | - Jianyu Wang
- Faculty of Robot Science and Engineering, Northeastern University, Shenyang, China.,Key Laboratory of Opto-Electronic Information Processing, Chinese Academy of Sciences, Shenyang, China.,Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang, China.,Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang, China
| | - Yingcai Wan
- Faculty of Robot Science and Engineering, Northeastern University, Shenyang, China
| | - Shuangfei Fu
- Key Laboratory of Opto-Electronic Information Processing, Chinese Academy of Sciences, Shenyang, China.,Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang, China.,Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang, China
| |
Collapse
|