1
|
Priorelli M, Stoianov IP. Dynamic planning in hierarchical active inference. Neural Netw 2025; 185:107075. [PMID: 39817980 DOI: 10.1016/j.neunet.2024.107075] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2024] [Revised: 11/13/2024] [Accepted: 12/18/2024] [Indexed: 01/18/2025]
Abstract
By dynamic planning, we refer to the ability of the human brain to infer and impose motor trajectories related to cognitive decisions. A recent paradigm, active inference, brings fundamental insights into the adaptation of biological organisms, constantly striving to minimize prediction errors to restrict themselves to life-compatible states. Over the past years, many studies have shown how human and animal behaviors could be explained in terms of active inference - either as discrete decision-making or continuous motor control - inspiring innovative solutions in robotics and artificial intelligence. Still, the literature lacks a comprehensive outlook on effectively planning realistic actions in changing environments. Setting ourselves the goal of modeling complex tasks such as tool use, we delve into the topic of dynamic planning in active inference, keeping in mind two crucial aspects of biological behavior: the capacity to understand and exploit affordances for object manipulation, and to learn the hierarchical interactions between the self and the environment, including other agents. We start from a simple unit and gradually describe more advanced structures, comparing recently proposed design choices and providing basic examples. This study distances itself from traditional views centered on neural networks and reinforcement learning, and points toward a yet unexplored direction in active inference: hybrid representations in hierarchical models.
Collapse
Affiliation(s)
- Matteo Priorelli
- Institute of Cognitive Sciences and Technologies, National Research Council, Padova, Italy; Sapienza University of Rome, Rome, Italy
| | - Ivilin Peev Stoianov
- Institute of Cognitive Sciences and Technologies, National Research Council, Padova, Italy.
| |
Collapse
|
2
|
Brucklacher M, Pezzulo G, Mannella F, Galati G, Pennartz CMA. Learning to segment self-generated from externally caused optic flow through sensorimotor mismatch circuits. Neural Netw 2025; 181:106716. [PMID: 39383679 DOI: 10.1016/j.neunet.2024.106716] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2024] [Revised: 06/09/2024] [Accepted: 09/07/2024] [Indexed: 10/11/2024]
Abstract
Efficient sensory detection requires the capacity to ignore task-irrelevant information, for example when optic flow patterns created by egomotion need to be disentangled from object perception. To investigate how this is achieved in the visual system, predictive coding with sensorimotor mismatch detection is an attractive starting point. Indeed, experimental evidence for sensorimotor mismatch signals in early visual areas exists, but it is not understood how they are integrated into cortical networks that perform input segmentation and categorization. Our model advances a biologically plausible solution by extending predictive coding models with the ability to distinguish self-generated from externally caused optic flow. We first show that a simple three neuron circuit produces experience-dependent sensorimotor mismatch responses, in agreement with calcium imaging data from mice. This microcircuit is then integrated into a neural network with two generative streams. The motor-to-visual stream consists of parallel microcircuits between motor and visual areas and learns to spatially predict optic flow resulting from self-motion. The second stream bidirectionally connects a motion-selective higher visual area (mHVA) to V1, assigning a crucial role to the abundant feedback connections to V1: the maintenance of a generative model of externally caused optic flow. In the model, area mHVA learns to segment moving objects from the background, and facilitates object categorization. Based on shared neurocomputational principles across species, the model also maps onto primate visual cortex. Our work extends Hebbian predictive coding to sensorimotor settings, in which the agent actively moves - and learns to predict the consequences of its own movements.
Collapse
Affiliation(s)
- Matthias Brucklacher
- Cognitive and Systems Neuroscience, University of Amsterdam, 1098XH Amsterdam, Netherlands.
| | - Giovanni Pezzulo
- Institute of Cognitive Sciences and Technologies, National Research Council, 00196 Rome, Italy
| | - Francesco Mannella
- Institute of Cognitive Sciences and Technologies, National Research Council, 00196 Rome, Italy
| | - Gaspare Galati
- Brain Imaging Laboratory, Department of Psychology, Sapienza University, 00185 Rome, Italy
| | - Cyriel M A Pennartz
- Cognitive and Systems Neuroscience, University of Amsterdam, 1098XH Amsterdam, Netherlands
| |
Collapse
|
3
|
Oliviers G, Bogacz R, Meulemans A. Learning probability distributions of sensory inputs with Monte Carlo predictive coding. PLoS Comput Biol 2024; 20:e1012532. [PMID: 39475902 PMCID: PMC11524488 DOI: 10.1371/journal.pcbi.1012532] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2024] [Accepted: 10/01/2024] [Indexed: 11/02/2024] Open
Abstract
It has been suggested that the brain employs probabilistic generative models to optimally interpret sensory information. This hypothesis has been formalised in distinct frameworks, focusing on explaining separate phenomena. On one hand, classic predictive coding theory proposed how the probabilistic models can be learned by networks of neurons employing local synaptic plasticity. On the other hand, neural sampling theories have demonstrated how stochastic dynamics enable neural circuits to represent the posterior distributions of latent states of the environment. These frameworks were brought together by variational filtering that introduced neural sampling to predictive coding. Here, we consider a variant of variational filtering for static inputs, to which we refer as Monte Carlo predictive coding (MCPC). We demonstrate that the integration of predictive coding with neural sampling results in a neural network that learns precise generative models using local computation and plasticity. The neural dynamics of MCPC infer the posterior distributions of the latent states in the presence of sensory inputs, and can generate likely inputs in their absence. Furthermore, MCPC captures the experimental observations on the variability of neural activity during perceptual tasks. By combining predictive coding and neural sampling, MCPC can account for both sets of neural data that previously had been explained by these individual frameworks.
Collapse
Affiliation(s)
- Gaspard Oliviers
- MRC Brain Network Dynamics Unit, Nuffield Department of Clinical Neurosciences, University of Oxford, Oxford, United Kingdom
| | - Rafal Bogacz
- MRC Brain Network Dynamics Unit, Nuffield Department of Clinical Neurosciences, University of Oxford, Oxford, United Kingdom
| | | |
Collapse
|
4
|
Malloy T, Gonzalez C. Applying Generative Artificial Intelligence to cognitive models of decision making. Front Psychol 2024; 15:1387948. [PMID: 38765837 PMCID: PMC11100990 DOI: 10.3389/fpsyg.2024.1387948] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/19/2024] [Accepted: 04/12/2024] [Indexed: 05/22/2024] Open
Abstract
Introduction Generative Artificial Intelligence has made significant impacts in many fields, including computational cognitive modeling of decision making, although these applications have not yet been theoretically related to each other. This work introduces a categorization of applications of Generative Artificial Intelligence to cognitive models of decision making. Methods This categorization is used to compare the existing literature and to provide insight into the design of an ablation study to evaluate our proposed model in three experimental paradigms. These experiments used for model comparison involve modeling human learning and decision making based on both visual information and natural language, in tasks that vary in realism and complexity. This comparison of applications takes as its basis Instance-Based Learning Theory, a theory of experiential decision making from which many models have emerged and been applied to a variety of domains and applications. Results The best performing model from the ablation we performed used a generative model to both create memory representations as well as predict participant actions. The results of this comparison demonstrates the importance of generative models in both forming memories and predicting actions in decision-modeling research. Discussion In this work, we present a model that integrates generative and cognitive models, using a variety of stimuli, applications, and training methods. These results can provide guidelines for cognitive modelers and decision making researchers interested in integrating Generative AI into their methods.
Collapse
Affiliation(s)
- Tyler Malloy
- Dynamic Decision Making Laboratory, Department of Social and Decision Sciences, Dietrich College, Carnegie Mellon University, Pittsburgh, PA, United States
| | | |
Collapse
|
5
|
Ambrogioni L. In Search of Dispersed Memories: Generative Diffusion Models Are Associative Memory Networks. ENTROPY (BASEL, SWITZERLAND) 2024; 26:381. [PMID: 38785630 PMCID: PMC11119823 DOI: 10.3390/e26050381] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/12/2024] [Revised: 04/19/2024] [Accepted: 04/25/2024] [Indexed: 05/25/2024]
Abstract
Uncovering the mechanisms behind long-term memory is one of the most fascinating open problems in neuroscience and artificial intelligence. Artificial associative memory networks have been used to formalize important aspects of biological memory. Generative diffusion models are a type of generative machine learning techniques that have shown great performance in many tasks. Similar to associative memory systems, these networks define a dynamical system that converges to a set of target states. In this work, we show that generative diffusion models can be interpreted as energy-based models and that, when trained on discrete patterns, their energy function is (asymptotically) identical to that of modern Hopfield networks. This equivalence allows us to interpret the supervised training of diffusion models as a synaptic learning process that encodes the associative dynamics of a modern Hopfield network in the weight structure of a deep neural network. Leveraging this connection, we formulate a generalized framework for understanding the formation of long-term memory, where creative generation and memory recall can be seen as parts of a unified continuum.
Collapse
Affiliation(s)
- Luca Ambrogioni
- Donders Institute for Brain, Cognition, and Behaviour, Radboud University, 6525 XZ Nijmegen, The Netherlands
| |
Collapse
|
6
|
Millidge B, Tang M, Osanlouy M, Harper NS, Bogacz R. Predictive coding networks for temporal prediction. PLoS Comput Biol 2024; 20:e1011183. [PMID: 38557984 PMCID: PMC11008833 DOI: 10.1371/journal.pcbi.1011183] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2023] [Revised: 04/11/2024] [Accepted: 03/12/2024] [Indexed: 04/04/2024] Open
Abstract
One of the key problems the brain faces is inferring the state of the world from a sequence of dynamically changing stimuli, and it is not yet clear how the sensory system achieves this task. A well-established computational framework for describing perceptual processes in the brain is provided by the theory of predictive coding. Although the original proposals of predictive coding have discussed temporal prediction, later work developing this theory mostly focused on static stimuli, and key questions on neural implementation and computational properties of temporal predictive coding networks remain open. Here, we address these questions and present a formulation of the temporal predictive coding model that can be naturally implemented in recurrent networks, in which activity dynamics rely only on local inputs to the neurons, and learning only utilises local Hebbian plasticity. Additionally, we show that temporal predictive coding networks can approximate the performance of the Kalman filter in predicting behaviour of linear systems, and behave as a variant of a Kalman filter which does not track its own subjective posterior variance. Importantly, temporal predictive coding networks can achieve similar accuracy as the Kalman filter without performing complex mathematical operations, but just employing simple computations that can be implemented by biological networks. Moreover, when trained with natural dynamic inputs, we found that temporal predictive coding can produce Gabor-like, motion-sensitive receptive fields resembling those observed in real neurons in visual areas. In addition, we demonstrate how the model can be effectively generalized to nonlinear systems. Overall, models presented in this paper show how biologically plausible circuits can predict future stimuli and may guide research on understanding specific neural circuits in brain areas involved in temporal prediction.
Collapse
Affiliation(s)
- Beren Millidge
- MRC Brain Network Dynamics Unit, University of Oxford, Oxford, United Kingdom
| | - Mufeng Tang
- MRC Brain Network Dynamics Unit, University of Oxford, Oxford, United Kingdom
| | - Mahyar Osanlouy
- Auckland Bioengineering Institute, University of Auckland, Auckland, New Zealand
| | - Nicol S. Harper
- Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford, United Kingdom
| | - Rafal Bogacz
- MRC Brain Network Dynamics Unit, University of Oxford, Oxford, United Kingdom
| |
Collapse
|
7
|
Priorelli M, Pezzulo G, Stoianov IP. Deep kinematic inference affords efficient and scalable control of bodily movements. Proc Natl Acad Sci U S A 2023; 120:e2309058120. [PMID: 38085784 PMCID: PMC10743426 DOI: 10.1073/pnas.2309058120] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2023] [Accepted: 10/24/2023] [Indexed: 12/18/2023] Open
Abstract
Performing goal-directed movements requires mapping goals from extrinsic (workspace-relative) to intrinsic (body-relative) coordinates and then to motor signals. Mainstream approaches based on optimal control realize the mappings by minimizing cost functions, which is computationally demanding. Instead, active inference uses generative models to produce sensory predictions, which allows a cheaper inversion to the motor signals. However, devising generative models to control complex kinematic chains like the human body is challenging. We introduce an active inference architecture that affords a simple but effective mapping from extrinsic to intrinsic coordinates via inference and easily scales up to drive complex kinematic chains. Rich goals can be specified in both intrinsic and extrinsic coordinates using attractive or repulsive forces. The proposed model reproduces sophisticated bodily movements and paves the way for computationally efficient and biologically plausible control of actuated systems.
Collapse
Affiliation(s)
- Matteo Priorelli
- National Research Council, Institute of Cognitive Sciences and Technologies, Padova35137, Italy
| | - Giovanni Pezzulo
- National Research Council, Institute of Cognitive Sciences and Technologies, Rome00185, Italy
| | - Ivilin Peev Stoianov
- National Research Council, Institute of Cognitive Sciences and Technologies, Padova35137, Italy
| |
Collapse
|
8
|
Muzik O, Diwadkar VA. Depth and hierarchies in the predictive brain: From reaction to action. WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2023; 14:e1664. [PMID: 37518831 DOI: 10.1002/wcs.1664] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/05/2022] [Revised: 05/18/2023] [Accepted: 07/10/2023] [Indexed: 08/01/2023]
Abstract
The human brain is a prediction device, a view widely accepted in neuroscience. Prediction is a rational and efficient response that relies on the brain's ability to create and employ generative models to optimize actions over unpredictable time horizons. We argue that extant predictive frameworks while compelling, have not explicitly accounted for the following: (a) The brain's generative models must incorporate predictive depth (i.e., rely on degrees of abstraction to enable predictions over different time horizons); (b) The brain's implementation scheme to account for varying predictive depth relies on dynamic predictive hierarchies formed using the brain's functional networks. We show that these hierarchies incorporate the ascending processes (driven by reaction), and the descending processes (related to prediction), eventually driving action. Because they are dynamically formed, predictive hierarchies allow the brain to address predictive challenges in virtually any domain. By way of application, we explain how this framework can be applied to heretofore poorly understood processes of human behavioral thermoregulation. Although mammalian thermoregulation has been closely tied to deep brain structures engaged in autonomic control such as the hypothalamus, this narrow conception does not translate well to humans. In addition to profound differences in evolutionary history, the human brain is bestowed with substantially increased functional complexity (that itself emerged from evolutionary differences). We argue that behavioral thermoregulation in humans is possible because, (a) ascending signals shaped by homeostatic sub-networks, interject with (b) descending signals related to prediction (implemented in interoceptive and executive sub-networks) and action (implemented in executive sub-networks). These sub-networks cumulatively form a predictive hierarchy for human thermoregulation, potentiating a range of viable responses to known and unknown thermoregulatory challenges. We suggest that our proposed extensions to the predictive framework provide a set of generalizable principles that can further illuminate the many facets of the predictive brain. This article is categorized under: Neuroscience > Behavior Philosophy > Action Psychology > Prediction.
Collapse
Affiliation(s)
- Otto Muzik
- Department of Pediatrics, Wayne State University School of Medicine, Children's Hospital of Michigan, Michigan, USA
| | - Vaibhav A Diwadkar
- Department of Psychiatry and Behavioral Neurosciences, Wayne State University School of Medicine, Detroit, Michigan, USA
| |
Collapse
|
9
|
Tang M, Salvatori T, Millidge B, Song Y, Lukasiewicz T, Bogacz R. Recurrent predictive coding models for associative memory employing covariance learning. PLoS Comput Biol 2023; 19:e1010719. [PMID: 37058541 PMCID: PMC10132551 DOI: 10.1371/journal.pcbi.1010719] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2022] [Revised: 04/26/2023] [Accepted: 03/07/2023] [Indexed: 04/15/2023] Open
Abstract
The computational principles adopted by the hippocampus in associative memory (AM) tasks have been one of the most studied topics in computational and theoretical neuroscience. Recent theories suggested that AM and the predictive activities of the hippocampus could be described within a unitary account, and that predictive coding underlies the computations supporting AM in the hippocampus. Following this theory, a computational model based on classical hierarchical predictive networks was proposed and was shown to perform well in various AM tasks. However, this fully hierarchical model did not incorporate recurrent connections, an architectural component of the CA3 region of the hippocampus that is crucial for AM. This makes the structure of the model inconsistent with the known connectivity of CA3 and classical recurrent models such as Hopfield Networks, which learn the covariance of inputs through their recurrent connections to perform AM. Earlier PC models that learn the covariance information of inputs explicitly via recurrent connections seem to be a solution to these issues. Here, we show that although these models can perform AM, they do it in an implausible and numerically unstable way. Instead, we propose alternatives to these earlier covariance-learning predictive coding networks, which learn the covariance information implicitly and plausibly, and can use dendritic structures to encode prediction errors. We show analytically that our proposed models are perfectly equivalent to the earlier predictive coding model learning covariance explicitly, and encounter no numerical issues when performing AM tasks in practice. We further show that our models can be combined with hierarchical predictive coding networks to model the hippocampo-neocortical interactions. Our models provide a biologically plausible approach to modelling the hippocampal network, pointing to a potential computational mechanism during hippocampal memory formation and recall, which employs both predictive coding and covariance learning based on the recurrent network structure of the hippocampus.
Collapse
Affiliation(s)
- Mufeng Tang
- MRC Brain Network Dynamics Unit, University of Oxford, Oxford, United Kingdom
| | - Tommaso Salvatori
- Department of Computer Science, University of Oxford, Oxford, United Kingdom
| | - Beren Millidge
- MRC Brain Network Dynamics Unit, University of Oxford, Oxford, United Kingdom
| | - Yuhang Song
- MRC Brain Network Dynamics Unit, University of Oxford, Oxford, United Kingdom
- Department of Computer Science, University of Oxford, Oxford, United Kingdom
| | - Thomas Lukasiewicz
- Department of Computer Science, University of Oxford, Oxford, United Kingdom
- Institute of Logic and Computation, TU Wien, Vienna, Austria
| | - Rafal Bogacz
- MRC Brain Network Dynamics Unit, University of Oxford, Oxford, United Kingdom
| |
Collapse
|
10
|
Salvatori T, Song Y, Hong Y, Sha L, Frieder S, Xu Z, Bogacz R, Lukasiewicz T. Associative Memories via Predictive Coding. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 2021; 34:3874-3886. [PMID: 35664437 PMCID: PMC7612799] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Associative memories in the brain receive and store patterns of activity registered by the sensory neurons, and are able to retrieve them when necessary. Due to their importance in human intelligence, computational models of associative memories have been developed for several decades now. In this paper, we present a novel neural model for realizing associative memories, which is based on a hierarchical generative network that receives external stimuli via sensory neurons. It is trained using predictive coding, an error-based learning algorithm inspired by information processing in the cortex. To test the model's capabilities, we perform multiple retrieval experiments from both corrupted and incomplete data points. In an extensive comparison, we show that this new model outperforms in retrieval accuracy and robustness popular associative memory models, such as autoencoders trained via backpropagation, and modern Hopfield networks. In particular, in completing partial data points, our model achieves remarkable results on natural image datasets, such as ImageNet, with a surprisingly high accuracy, even when only a tiny fraction of pixels of the original images is presented. Our model provides a plausible framework to study learning and retrieval of memories in the brain, as it closely mimics the behavior of the hippocampus as a memory index and generative model.
Collapse
Affiliation(s)
| | - Yuhang Song
- Department of Computer Science, University of Oxford, UK
- MRC Brain Network Dynamics Unit, University of Oxford, UK
| | - Yujian Hong
- Department of Computer Science, University of Oxford, UK
| | - Lei Sha
- Department of Computer Science, University of Oxford, UK
| | - Simon Frieder
- Department of Computer Science, University of Oxford, UK
| | - Zhenghua Xu
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, Hebei University of Technology, Tianjin, China
| | - Rafal Bogacz
- MRC Brain Network Dynamics Unit, University of Oxford, UK
| | | |
Collapse
|