1
|
Dehaene S, Sablé-Meyer M, Ciccione L. Origins of numbers: a shared language-of-thought for arithmetic and geometry? Trends Cogn Sci 2025:S1364-6613(25)00059-2. [PMID: 40234140 DOI: 10.1016/j.tics.2025.03.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2024] [Revised: 02/07/2025] [Accepted: 03/06/2025] [Indexed: 04/17/2025]
Abstract
Concepts of exact number are often thought to originate from counting and the successor function, or from a refinement of the approximate number system (ANS). We argue here for a third origin: a shared language-of-thought (LoT) for geometry and arithmetic that involves primitives of repetition, concatenation, and recursive embedding. Applied to sets, those primitives engender concepts of exact integers through recursive applications of additions and multiplications. Links between geometry and arithmetic also explain the emergence of higher-level notions (squares, primes, etc.). Under our hypothesis, understanding a number means having one or several mental expressions for it, and their minimal description length (MDL) determines how easily they can be mentally manipulated. Several historical, developmental, linguistic, and brain imaging phenomena provide preliminary support for our proposal.
Collapse
Affiliation(s)
- Stanislas Dehaene
- Cognitive Neuroimaging Unit, Commissariat à l'Energie Atomique (CEA), Institut National de la Santé et de la Recherche Médicale (INSERM), NeuroSpin Center, Université Paris-Saclay, 91191 Gif-sur-Yvette, France; Collège de France, Université Paris-Sciences-Lettres (PSL), 11 Place Marcelin Berthelot, 75005 Paris, France
| | - Mathias Sablé-Meyer
- Cognitive Neuroimaging Unit, Commissariat à l'Energie Atomique (CEA), Institut National de la Santé et de la Recherche Médicale (INSERM), NeuroSpin Center, Université Paris-Saclay, 91191 Gif-sur-Yvette, France; Collège de France, Université Paris-Sciences-Lettres (PSL), 11 Place Marcelin Berthelot, 75005 Paris, France
| | - Lorenzo Ciccione
- Cognitive Neuroimaging Unit, Commissariat à l'Energie Atomique (CEA), Institut National de la Santé et de la Recherche Médicale (INSERM), NeuroSpin Center, Université Paris-Saclay, 91191 Gif-sur-Yvette, France; Collège de France, Université Paris-Sciences-Lettres (PSL), 11 Place Marcelin Berthelot, 75005 Paris, France.
| |
Collapse
|
2
|
Schmidbauer P, Hahn M, Nieder A. Crows recognize geometric regularity. SCIENCE ADVANCES 2025; 11:eadt3718. [PMID: 40215319 PMCID: PMC11988402 DOI: 10.1126/sciadv.adt3718] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/24/2024] [Accepted: 03/07/2025] [Indexed: 04/14/2025]
Abstract
The perception of geometric regularity in shapes, a form of elementary Euclidean geometry, is a fundamental mathematical intuition in humans. We demonstrate this geometric understanding in an animal, the carrion crow. Crows were trained to detect a visually distinct intruder shape among six concurrent arbitrary shapes. The crows were able to immediately apply this intruder concept to quadrilaterals, identifying the one that exhibited differing geometric properties compared to the others in the set. The crows exhibited a geometric regularity effect, showing better performance with shapes featuring right angles, parallel lines, or symmetry over more irregular shapes. This performance advantage did not require learning. Our findings suggest that geometric intuitions are not specific to humans but are deeply rooted in biological evolution.
Collapse
Affiliation(s)
- Philipp Schmidbauer
- Animal Physiology Unit, Institute of Neurobiology, University of Tübingen, 72076 Tübingen, Germany
| | - Madita Hahn
- Animal Physiology Unit, Institute of Neurobiology, University of Tübingen, 72076 Tübingen, Germany
| | - Andreas Nieder
- Animal Physiology Unit, Institute of Neurobiology, University of Tübingen, 72076 Tübingen, Germany
| |
Collapse
|
3
|
Chen W, Ye S, Ding X, Shen M, Gao Z. Selectively maintaining an object's feature in visual working memory: A comparison between highly discriminable and fine-grained features. Mem Cognit 2025; 53:853-868. [PMID: 39048836 DOI: 10.3758/s13421-024-01612-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/03/2024] [Indexed: 07/27/2024]
Abstract
Selectively maintaining information is an essential function of visual working memory (VWM). Recent VWM studies have mainly focused on selective maintenance of objects, leaving the mechanisms of selectively maintaining an object's feature in VWM unknown. Based on the interactive model of perception and VWM, we hypothesized that there are distinct selective maintenance mechanisms for objects containing fine-grained features versus objects containing highly discriminable features. To test this hypothesis, we first required participants to memorize a dual-feature object (colored simple shapes vs. colored polygons), and informed them about the target feature via a retro-cue. Then a visual search task was added to examine the fate of the irrelevant feature. The selective maintenance of an object's feature predicted that the irrelevant feature should be removed from the active state of VWM and should not capture attention when presented as a distractor in the visual search task. We found that irrelevant simple shapes impaired performance in the visual search task (Experiment 1). However, irrelevant polygons did not affect visual search performance (Experiment 2), and this could not be explained by decay of polygons (Experiment 3) or by polygons not capturing attention (Experiment 4). These findings suggest that VWM adopts dissociable mechanisms to selectively maintain an object's feature, depending on the feature's perceptual characteristics.
Collapse
Affiliation(s)
- Wei Chen
- Department of Psychology, Sun Yat-sen University, Guangzhou, China
| | - Shujuan Ye
- Department of Psychology, Sun Yat-sen University, Guangzhou, China
| | - Xiaowei Ding
- Department of Psychology, Sun Yat-sen University, Guangzhou, China.
| | - Mowei Shen
- Department of Psychology and Behavioral Sciences, Zhejiang University, Zhejiang, China.
| | - Zaifeng Gao
- Department of Psychology and Behavioral Sciences, Zhejiang University, Zhejiang, China.
| |
Collapse
|
4
|
Yousif SR, Goldstein LB, Brannon EM. Children's Understanding of Topological Relations. Open Mind (Camb) 2025; 9:401-417. [PMID: 40177301 PMCID: PMC11964115 DOI: 10.1162/opmi_a_00194] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2024] [Accepted: 02/02/2025] [Indexed: 04/05/2025] Open
Abstract
A core aim of developmental cognitive science is to uncover the basic building blocks of human thought. For instance, work revealing that even young children, adults without formal education, and distant animal species are sensitive to basic Euclidean properties indicates that humans may be endowed with some primitive understanding of Euclidean geometry. But what about other forms of geometry? Here, we explore children's sensitivity to topological spatial forms. We show that children, like adults, spontaneously distinguish and match items in accordance with their topological relations. As well, we show that children's judgments about object similarity are remarkably consistent with adults', indicating stability in object concepts throughout the lifespan. Finally, we compare children's sensitivity to various topological forms with their sensitivity to geometric properties like curvature, perpendicularity, and symmetry, and find that while there is some variability in performance across all the features tested, overall performance for geometric vs. topological is comparable. Collectively, these findings suggest that even young children have an intuitive understanding of topological relations and suggest that topological relations may be among the building blocks of human visuospatial representation.
Collapse
Affiliation(s)
- Sami R. Yousif
- Department of Psychology & Neuroscience, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
| | - Lily B. Goldstein
- Department of Psychology, University of Pennsylvania, Philadelphia, PA, USA
| | | |
Collapse
|
5
|
Yousif SR, Brannon EM. Perceiving Topological Relations. Psychol Sci 2025; 36:71-86. [PMID: 39965204 DOI: 10.1177/09567976241309615] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/20/2025] Open
Abstract
There are many ways to describe and represent the visuospatial world. A space can be described by its euclidean properties-the size of objects, the angles of boundaries, the distances between them. A space can also be described in nonspatial terms: One could explain the layout of a city by the order of its streets. Somewhere in between, topological representations-such as those commonly depicted in public-transit maps-capture coarse relational structure without precise euclidean detail, offering a relatively efficient, low-dimensional way of capturing spatial content. Here, we ask whether human adults quickly and automatically perceive such relations. In six experiments, we show that differences in simple topological features influence a range of visual tasks from object matching to number estimation to visual search. We discuss the possibility that topological relations are a kind of visual primitive that supports visuospatial representation.
Collapse
Affiliation(s)
- Sami R Yousif
- Department of Psychology and Neuroscience, University of North Carolina, Chapel Hill
| | | |
Collapse
|
6
|
Yang Q, Zhu Z, Si R, Li Y, Zhang J, Yang T. A language model of problem solving in humans and macaque monkeys. Curr Biol 2025; 35:11-20.e10. [PMID: 39631400 DOI: 10.1016/j.cub.2024.10.074] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Revised: 09/30/2024] [Accepted: 10/29/2024] [Indexed: 12/07/2024]
Abstract
Human intelligence is characterized by the remarkable ability to solve complex problems by planning a sequence of actions that takes us from an initial state to a desired goal state. Quantifying and comparing problem-solving capabilities across species and finding their evolutionary roots are critical for understanding how the brain carries out this intricate process. We introduce the Language of Problem Solving (LoPS) model as a novel quantitative framework that investigates the structure of problem-solving behavior through a language model. We applied the model to an adapted classic Pac-Man game as a cross-species behavioral paradigm to test both humans and macaque monkeys. The LoPS model extracted the latent structure, or grammar, embedded in the agents' gameplay, revealing the non-Markovian temporal dependency structure of their problem-solving behavior and the hierarchical structures of problem solving in both species. The complexity of LoPS grammar correlated with individuals' game performance and reflected the difference in problem-solving capacity between humans and monkeys. Both species evolved their LoPS grammars during learning, progressing from simpler to more complex ones, suggesting that the structure of problem solving is not fixed but evolves to support more sophisticated and efficient problem solving. Our study provides insights into how humans and monkeys break down problem solving into compositional units and navigate complex tasks, deepening our understanding of human intelligence and its evolution and establishing a foundation for future investigations of the neural mechanisms of problem solving.
Collapse
Affiliation(s)
- Qianli Yang
- Institute of Neuroscience, Key Laboratory of Brain Cognition and Brain-inspired Intelligence Technology, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai 200031, China.
| | - Zhihua Zhu
- Institute of Neuroscience, Key Laboratory of Brain Cognition and Brain-inspired Intelligence Technology, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai 200031, China
| | - Ruoguang Si
- Institute of Neuroscience, Key Laboratory of Brain Cognition and Brain-inspired Intelligence Technology, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai 200031, China; Cardiff University Brain Research Imaging Centre, School of Psychology, Cardiff University, Maindy Road, Cardiff CF24 4HQ, UK
| | - Yunwei Li
- Institute of Neuroscience, Key Laboratory of Brain Cognition and Brain-inspired Intelligence Technology, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai 200031, China; University of Chinese Academy of Sciences, Beijing 101408, China
| | - Jiaxiang Zhang
- School of Mathematics and Computer Science, Swansea University, Swansea SA1 8DD, UK
| | - Tianming Yang
- Institute of Neuroscience, Key Laboratory of Brain Cognition and Brain-inspired Intelligence Technology, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai 200031, China.
| |
Collapse
|
7
|
Debray S, Dehaene S. Mapping and modeling the semantic space of math concepts. Cognition 2025; 254:105971. [PMID: 39369595 DOI: 10.1016/j.cognition.2024.105971] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2024] [Revised: 09/26/2024] [Accepted: 09/27/2024] [Indexed: 10/08/2024]
Abstract
Mathematics is an underexplored domain of human cognition. While many studies have focused on subsets of math concepts such as numbers, fractions, or geometric shapes, few have ventured beyond these elementary domains. Here, we attempted to map out the full space of math concepts and to answer two specific questions: can distributed semantic models, such a GloVe, provide a satisfactory fit to human semantic judgements in mathematics? And how does this fit vary with education? We first analyzed all of the French and English Wikipedia pages with math contents, and used a semi-automatic procedure to extract the 1000 most frequent math terms in both languages. In a second step, we collected extensive behavioral judgements of familiarity and semantic similarity between them. About half of the variance in human similarity judgements was explained by vector embeddings that attempt to capture latent semantic structures based on cooccurence statistics. Participants' self-reported level of education modulated familiarity and similarity, allowing us to create a partial hierarchy among high-level math concepts. Our results converge onto the proposal of a map of math space, organized as a database of math terms with information about their frequency, familiarity, grade of acquisition, and entanglement with other concepts.
Collapse
Affiliation(s)
- Samuel Debray
- Cognitive Neuroimaging Unit, Institut National de la Santé et de la Recherche Médicale, Commissariat à l'Energie Atomique et aux énergies alternatives, Centre National de la Recherche Scientifique, Université Paris-Saclay, NeuroSpin center, Gif-sur-Yvette, France.
| | - Stanislas Dehaene
- Cognitive Neuroimaging Unit, Institut National de la Santé et de la Recherche Médicale, Commissariat à l'Energie Atomique et aux énergies alternatives, Centre National de la Recherche Scientifique, Université Paris-Saclay, NeuroSpin center, Gif-sur-Yvette, France; Collège de France, Université Paris Sciences & Lettres, Paris, France.
| |
Collapse
|
8
|
Butz MV, Mittenbühler M, Schwöbel S, Achimova A, Gumbsch C, Otte S, Kiebel S. Contextualizing predictive minds. Neurosci Biobehav Rev 2025; 168:105948. [PMID: 39580009 DOI: 10.1016/j.neubiorev.2024.105948] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2023] [Revised: 09/13/2024] [Accepted: 11/16/2024] [Indexed: 11/25/2024]
Abstract
The structure of human memory seems to be optimized for efficient prediction, planning, and behavior. We propose that these capacities rely on a tripartite structure of memory that includes concepts, events, and contexts-three layers that constitute the mental world model. We suggest that the mechanism that critically increases adaptivity and flexibility is the tendency to contextualize. This tendency promotes local, context-encoding abstractions, which focus event- and concept-based planning and inference processes on the task and situation at hand. As a result, cognitive contextualization offers a solution to the frame problem-the need to select relevant features of the environment from the rich stream of sensorimotor signals. We draw evidence for our proposal from developmental psychology and neuroscience. Adopting a computational stance, we present evidence from cognitive modeling research which suggests that context sensitivity is a feature that is critical for maximizing the efficiency of cognitive processes. Finally, we turn to recent deep-learning architectures which independently demonstrate how context-sensitive memory can emerge in a self-organized learning system constrained by cognitively-inspired inductive biases.
Collapse
Affiliation(s)
- Martin V Butz
- Cognitive Modeling, Faculty of Science, University of Tübingen, Sand 14, Tübingen 72076, Germany.
| | - Maximilian Mittenbühler
- Cognitive Modeling, Faculty of Science, University of Tübingen, Sand 14, Tübingen 72076, Germany
| | - Sarah Schwöbel
- Cognitive Computational Neuroscience, Faculty of Psychology, TU Dresden, School of Science, Dresden 01062, Germany
| | - Asya Achimova
- Cognitive Modeling, Faculty of Science, University of Tübingen, Sand 14, Tübingen 72076, Germany
| | - Christian Gumbsch
- Cognitive Modeling, Faculty of Science, University of Tübingen, Sand 14, Tübingen 72076, Germany; Chair of Cognitive and Clinical Neuroscience, Faculty of Psychology, TU Dresden, Dresden 01069, Germany
| | - Sebastian Otte
- Cognitive Modeling, Faculty of Science, University of Tübingen, Sand 14, Tübingen 72076, Germany; Adaptive AI Lab, Institute of Robotics and Cognitive Systems, University of Lübeck, Ratzeburger Allee 160, Lübeck 23562, Germany
| | - Stefan Kiebel
- Cognitive Computational Neuroscience, Faculty of Psychology, TU Dresden, School of Science, Dresden 01062, Germany
| |
Collapse
|
9
|
Rule JS, Piantadosi ST, Cropper A, Ellis K, Nye M, Tenenbaum JB. Symbolic metaprogram search improves learning efficiency and explains rule learning in humans. Nat Commun 2024; 15:6847. [PMID: 39127796 DOI: 10.1038/s41467-024-50966-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2024] [Accepted: 07/23/2024] [Indexed: 08/12/2024] Open
Abstract
Throughout their lives, humans seem to learn a variety of rules for things like applying category labels, following procedures, and explaining causal relationships. These rules are often algorithmically rich but are nonetheless acquired with minimal data and computation. Symbolic models based on program learning successfully explain rule-learning in many domains, but performance degrades quickly as program complexity increases. It remains unclear how to scale symbolic rule-learning methods to model human performance in challenging domains. Here we show that symbolic search over the space of metaprograms-programs that revise programs-dramatically improves learning efficiency. On a behavioral benchmark of 100 algorithmically rich rules, this approach fits human learning more accurately than alternative models while also using orders of magnitude less search. The computation required to match median human performance is consistent with conservative estimates of human thinking time. Our results suggest that metaprogram-like representations may help human learners to efficiently acquire rules.
Collapse
Affiliation(s)
- Joshua S Rule
- Psychology, University of California, Berkeley, Berkeley, CA, 94704, USA.
| | | | | | - Kevin Ellis
- Computer Science, Cornell University, Ithaca, NY, 14850, USA
| | - Maxwell Nye
- Adept AI Labs, San Francisco, CA, 94110, USA
| | - Joshua B Tenenbaum
- Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA, 02139, USA
| |
Collapse
|
10
|
Hodgson D. Commentary on " 'Snakes and ladders' in paleoanthropology: From cognitive surprise to skillfulness a million years ago". Phys Life Rev 2024; 49:134-135. [PMID: 38718470 DOI: 10.1016/j.plrev.2024.04.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2024] [Accepted: 04/21/2024] [Indexed: 05/25/2024]
|
11
|
Sun Z, Han S, Firestone C. Caricaturing Shapes in Visual Memory. Psychol Sci 2024; 35:722-735. [PMID: 38648201 DOI: 10.1177/09567976231225091] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/25/2024] Open
Abstract
When representing high-level stimuli, such as faces and animals, we tend to emphasize salient features-such as a face's prominent cheekbones or a bird's pointed beak. Such mental caricaturing leaves traces in memory, which exaggerates these distinctive qualities. How broadly does this phenomenon extend? Here, in six experiments (N = 700 adults), we explored how memory automatically caricatures basic units of visual processing-simple geometric shapes-even without task-related demands to do so. Participants saw a novel shape and then immediately adjusted a copy of that shape to match what they had seen. Surprisingly, participants reconstructed shapes in exaggerated form, amplifying curvature, enlarging salient parts, and so on. Follow-up experiments generalized this bias to new parameters, ruled out strategic responding, and amplified the effects in serial transmission. Thus, even the most basic stimuli we encounter are remembered as caricatures of themselves.
Collapse
Affiliation(s)
- Zekun Sun
- Department of Psychological and Brain Sciences, Johns Hopkins University
| | - Subin Han
- Department of Psychological and Brain Sciences, Johns Hopkins University
| | - Chaz Firestone
- Department of Psychological and Brain Sciences, Johns Hopkins University
| |
Collapse
|
12
|
Dillon MR. Divisive language. Behav Brain Sci 2024; 47:e124. [PMID: 38934439 DOI: 10.1017/s0140525x23003047] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/28/2024]
Abstract
What language devises, it might divide. By exploring the relations among the core geometries of the physical world, the abstract geometry of Euclid, and language, I give new insight into both the persistence of core knowledge into adulthood and our access to it through language. My extension of Spelke's language argument has implications for pedagogy, philosophy, and artificial intelligence.
Collapse
Affiliation(s)
- Moira R Dillon
- Department of Psychology, New York University, New York, NY, USA
| |
Collapse
|
13
|
Al Roumi F, Planton S, Wang L, Dehaene S. Brain-imaging evidence for compression of binary sound sequences in human memory. eLife 2023; 12:e84376. [PMID: 37910588 PMCID: PMC10619979 DOI: 10.7554/elife.84376] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2022] [Accepted: 10/14/2023] [Indexed: 11/03/2023] Open
Abstract
According to the language-of-thought hypothesis, regular sequences are compressed in human memory using recursive loops akin to a mental program that predicts future items. We tested this theory by probing memory for 16-item sequences made of two sounds. We recorded brain activity with functional MRI and magneto-encephalography (MEG) while participants listened to a hierarchy of sequences of variable complexity, whose minimal description required transition probabilities, chunking, or nested structures. Occasional deviant sounds probed the participants' knowledge of the sequence. We predicted that task difficulty and brain activity would be proportional to the complexity derived from the minimal description length in our formal language. Furthermore, activity should increase with complexity for learned sequences, and decrease with complexity for deviants. These predictions were upheld in both fMRI and MEG, indicating that sequence predictions are highly dependent on sequence structure and become weaker and delayed as complexity increases. The proposed language recruited bilateral superior temporal, precentral, anterior intraparietal, and cerebellar cortices. These regions overlapped extensively with a localizer for mathematical calculation, and much less with spoken or written language processing. We propose that these areas collectively encode regular sequences as repetitions with variations and their recursive composition into nested structures.
Collapse
Affiliation(s)
- Fosca Al Roumi
- Cognitive Neuroimaging Unit, Université Paris-Saclay, INSERM, CEA, CNRS, NeuroSpin centerGif/YvetteFrance
| | - Samuel Planton
- Cognitive Neuroimaging Unit, Université Paris-Saclay, INSERM, CEA, CNRS, NeuroSpin centerGif/YvetteFrance
| | - Liping Wang
- Institute of Neuroscience, Key Laboratory of Primate Neurobiology, CAS Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of SciencesShanghaiChina
| | - Stanislas Dehaene
- Cognitive Neuroimaging Unit, Université Paris-Saclay, INSERM, CEA, CNRS, NeuroSpin centerGif/YvetteFrance
- Collège de France, Université Paris Sciences Lettres (PSL)ParisFrance
| |
Collapse
|
14
|
Canudas-Grabolosa I, Martín-Salguero A, Bonatti LL. Natural logic and baby LoTH. Behav Brain Sci 2023; 46:e266. [PMID: 37766633 DOI: 10.1017/s0140525x23001942] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/29/2023]
Abstract
Language-of-thought hypothesis (LoTH) is having a profound impact on cognition studies. However, much remains unknown about its basic primitives and generative operations. Infant studies are fundamental, but methodologically very challenging. By distilling potential primitives from work in natural-language semantics, an approach beyond the corset of standard formal logic may be undertaken. Still, the road ahead is challenging and long.
Collapse
Affiliation(s)
| | - Ana Martín-Salguero
- Center for Brain and Cognition, Universitat Pompeu Fabra, Barcelona, Spain
- Cognitive Neuroimaging Unit, CEA, INSERM, Université Paris-Saclay, NeuroSpin Center, Gif/Yvette, France
| | - Luca L Bonatti
- Center for Brain and Cognition, Universitat Pompeu Fabra, Barcelona, Spain
- ICREA, Barcelona, Spain
| |
Collapse
|
15
|
Kumar S, Dasgupta I, Daw ND, Cohen JD, Griffiths TL. Disentangling Abstraction from Statistical Pattern Matching in Human and Machine Learning. PLoS Comput Biol 2023; 19:e1011316. [PMID: 37624841 PMCID: PMC10497163 DOI: 10.1371/journal.pcbi.1011316] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2023] [Revised: 09/12/2023] [Accepted: 06/29/2023] [Indexed: 08/27/2023] Open
Abstract
The ability to acquire abstract knowledge is a hallmark of human intelligence and is believed by many to be one of the core differences between humans and neural network models. Agents can be endowed with an inductive bias towards abstraction through meta-learning, where they are trained on a distribution of tasks that share some abstract structure that can be learned and applied. However, because neural networks are hard to interpret, it can be difficult to tell whether agents have learned the underlying abstraction, or alternatively statistical patterns that are characteristic of that abstraction. In this work, we compare the performance of humans and agents in a meta-reinforcement learning paradigm in which tasks are generated from abstract rules. We define a novel methodology for building "task metamers" that closely match the statistics of the abstract tasks but use a different underlying generative process, and evaluate performance on both abstract and metamer tasks. We find that humans perform better at abstract tasks than metamer tasks whereas common neural network architectures typically perform worse on the abstract tasks than the matched metamers. This work provides a foundation for characterizing differences between humans and machine learning that can be used in future work towards developing machines with more human-like behavior.
Collapse
Affiliation(s)
- Sreejan Kumar
- Neuroscience Institute, Princeton University, Princeton, New Jersey, United States of America
| | | | - Nathaniel D. Daw
- Neuroscience Institute, Princeton University, Princeton, New Jersey, United States of America
- Department of Psychology, Princeton University, Princeton, New Jersey, United States of America
| | - Jonathan. D. Cohen
- Neuroscience Institute, Princeton University, Princeton, New Jersey, United States of America
- Department of Psychology, Princeton University, Princeton, New Jersey, United States of America
| | - Thomas L. Griffiths
- Department of Psychology, Princeton University, Princeton, New Jersey, United States of America
- Department of Computer Science, Princeton University, Princeton, New Jersey, United States of America
| |
Collapse
|
16
|
Ellis K, Wong L, Nye M, Sablé-Meyer M, Cary L, Anaya Pozo L, Hewitt L, Solar-Lezama A, Tenenbaum JB. DreamCoder: growing generalizable, interpretable knowledge with wake-sleep Bayesian program learning. PHILOSOPHICAL TRANSACTIONS. SERIES A, MATHEMATICAL, PHYSICAL, AND ENGINEERING SCIENCES 2023; 381:20220050. [PMID: 37271169 DOI: 10.1098/rsta.2022.0050] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/31/2023] [Accepted: 03/23/2023] [Indexed: 06/06/2023]
Abstract
Expert problem-solving is driven by powerful languages for thinking about problems and their solutions. Acquiring expertise means learning these languages-systems of concepts, alongside the skills to use them. We present DreamCoder, a system that learns to solve problems by writing programs. It builds expertise by creating domain-specific programming languages for expressing domain concepts, together with neural networks to guide the search for programs within these languages. A 'wake-sleep' learning algorithm alternately extends the language with new symbolic abstractions and trains the neural network on imagined and replayed problems. DreamCoder solves both classic inductive programming tasks and creative tasks such as drawing pictures and building scenes. It rediscovers the basics of modern functional programming, vector algebra and classical physics, including Newton's and Coulomb's laws. Concepts are built compositionally from those learned earlier, yielding multilayered symbolic representations that are interpretable and transferrable to new tasks, while still growing scalably and flexibly with experience. This article is part of a discussion meeting issue 'Cognitive artificial intelligence'.
Collapse
|