1
|
Watson EAG, Ewing L, Malcolm GL. When children get the gist: The development of rapid scene categorisation. Vision Res 2025; 233:108620. [PMID: 40382938 DOI: 10.1016/j.visres.2025.108620] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2024] [Revised: 04/15/2025] [Accepted: 05/07/2025] [Indexed: 05/20/2025]
Abstract
Research surrounding adult recognition of scene gist is extensive; however, very little is known of its development. Behavioural studies of scene processing tend to broadly support a protracted developmental trajectory, with a quantitative and perhaps also qualitative shift towards more adultlike processing across middle childhood. Here we sought to better understand the very early stages of children's scene processing by targeting gist perception. Children aged 5-10 years categorised backwards-masked scenes presented at very brief durations. We drew inferences about the processing speed with which each age group extracted category-diagnostic information by varying presentation durations, and the quality of information extracted by varying the level they were prompted to make their judgments (superordinate-level indicative of coarse global information, basic-level indicative of more detailed information). Children across all ages demonstrated a remarkably sophisticated ability to extract scene gist, with 5-6-year-old children performing above chance for scenes presented for as little as 32 ms for both superordinate and basic-level judgements. Categorisation performance also became more efficient with age. Overall, our novel findings indicate that young children possess an impressive ability to process a scene's gist, which is followed by a protracted development towards expertise across middle childhood.
Collapse
Affiliation(s)
| | - Louise Ewing
- School of Psychology, University of East Anglia, Norwich, UK
| | | |
Collapse
|
2
|
Jung Y, Dilks DD. Early development of navigationally relevant location information in the retrosplenial complex. Proc Natl Acad Sci U S A 2025; 122:e2503569122. [PMID: 40324094 DOI: 10.1073/pnas.2503569122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2025] [Accepted: 04/05/2025] [Indexed: 05/07/2025] Open
Abstract
Representing the locations of places so that we can use them as landmarks is critical to our ability to navigate through large-scale spaces-a process referred to as "map-based navigation." While many neuroimaging studies in adults have revealed that this ability involves the retrosplenial complex (RSC)-a scene-selective region in the medial parietal cortex-nothing is known about how this cortical system develops. So, does it develop only late in childhood, as generally assumed from some behavioral studies? Or is it, perhaps counterintuitively, present in the first few years of life? To test this question, using functional magnetic resonance imaging (fMRI) multivoxel pattern analysis and a virtual town paradigm, we investigated the representation of location information in the RSC of 5-y-olds. We found that i) the RSC in 5-y-olds already represents the locations of particular buildings in the town (e.g., the ice cream store by the mountain versus by the lake), but not their category membership (e.g., ice cream store, regardless of location), and ii) this neural representation is correlated with their performance on a location task. Using multidimensional scaling, we also found that the neural representation of the buildings in RSC reflects the actual layout of the virtual town. Finally, the parahippocampal place area-a scene-selective region implicated in scene categorization, not map-based navigation-did not represent location information, but instead category information, the exact opposite of RSC. Taken together, these findings reveal the early development of navigationally relevant location information in RSC and thus the early origins of map-based navigation.
Collapse
Affiliation(s)
- Yaelan Jung
- Department of Psychology, Emory University, Atlanta, GA 30322
| | - Daniel D Dilks
- Department of Psychology, Emory University, Atlanta, GA 30322
| |
Collapse
|
3
|
Hauptman M, Elli G, Pant R, Bedny M. Neural specialization for 'visual' concepts emerges in the absence of vision. Cognition 2025; 257:106058. [PMID: 39827755 DOI: 10.1016/j.cognition.2024.106058] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2024] [Revised: 12/20/2024] [Accepted: 12/28/2024] [Indexed: 01/22/2025]
Abstract
The 'different-body/different-concepts hypothesis' central to some embodiment theories proposes that the sensory capacities of our bodies shape the cognitive and neural basis of our concepts. We tested this hypothesis by comparing behavioral semantic similarity judgments and neural signatures (fMRI) of 'visual' categories ('living things,' or animals, e.g., tiger, and light events, e.g., sparkle) across congenitally blind (n = 21) and sighted (n = 22) adults. Words referring to 'visual' entities/nouns and events/verbs (animals and light events) were compared to less vision-dependent categories from the same grammatical class (animal vs. place nouns, light vs. sound, mouth, and hand verbs). Within-category semantic similarity judgments about animals (e.g., sparrow vs. finch) were partially different across groups, consistent with the idea that sighted people rely on visually learned information to make such judgments about animals. However, robust neural specialization for living things in temporoparietal semantic networks, including in the precuneus, was observed in blind and sighted people alike. For light events, which are directly accessible only through vision, behavioral judgments were indistinguishable across groups. Neural responses to light events were also similar across groups: in both blind and sighted people, the left middle temporal gyrus (LMTG+) responded more to event concepts, including light events, compared to entity concepts. Multivariate patterns of neural activity in LMTG+ distinguished among different event types, including light events vs. other event types. In sum, we find that neural signatures of concepts previously attributed to visual experience do not require vision. Across a wide range of semantic types, conceptual representations develop independent of sensory experience.
Collapse
Affiliation(s)
- Miriam Hauptman
- Department of Psychological & Brain Sciences, Johns Hopkins University, Baltimore, MD, USA.
| | - Giulia Elli
- Department of Psychological & Brain Sciences, Johns Hopkins University, Baltimore, MD, USA
| | - Rashi Pant
- Department of Psychological & Brain Sciences, Johns Hopkins University, Baltimore, MD, USA; Department of Biological Psychology & Neuropsychology, Universität Hamburg, Germany.
| | - Marina Bedny
- Department of Psychological & Brain Sciences, Johns Hopkins University, Baltimore, MD, USA.
| |
Collapse
|
4
|
Persichetti AS, Shao J, Gotts SJ, Martin A. A functional parcellation of the whole brain in high-functioning individuals with autism spectrum disorder reveals atypical patterns of network organization. Mol Psychiatry 2025; 30:1518-1528. [PMID: 39349967 PMCID: PMC11919759 DOI: 10.1038/s41380-024-02764-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Revised: 09/19/2024] [Accepted: 09/23/2024] [Indexed: 10/09/2024]
Abstract
Researchers studying autism spectrum disorder (ASD) lack a comprehensive map of the functional network topography in the ASD brain. We used high-quality resting state functional MRI (rs-fMRI) connectivity data and a robust parcellation routine to provide a whole-brain map of functional networks in a group of seventy high-functioning individuals with ASD and a group of seventy typically developing (TD) individuals. The rs-fMRI data were collected using an imaging sequence optimized to achieve high temporal signal-to-noise ratio (tSNR) across the whole-brain. We identified functional networks using a parcellation routine that intrinsically incorporates internal consistency and repeatability of the networks by keeping only network distinctions that agree across halves of the data over multiple random iterations in each group. The groups were tightly matched on tSNR, in-scanner motion, age, and IQ. We compared the maps from each group and found that functional networks in the ASD group are atypical in three seemingly related ways: (1) whole-brain connectivity patterns are less stable across voxels within multiple functional networks, (2) the cerebellum, subcortex, and hippocampus show weaker differentiation of functional subnetworks, and (3) subcortical structures and the hippocampus are atypically integrated with the neocortex. These results were statistically robust and suggest that patterns of network connectivity between the neocortex and the cerebellum, subcortical structures, and hippocampus are atypical in ASD individuals.
Collapse
Affiliation(s)
- Andrew S Persichetti
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, National Institute of Mental Health, National Institutes of Health, Bethesda, MD, USA.
| | - Jiayu Shao
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, National Institute of Mental Health, National Institutes of Health, Bethesda, MD, USA
| | - Stephen J Gotts
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, National Institute of Mental Health, National Institutes of Health, Bethesda, MD, USA
| | - Alex Martin
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, National Institute of Mental Health, National Institutes of Health, Bethesda, MD, USA
| |
Collapse
|
5
|
Yoon HK, Jung Y, Persichetti AS, Dilks DD. A scene-selective region in the superior parietal lobule for visually guided navigation. Cereb Cortex 2025; 35:bhaf082. [PMID: 40264261 PMCID: PMC12014905 DOI: 10.1093/cercor/bhaf082] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2024] [Revised: 03/10/2025] [Accepted: 03/14/2025] [Indexed: 04/24/2025] Open
Abstract
Growing evidence indicates that the occipital place area (OPA) is involved in "visually guided navigation." Here, we propose that a recently uncovered scene-selective region in the superior parietal lobule is also involved in visually guided navigation. First, using functional magnetic resonance imaging (fMRI), we found that the superior parietal lobule (SPL) responds significantly more to scene stimuli than to face and object stimuli across two sets of stimuli (i.e. dynamic and static), confirming its scene selectivity. Second, we found that the SPL, like the OPA, processes two kinds of information necessary for visually guided navigation: first-person perspective motion and sense (left/right) information in scenes. Third, resting-state fMRI data revealed that SPL is preferentially connected to OPA, compared to other scene-selective regions, indicating that SPL and OPA are part of the same system. Fourth, analysis of previously published fMRI data showed that SPL, like OPA, responds significantly more while participants perform a visually guided navigation task compared to both a scene categorization task and a baseline task, further supporting our hypothesis in an independent dataset. Taken together, these findings indicate the existence of a new scene-selective region for visually guided navigation and raise interesting questions about the precise role that SPL, compared to OPA, may play within visually guided navigation.
Collapse
Affiliation(s)
- Hee Kyung Yoon
- Department of Psychology, Emory University, 36 Eagle Row, Atlanta, GA, 30322, United States
| | - Yaelan Jung
- Department of Psychology, Emory University, 36 Eagle Row, Atlanta, GA, 30322, United States
| | - Andrew S Persichetti
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, National Institute of Mental Health, National Institutes of Health, Bethesda, MD 20892, United States
| | - Daniel D Dilks
- Department of Psychology, Emory University, 36 Eagle Row, Atlanta, GA, 30322, United States
| |
Collapse
|
6
|
Steel A, Prasad D, Garcia BD, Robertson CE. Relating scene memory and perception activity to functional properties, networks, and landmarks of posterior cerebral cortex - a probabilistic atlas. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2025:2025.01.06.631538. [PMID: 39829755 PMCID: PMC11741410 DOI: 10.1101/2025.01.06.631538] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Indexed: 01/22/2025]
Abstract
Adaptive behavior in complex environments requires integrating visual perception with memory of our spatial environment. Recent work has implicated three brain areas in posterior cerebral cortex - the place memory areas (PMAs) that are anterior to the three visual scene perception areas (SPAs) - in this function. However, PMAs' relationship to the broader cortical hierarchy remains unclear due to limited group-level characterization. Here, we examined the PMA and SPA locations across three fMRI datasets (44 participants, 29 female). SPAs were identified using a standard visual localizer where participants viewed scenes versus faces. PMAs were identified by contrasting activity when participants recalled personally familiar places versus familiar faces (Datasets 1-2) or places versus multiple categories (familiar faces, bodies, and objects, and famous faces; Dataset 3). Across datasets, the PMAs were located anterior to the SPAs on the ventral and lateral cortical surfaces. The anterior displacement between PMAs and SPAs was highly reproducible. Compared to public atlases, the PMAs fell at the boundary between externally-oriented networks (dorsal attention) and internally-oriented networks (default mode). Additionally, while SPAs overlapped with retinotopic maps, the PMAs were consistently located anterior to mapped visual cortex. These results establish the anatomical position of the PMAs at inflection points along the cortical hierarchy between unimodal sensory and transmodal, apical regions, which informs broader theories of how the brain integrates perception and memory for scenes. We have released probabilistic parcels of these regions to facilitate future research into their roles in spatial cognition.
Collapse
Affiliation(s)
- Adam Steel
- Department of Psychology, University of Illinois
- Beckman Institute for Advanced Science and Technology, University of Illinois
| | | | - Brenda D. Garcia
- University of California San Diego Medical School, University of California San Diego
| | | |
Collapse
|
7
|
Mononen R, Saarela T, Vallinoja J, Olkkonen M, Henriksson L. Cortical Encoding of Spatial Structure and Semantic Content in 3D Natural Scenes. J Neurosci 2025; 45:e2157232024. [PMID: 39788741 PMCID: PMC11866997 DOI: 10.1523/jneurosci.2157-23.2024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2023] [Revised: 11/25/2024] [Accepted: 12/24/2024] [Indexed: 01/12/2025] Open
Abstract
Our visual system enables us to effortlessly navigate and recognize real-world visual environments. Functional magnetic resonance imaging (fMRI) studies suggest a network of scene-responsive cortical visual areas, but much less is known about the temporal order in which different scene properties are analyzed by the human visual system. In this study, we selected a set of 36 full-color natural scenes that varied in spatial structure and semantic content that our male and female human participants viewed both in 2D and 3D while we recorded magnetoencephalography (MEG) data. MEG enables tracking of cortical activity in humans at millisecond timescale. We compared the representational geometry in the MEG responses with predictions based on the scene stimuli using the representational similarity analysis framework. The representational structure first reflected the spatial structure in the scenes in time window 90-125 ms, followed by the semantic content in time window 140-175 ms after stimulus onset. The 3D stereoscopic viewing of the scenes affected the responses relatively late, from ∼140 ms from stimulus onset. Taken together, our results indicate that the human visual system rapidly encodes a scene's spatial structure and suggest that this information is based on monocular instead of binocular depth cues.
Collapse
Affiliation(s)
- Riikka Mononen
- Department of Neuroscience and Biomedical Engineering, Aalto University, Espoo FI-00076, Finland
- MEG Core, Aalto NeuroImaging, Aalto University, Espoo FI-00076, Finland
| | - Toni Saarela
- Department of Psychology, University of Helsinki, Helsinki FI-00014, Finland
| | - Jaakko Vallinoja
- Department of Neuroscience and Biomedical Engineering, Aalto University, Espoo FI-00076, Finland
- MEG Core, Aalto NeuroImaging, Aalto University, Espoo FI-00076, Finland
| | - Maria Olkkonen
- Department of Psychology, University of Helsinki, Helsinki FI-00014, Finland
| | - Linda Henriksson
- Department of Neuroscience and Biomedical Engineering, Aalto University, Espoo FI-00076, Finland
- MEG Core, Aalto NeuroImaging, Aalto University, Espoo FI-00076, Finland
| |
Collapse
|
8
|
Koc AN, Urgen BA, Afacan Y. Task-modulated neural responses in scene-selective regions of the human brain. Vision Res 2025; 227:108539. [PMID: 39733756 DOI: 10.1016/j.visres.2024.108539] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2023] [Revised: 10/29/2024] [Accepted: 12/20/2024] [Indexed: 12/31/2024]
Abstract
The study of scene perception is crucial to the understanding of how one interprets and interacts with their environment, and how the environment impacts various cognitive functions. The literature so far has mainly focused on the impact of low-level and categorical properties of scenes and how they are represented in the scene-selective regions in the brain, PPA, RSC, and OPA. However, higher-level scene perception and the impact of behavioral goals is a developing research area. Moreover, the selection of the stimuli has not been systematic and mainly focused on outdoor environments. In this fMRI experiment, we adopted multiple behavioral tasks, selected real-life indoor stimuli with a systematic categorization approach, and used various multivariate analysis techniques to explain the neural modulation of scene perception in the scene-selective regions of the human brain. Participants (N = 21) performed categorization and approach-avoidance tasks during fMRI scans while they were viewing scenes from built environment categories based on different affordances ((i)access and (ii)circulation elements, (iii)restrooms and (iv)eating/seating areas). ROI-based classification analysis revealed that the OPA was significantly successful in decoding scene category regardless of the task, and that the task condition affected category decoding performances of all the scene-selective regions. Model-based representational similarity analysis (RSA) revealed that the activity patterns in scene-selective regions are best explained by task. These results contribute to the literature by extending the task and stimulus content of scene perception research, and uncovering the impact of behavioral goals on the scene-selective regions of the brain.
Collapse
Affiliation(s)
- Aysu Nur Koc
- Department of Psychology, Justus Liebig University Giessen, Giessen, Germany; Interdisciplinary Neuroscience Program, Bilkent University, Ankara, Turkey.
| | - Burcu A Urgen
- Interdisciplinary Neuroscience Program, Bilkent University, Ankara, Turkey; Department of Psychology, Bilkent University, Ankara, Turkey; Aysel Sabuncu Brain Research Center and National Magnetic Resonance Imaging Center, Bilkent University, Ankara, Turkey.
| | - Yasemin Afacan
- Interdisciplinary Neuroscience Program, Bilkent University, Ankara, Turkey; Department of Interior Architecture and Environmental Design, Bilkent University, Ankara, Turkey; Aysel Sabuncu Brain Research Center and National Magnetic Resonance Imaging Center, Bilkent University, Ankara, Turkey.
| |
Collapse
|
9
|
Watson DM, Andrews TJ. Functional Connectivity of the Scene Processing Network at Rest Does Not Reliably Predict Human Behavior on Scene Processing Tasks. eNeuro 2025; 12:ENEURO.0375-24.2024. [PMID: 39890456 PMCID: PMC11820959 DOI: 10.1523/eneuro.0375-24.2024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2024] [Revised: 12/05/2024] [Accepted: 12/11/2024] [Indexed: 02/03/2025] Open
Abstract
The perception of scenes is associated with processing in a network of scene-selective regions in the human brain. Prior research has identified a posterior-anterior bias within this network. Posterior scene regions exhibit preferential connectivity with early visual and posterior parietal regions, indicating a role in representing egocentric visual features. In contrast, anterior scene regions demonstrate stronger connectivity with frontoparietal control and default mode networks, suggesting a role in mnemonic processing of locations. Despite these findings, evidence linking connectivity in these regions to cognitive scene processing remains limited. In this preregistered study, we obtained cognitive behavioral measures alongside resting-state fMRI data from a large-scale public dataset to investigate interindividual variation in scene processing abilities relative to the functional connectivity of the scene network. Our results revealed substantial individual differences in scene recognition, spatial memory, and navigational abilities. Resting-state functional connectivity reproduced the posterior-anterior bias within the scene network. However, contrary to our preregistered hypothesis, we did not observe any consistent associations between interindividual variation in this connectivity and behavioral performance. These findings highlight the need for further research to clarify the role of these connections in scene processing, potentially through assessments of functional connectivity during scene-relevant tasks or in naturalistic conditions.
Collapse
Affiliation(s)
- David M Watson
- Department of Psychology and York Neuroimaging Centre, University of York, York YO10 5DD, United Kingdom
| | - Timothy J Andrews
- Department of Psychology and York Neuroimaging Centre, University of York, York YO10 5DD, United Kingdom
| |
Collapse
|
10
|
Yao JK, Choo J, Finzi D, Grill-Spector K. Visuospatial computations vary by category and stream and continue to develop in adolescence. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2025:2025.01.14.633067. [PMID: 39868259 PMCID: PMC11761743 DOI: 10.1101/2025.01.14.633067] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Indexed: 01/28/2025]
Abstract
Reading, face recognition, and navigation are supported by visuospatial computations in category-selective regions across ventral, lateral, and dorsal visual streams. However, the nature of visuospatial computations across streams and their development in adolescence remain unknown. Using fMRI and population receptive field (pRF) modeling in adolescents and adults, we estimate pRFs in high-level visual cortex and determine their development. Results reveal that pRF location, size, and visual field coverage vary across category, stream, and hemisphere in both adolescents and adults. While pRF location is mature by adolescence, pRF size and visual field coverage continue to develop - increasing in face-selective and decreasing in place-selective regions - alongside similar development of category selectivity. These findings provide a timeline for differential development of visual functions and suggest that visuospatial computations in high-level visual cortex continue to be optimized to accommodate both category and stream demands through adolescence.
Collapse
Affiliation(s)
- Jewelia K Yao
- Department of Psychology, Stanford University, Stanford, CA 94305
| | - Justin Choo
- Department of Symbolic Systems, Stanford University, Stanford, CA, 94305
| | - Dawn Finzi
- Department of Psychology, Stanford University, Stanford, CA 94305
| | - Kalanit Grill-Spector
- Department of Psychology, Stanford University, Stanford, CA 94305
- Wu Tsai Neuroscience Institute, Stanford University, Stanford, CA 94305
| |
Collapse
|
11
|
Persichetti AS, Li TL, Stevens WD, Martin A, Gilmore AW. Atypical scene-selectivity in the retrosplenial complex in individuals with autism spectrum disorder. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.12.16.628702. [PMID: 39803432 PMCID: PMC11722278 DOI: 10.1101/2024.12.16.628702] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 01/26/2025]
Abstract
A small behavioral literature on individuals with autism spectrum disorder (ASD) has shown that they can be impaired when navigating using map-based strategies (i.e., memory-guided navigation), but not during visually guided navigation. Meanwhile, there is neuroimaging evidence in typically developing (TD) individuals demonstrating that the retrosplenial complex (RSC) is part of a memory-guided navigation system, while the occipital place area (OPA) is part of a visually-guided navigation system. A key identifying feature of the RSC and OPA is that they respond significantly more to pictures of places compared to faces or objects - i.e., they demonstrate scene-selectivity. Therefore, we predicted that scene-selectivity would be weaker in the RSC of individuals with ASD compared to a TD control group, while the OPA would not show such a difference between the groups. We used functional MRI to scan groups of ASD individuals and matched TD individuals while they viewed pictures of places and faces and performed a one-back task. As predicted, scene-selectivity was significantly lower in the RSC, but not OPA, in the ASD group compared to the TD group. These results suggest that impaired memory-guided navigation in individuals with ASD may, in part, be due to atypical functioning in the RSC.
Collapse
Affiliation(s)
- Andrew S. Persichetti
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, National Institute of Mental Health, National Institutes of Health, Bethesda, MD, USA
| | - Taylor L. Li
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, National Institute of Mental Health, National Institutes of Health, Bethesda, MD, USA
| | | | - Alex Martin
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, National Institute of Mental Health, National Institutes of Health, Bethesda, MD, USA
| | - Adrian W. Gilmore
- Department of Psychological and Brain Sciences, University of Delaware, Newark, DE, USA
| |
Collapse
|
12
|
Son G, Walther DB, Mack ML. Brief category learning distorts perceptual space for complex scenes. Psychon Bull Rev 2024; 31:2234-2248. [PMID: 38438711 DOI: 10.3758/s13423-024-02484-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 02/20/2024] [Indexed: 03/06/2024]
Abstract
The formation of categories is known to distort perceptual space: representations are pushed away from category boundaries and pulled toward categorical prototypes. This phenomenon has been studied with artificially constructed objects, whose feature dimensions are easily defined and manipulated. How such category-induced perceptual distortions arise for complex, real-world scenes, however, remains largely unknown due to the technical challenge of measuring and controlling scene features. We address this question by generating realistic scene images from a high-dimensional continuous space using generative adversarial networks and using the images as stimuli in a novel learning task. Participants learned to categorize the scene images along arbitrary category boundaries and later reconstructed the same scenes from memory. Systematic biases in reconstruction errors closely tracked each participant's subjective category boundaries. These findings suggest that the perception of global scene properties is warped to align with a newly learned category structure after only a brief learning experience.
Collapse
Affiliation(s)
- Gaeun Son
- Department of Psychology, University of Toronto, Toronto, Ontario, Canada.
| | - Dirk B Walther
- Department of Psychology, University of Toronto, Toronto, Ontario, Canada
| | - Michael L Mack
- Department of Psychology, University of Toronto, Toronto, Ontario, Canada
| |
Collapse
|
13
|
Naveilhan C, Saulay-Carret M, Zory R, Ramanoël S. Spatial Contextual Information Modulates Affordance Processing and Early Electrophysiological Markers of Scene Perception. J Cogn Neurosci 2024; 36:2084-2099. [PMID: 39023371 DOI: 10.1162/jocn_a_02223] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/20/2024]
Abstract
Scene perception allows humans to extract information from their environment and plan navigation efficiently. The automatic extraction of potential paths in a scene, also referred to as navigational affordance, is supported by scene-selective regions (SSRs) that enable efficient human navigation. Recent evidence suggests that the activity of these SSRs can be influenced by information from adjacent spatial memory areas. However, it remains unexplored how this contextual information could influence the extraction of bottom-up information, such as navigational affordances, from a scene and the underlying neural dynamics. Therefore, we analyzed ERPs in 26 young adults performing scene and spatial memory tasks in artificially generated rooms with varying numbers and locations of available doorways. We found that increasing the number of navigational affordances only impaired performance in the spatial memory task. ERP results showed a similar pattern of activity for both tasks, but with increased P2 amplitude in the spatial memory task compared with the scene memory. Finally, we reported no modulation of the P2 component by the number of affordances in either task. This modulation of early markers of visual processing suggests that the dynamics of SSR activity are influenced by a priori knowledge, with increased amplitude when participants have more contextual information about the perceived scene. Overall, our results suggest that prior spatial knowledge about the scene, such as the location of a goal, modulates early cortical activity associated with SSRs, and that this information may interact with bottom-up processing of scene content, such as navigational affordances.
Collapse
Affiliation(s)
| | | | - Raphaël Zory
- LAMHESS, Université Côte d'Azur, Nice, France
- Institut Universitaire de France (IUF)
| | - Stephen Ramanoël
- LAMHESS, Université Côte d'Azur, Nice, France
- INSERM, CNRS, Institut de la Vision, Sorbonne Université, Paris, France
| |
Collapse
|
14
|
Kang J, Park S. Combined representation of visual features in the scene-selective cortex. Behav Brain Res 2024; 471:115110. [PMID: 38871131 PMCID: PMC11375617 DOI: 10.1016/j.bbr.2024.115110] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2024] [Revised: 06/05/2024] [Accepted: 06/10/2024] [Indexed: 06/15/2024]
Abstract
Visual features of separable dimensions conjoin to represent an integrated entity. We investigated how visual features bind to form a complex visual scene. Specifically, we focused on features important for visually guided navigation: direction and distance. Previously, separate works have shown that directions and distances of navigable paths are coded in the occipital place area (OPA). Using functional magnetic resonance imaging (fMRI), we tested how separate features are concurrently represented in the OPA. Participants saw eight types of scenes, in which four of them had one path and the other four had two paths. In single-path scenes, path direction was either to the left or to the right. In double-path scenes, both directions were present. A glass wall was placed in some paths to restrict navigational distance. To test how the OPA represents path directions and distances, we took three approaches. First, the independent-features approach examined whether the OPA codes each direction and distance. Second, the integrated-features approach explored how directions and distances are integrated into path units, as compared to pooled features, using double-path scenes. Finally, the integrated-paths approach asked how separate paths are combined into a scene. Using multi-voxel pattern similarity analysis, we found that the OPA's representations of single-path scenes were similar to other single-path scenes of either the same direction or the same distance. Representations of double-path scenes were similar to the combination of two constituent single-paths, as a combined unit of direction and distance rather than as a pooled representation of all features. These results show that the OPA combines the two features to form path units, which are then used to build multiple-path scenes. Altogether, these results suggest that visually guided navigation may be supported by the OPA that automatically and efficiently combines multiple features relevant for navigation and represent a navigation file.
Collapse
Affiliation(s)
- Jisu Kang
- Department of Psychology, Yonsei University, 50, Yonsei-ro, Seodaemun-gu, Seoul 03722, the Republic of Korea
| | - Soojin Park
- Department of Psychology, Yonsei University, 50, Yonsei-ro, Seodaemun-gu, Seoul 03722, the Republic of Korea.
| |
Collapse
|
15
|
Tansey R, Graff K, Rai S, Merrikh D, Godfrey KJ, Vanderwal T, Bray S. Development of human visual cortical function: A scoping review of task- and naturalistic-fMRI studies through the interactive specialization and maturational frameworks. Neurosci Biobehav Rev 2024; 162:105729. [PMID: 38763178 DOI: 10.1016/j.neubiorev.2024.105729] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2024] [Revised: 05/12/2024] [Accepted: 05/14/2024] [Indexed: 05/21/2024]
Abstract
Overarching theories such as the interactive specialization and maturational frameworks have been proposed to describe human functional brain development. However, these frameworks have not yet been systematically examined across the fMRI literature. Visual processing is one of the most well-studied fields in neuroimaging, and research in this area has recently expanded to include naturalistic paradigms that facilitate study in younger age ranges, allowing for an in-depth critical appraisal of these frameworks across childhood. To this end, we conducted a scoping review of 94 developmental visual fMRI studies, including both traditional experimental task and naturalistic studies, across multiple sub-domains (early visual processing, category-specific higher order processing, naturalistic visual processing). We found that across domains, many studies reported progressive development, but few studies describe regressive or emergent changes necessary to fit the maturational or interactive specialization frameworks. Our findings suggest a need for the expansion of developmental frameworks and clearer reporting of both progressive and regressive changes, along with well-powered, longitudinal studies.
Collapse
Affiliation(s)
- Ryann Tansey
- Department of Psychiatry, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada; Alberta Children's Hospital Research Institute, University of Calgary, Calgary, AB, Canada; Hotchkiss Brain Institute, University of Calgary, Calgary, AB, Canada.
| | - Kirk Graff
- Alberta Children's Hospital Research Institute, University of Calgary, Calgary, AB, Canada; Hotchkiss Brain Institute, University of Calgary, Calgary, AB, Canada; Department of Psychiatry, Washington University in St. Louis, St. Louis, MO, USA
| | - Shefali Rai
- Alberta Children's Hospital Research Institute, University of Calgary, Calgary, AB, Canada; Hotchkiss Brain Institute, University of Calgary, Calgary, AB, Canada
| | - Daria Merrikh
- Alberta Children's Hospital Research Institute, University of Calgary, Calgary, AB, Canada; Hotchkiss Brain Institute, University of Calgary, Calgary, AB, Canada
| | - Kate J Godfrey
- Alberta Children's Hospital Research Institute, University of Calgary, Calgary, AB, Canada; Hotchkiss Brain Institute, University of Calgary, Calgary, AB, Canada
| | - Tamara Vanderwal
- Department of Psychiatry, University of British Columbia, Vancouver, BC, Canada; BC Children's Hospital Research Institute, Vancouver, BC, Canada
| | - Signe Bray
- Alberta Children's Hospital Research Institute, University of Calgary, Calgary, AB, Canada; Hotchkiss Brain Institute, University of Calgary, Calgary, AB, Canada; Department of Radiology, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada; Department of Pediatrics, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
| |
Collapse
|
16
|
Park J, Soucy E, Segawa J, Mair R, Konkle T. Immersive scene representation in human visual cortex with ultra-wide-angle neuroimaging. Nat Commun 2024; 15:5477. [PMID: 38942766 PMCID: PMC11213904 DOI: 10.1038/s41467-024-49669-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2023] [Accepted: 06/13/2024] [Indexed: 06/30/2024] Open
Abstract
While human vision spans 220°, traditional functional MRI setups display images only up to central 10-15°. Thus, it remains unknown how the brain represents a scene perceived across the full visual field. Here, we introduce a method for ultra-wide angle display and probe signatures of immersive scene representation. An unobstructed view of 175° is achieved by bouncing the projected image off angled-mirrors onto a custom-built curved screen. To avoid perceptual distortion, scenes are created with wide field-of-view from custom virtual environments. We find that immersive scene representation drives medial cortex with far-peripheral preferences, but shows minimal modulation in classic scene regions. Further, scene and face-selective regions maintain their content preferences even with extreme far-periphery stimulation, highlighting that not all far-peripheral information is automatically integrated into scene regions computations. This work provides clarifying evidence on content vs. peripheral preferences in scene representation and opens new avenues to research immersive vision.
Collapse
Affiliation(s)
- Jeongho Park
- Department of Psychology, Harvard University, Cambridge, MA, USA.
| | - Edward Soucy
- Center for Brain Science, Harvard University, Cambridge, MA, USA
| | - Jennifer Segawa
- Center for Brain Science, Harvard University, Cambridge, MA, USA
| | - Ross Mair
- Center for Brain Science, Harvard University, Cambridge, MA, USA
- Department of Radiology, Harvard Medical School, Boston, MA, USA
- Department of Radiology, Massachusetts General Hospital, Boston, MA, USA
| | - Talia Konkle
- Department of Psychology, Harvard University, Cambridge, MA, USA
- Center for Brain Science, Harvard University, Cambridge, MA, USA
- Kempner Institute for Biological and Artificial Intelligence, Harvard University, Boston, MA, USA
| |
Collapse
|
17
|
Kamps FS, Chen EM, Kanwisher N, Saxe R. Representation of navigational affordances and ego-motion in the occipital place area. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.04.30.591964. [PMID: 38746251 PMCID: PMC11092631 DOI: 10.1101/2024.04.30.591964] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2024]
Abstract
Humans effortlessly use vision to plan and guide navigation through the local environment, or "scene". A network of three cortical regions responds selectively to visual scene information, including the occipital (OPA), parahippocampal (PPA), and medial place areas (MPA) - but how this network supports visually-guided navigation is unclear. Recent evidence suggests that one region in particular, the OPA, supports visual representations for navigation, while PPA and MPA support other aspects of scene processing. However, most previous studies tested only static scene images, which lack the dynamic experience of navigating through scenes. We used dynamic movie stimuli to test whether OPA, PPA, and MPA represent two critical kinds of navigationally-relevant information: navigational affordances (e.g., can I walk to the left, right, or both?) and ego-motion (e.g., am I walking forward or backward? turning left or right?). We found that OPA is sensitive to both affordances and ego-motion, as well as the conflict between these cues - e.g., turning toward versus away from an open doorway. These effects were significantly weaker or absent in PPA and MPA. Responses in OPA were also dissociable from those in early visual cortex, consistent with the idea that OPA responses are not merely explained by lower-level visual features. OPA responses to affordances and ego-motion were stronger in the contralateral than ipsilateral visual field, suggesting that OPA encodes navigationally relevant information within an egocentric reference frame. Taken together, these results support the hypothesis that OPA contains visual representations that are useful for planning and guiding navigation through scenes.
Collapse
|
18
|
McMullin MA, Kumar R, Higgins NC, Gygi B, Elhilali M, Snyder JS. Preliminary Evidence for Global Properties in Human Listeners During Natural Auditory Scene Perception. Open Mind (Camb) 2024; 8:333-365. [PMID: 38571530 PMCID: PMC10990578 DOI: 10.1162/opmi_a_00131] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2023] [Accepted: 02/10/2024] [Indexed: 04/05/2024] Open
Abstract
Theories of auditory and visual scene analysis suggest the perception of scenes relies on the identification and segregation of objects within it, resembling a detail-oriented processing style. However, a more global process may occur while analyzing scenes, which has been evidenced in the visual domain. It is our understanding that a similar line of research has not been explored in the auditory domain; therefore, we evaluated the contributions of high-level global and low-level acoustic information to auditory scene perception. An additional aim was to increase the field's ecological validity by using and making available a new collection of high-quality auditory scenes. Participants rated scenes on 8 global properties (e.g., open vs. enclosed) and an acoustic analysis evaluated which low-level features predicted the ratings. We submitted the acoustic measures and average ratings of the global properties to separate exploratory factor analyses (EFAs). The EFA of the acoustic measures revealed a seven-factor structure explaining 57% of the variance in the data, while the EFA of the global property measures revealed a two-factor structure explaining 64% of the variance in the data. Regression analyses revealed each global property was predicted by at least one acoustic variable (R2 = 0.33-0.87). These findings were extended using deep neural network models where we examined correlations between human ratings of global properties and deep embeddings of two computational models: an object-based model and a scene-based model. The results support that participants' ratings are more strongly explained by a global analysis of the scene setting, though the relationship between scene perception and auditory perception is multifaceted, with differing correlation patterns evident between the two models. Taken together, our results provide evidence for the ability to perceive auditory scenes from a global perspective. Some of the acoustic measures predicted ratings of global scene perception, suggesting representations of auditory objects may be transformed through many stages of processing in the ventral auditory stream, similar to what has been proposed in the ventral visual stream. These findings and the open availability of our scene collection will make future studies on perception, attention, and memory for natural auditory scenes possible.
Collapse
Affiliation(s)
| | - Rohit Kumar
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD, USA
| | - Nathan C. Higgins
- Department of Communication Sciences & Disorders, University of South Florida, Tampa, FL, USA
| | - Brian Gygi
- East Bay Institute for Research and Education, Martinez, CA, USA
| | - Mounya Elhilali
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD, USA
| | - Joel S. Snyder
- Department of Psychology, University of Nevada, Las Vegas, Las Vegas, NV, USA
| |
Collapse
|
19
|
Jung Y, Hsu D, Dilks DD. "Walking selectivity" in the occipital place area in 8-year-olds, not 5-year-olds. Cereb Cortex 2024; 34:bhae101. [PMID: 38494889 PMCID: PMC10945045 DOI: 10.1093/cercor/bhae101] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2023] [Revised: 02/20/2024] [Accepted: 02/21/2024] [Indexed: 03/19/2024] Open
Abstract
A recent neuroimaging study in adults found that the occipital place area (OPA)-a cortical region involved in "visually guided navigation" (i.e. moving about the immediately visible environment, avoiding boundaries, and obstacles)-represents visual information about walking, not crawling, suggesting that OPA is late developing, emerging only when children are walking, not beforehand. But when precisely does this "walking selectivity" in OPA emerge-when children first begin to walk in early childhood, or perhaps counterintuitively, much later in childhood, around 8 years of age, when children are adult-like walking? To directly test these two hypotheses, using functional magnetic resonance imaging (fMRI) in two groups of children, 5- and 8-year-olds, we measured the responses in OPA to first-person perspective videos through scenes from a "walking" perspective, as well as three control perspectives ("crawling," "flying," and "scrambled"). We found that the OPA in 8-year-olds-like adults-exhibited walking selectivity (i.e. responding significantly more to the walking videos than to any of the others, and no significant differences across the crawling, flying, and scrambled videos), while the OPA in 5-year-olds exhibited no walking selectively. These findings reveal that OPA undergoes protracted development, with walking selectivity only emerging around 8 years of age.
Collapse
Affiliation(s)
- Yaelan Jung
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
| | - Debbie Hsu
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
| | - Daniel D Dilks
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
| |
Collapse
|
20
|
Park J, Soucy E, Segawa J, Mair R, Konkle T. Immersive scene representation in human visual cortex with ultra-wide angle neuroimaging. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2023.05.14.540275. [PMID: 37292806 PMCID: PMC10245572 DOI: 10.1101/2023.05.14.540275] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
While humans experience the visual environment in a panoramic 220° view, traditional functional MRI setups are limited to display images like postcards in the central 10-15° of the visual field. Thus, it remains unknown how a scene is represented in the brain when perceived across the full visual field. Here, we developed a novel method for ultra-wide angle visual presentation and probed for signatures of immersive scene representation. To accomplish this, we bounced the projected image off angled-mirrors directly onto a custom-built curved screen, creating an unobstructed view of 175°. Scene images were created from custom-built virtual environments with a compatible wide field-of-view to avoid perceptual distortion. We found that immersive scene representation drives medial cortex with far-peripheral preferences, but surprisingly had little effect on classic scene regions. That is, scene regions showed relatively minimal modulation over dramatic changes of visual size. Further, we found that scene and face-selective regions maintain their content preferences even under conditions of central scotoma, when only the extreme far-peripheral visual field is stimulated. These results highlight that not all far-peripheral information is automatically integrated into the computations of scene regions, and that there are routes to high-level visual areas that do not require direct stimulation of the central visual field. Broadly, this work provides new clarifying evidence on content vs. peripheral preferences in scene representation, and opens new neuroimaging research avenues to understand immersive visual representation.
Collapse
Affiliation(s)
| | | | | | - Ross Mair
- Center for Brain Science, Harvard University
- Department of Radiology, Harvard Medical School
- Department of Radiology, Massachusetts General Hospital
| | - Talia Konkle
- Department of Psychology, Harvard University
- Center for Brain Science, Harvard University
- Kempner Institute for Biological and Artificial Intelligence, Harvard University
| |
Collapse
|
21
|
Chai XJ, Tang L, Gabrieli JDE, Ofen N. From vision to memory: How scene-sensitive regions support episodic memory formation during child development. Dev Cogn Neurosci 2024; 65:101340. [PMID: 38218015 PMCID: PMC10825658 DOI: 10.1016/j.dcn.2024.101340] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2022] [Revised: 12/21/2023] [Accepted: 01/02/2024] [Indexed: 01/15/2024] Open
Abstract
Previous brain imaging studies have identified three brain regions that selectively respond to visual scenes, the parahippocampal place area (PPA), the occipital place area (OPA), and the retrosplenial cortex (RSC). There is growing evidence that these visual scene-sensitive regions process different types of scene information and may have different developmental timelines in supporting scene perception. How these scene-sensitive regions support memory functions during child development is largely unknown. We investigated PPA, OPA and RSC activations associated with episodic memory formation in childhood (5-7 years of age) and young adulthood, using a subsequent scene memory paradigm and a functional localizer for scenes. PPA, OPA, and RSC subsequent memory activation and functional connectivity differed between children and adults. Subsequent memory effects were found in activations of all three scene regions in adults. In children, however, robust subsequent memory effects were only found in the PPA. Functional connectivity during successful encoding was significant among the three regions in adults, but not in children. PPA subsequently memory activations and PPA-RSC subsequent memory functional connectivity correlated with accuracy in adults, but not children. These age-related differences add new evidence linking protracted development of the scene-sensitive regions to the protracted development of episodic memory.
Collapse
Affiliation(s)
- Xiaoqian J Chai
- Department of Neurology and Neurosurgery, McGill University, USA.
| | - Lingfei Tang
- Department of Psychology and the Institute of Gerontology, Wayne State University, USA
| | - John DE Gabrieli
- Department of Brain and Cognitive Sciences and McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA, USA
| | - Noa Ofen
- Department of Psychology and the Institute of Gerontology, Wayne State University, USA; Center for Vital Longevity and School of Behavioral and Brain Sciences, University of Texas at Dallas, Dallas, TX, USA.
| |
Collapse
|
22
|
Agron AM, Martin A, Gilmore AW. Scene construction and autobiographical memory retrieval in autism spectrum disorder. Autism Res 2024; 17:204-214. [PMID: 38037250 PMCID: PMC10922094 DOI: 10.1002/aur.3066] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2023] [Accepted: 11/06/2023] [Indexed: 12/02/2023]
Abstract
Individuals with autism spectrum disorder (ASD) frequently exhibit difficulties in retrieving autobiographical memories (AMs) of specific events from their life. Such memory deficits are frequently attributed to underlying disruptions in self-referential or social cognition processes. This makes intuitive sense as these are hallmarks of ASD. However, an emerging literature suggests that parallel deficits also exist in ASD individuals' ability to reconstruct the rich spatial contexts in which events occur. This is a capacity known as scene construction, and in typically developing individuals is considered a core process in retrieving AMs. In this review, we discuss evidence of difficulties with scene construction in ASD, drawing upon experiments that involve AM retrieval, other forms of mental time travel, and spatial navigation. We also highlight aspects of extant data that cannot be accounted for using purely social explanations of memory deficits in ASD. We conclude by identifying key questions raised by our framework and suggest how they might be addressed in future research.
Collapse
Affiliation(s)
- Anna M. Agron
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, NIMH/NIH, Bethesda, MD 20892
| | - Alex Martin
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, NIMH/NIH, Bethesda, MD 20892
| | - Adrian W. Gilmore
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, NIMH/NIH, Bethesda, MD 20892
| |
Collapse
|
23
|
Persichetti AS, Shao J, Gotts SJ, Martin A. A functional parcellation of the whole brain in individuals with autism spectrum disorder reveals atypical patterns of network organization. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.12.15.571854. [PMID: 38168156 PMCID: PMC10760210 DOI: 10.1101/2023.12.15.571854] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/05/2024]
Abstract
BACKGROUND Researchers studying autism spectrum disorder (ASD) lack a comprehensive map of the functional network topography in the ASD brain. We used high-quality resting state functional MRI (rs-fMRI) connectivity data and a robust parcellation routine to provide a whole-brain map of functional networks in a group of seventy individuals with ASD and a group of seventy typically developing (TD) individuals. METHODS The rs-fMRI data were collected using an imaging sequence optimized to achieve high temporal signal-to-noise ratio (tSNR) across the whole-brain. We identified functional networks using a parcellation routine that intrinsically incorporates stability and replicability of the networks by keeping only network distinctions that agree across halves of the data over multiple random iterations in each group. The groups were tightly matched on tSNR, in-scanner motion, age, and IQ. RESULTS We compared the maps from each group and found that functional networks in the ASD group are atypical in three seemingly related ways: 1) whole-brain connectivity patterns are less stable across voxels within multiple functional networks, 2) the cerebellum, subcortex, and hippocampus show weaker differentiation of functional subnetworks, and 3) subcortical structures and the hippocampus are atypically integrated with the neocortex. CONCLUSIONS These results were statistically robust and suggest that patterns of network connectivity between the neocortex and the cerebellum, subcortical structures, and hippocampus are atypical in ASD individuals.
Collapse
Affiliation(s)
- Andrew S Persichetti
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, National Institute of Mental Health, National Institutes of Health, Bethesda, Maryland 20892
| | - Jiayu Shao
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, National Institute of Mental Health, National Institutes of Health, Bethesda, Maryland 20892
| | - Stephen J Gotts
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, National Institute of Mental Health, National Institutes of Health, Bethesda, Maryland 20892
| | - Alex Martin
- Section on Cognitive Neuropsychology, Laboratory of Brain and Cognition, National Institute of Mental Health, National Institutes of Health, Bethesda, Maryland 20892
| |
Collapse
|
24
|
Dilks DD, Jung Y, Kamps FS. The development of human cortical scene processing. CURRENT DIRECTIONS IN PSYCHOLOGICAL SCIENCE 2023; 32:479-486. [PMID: 38283826 PMCID: PMC10815932 DOI: 10.1177/09637214231191772] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/30/2024]
Abstract
Decades of research have uncovered the neural basis of place (or "scene") processing in adulthood, revealing a set of three regions that respond selectively to visual scene information, each hypothesized to support distinct functions within scene processing (e.g., recognizing a particular kind of place versus navigating through it). Despite this considerable progress, surprisingly little is known about how these cortical regions develop. Here we review the limited evidence to date, highlighting the first few studies exploring the origins of cortical scene processing in infancy, and the several studies addressing when the scene regions reach full maturity, unfortunately with inconsistent findings. This inconsistency likely stems from common pitfalls in pediatric functional magnetic resonance imaging, and accordingly, we discuss how these pitfalls may be avoided. Furthermore, we point out that almost all studies to date have focused only on general scene selectivity and argue that greater insight could be gleaned by instead exploring the more distinct functions of each region, as well as their connectivity. Finally, with this last point in mind, we offer a novel hypothesis that scene regions supporting navigation (including the occipital place area and retrosplenial complex) mature later than those supporting scene categorization (including the parahippocampal place area).
Collapse
Affiliation(s)
- Daniel D. Dilks
- Department of Psychology, Emory University, Atlanta, Georgia, USA
| | - Yaelan Jung
- Department of Psychology, Emory University, Atlanta, Georgia, USA
| | - Frederik S. Kamps
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, Massachusetts, USA
| |
Collapse
|
25
|
Molloy MF, Osher DE. A personalized cortical atlas for functional regions of interest. J Neurophysiol 2023; 130:1067-1080. [PMID: 37727907 PMCID: PMC10994647 DOI: 10.1152/jn.00108.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2023] [Revised: 09/18/2023] [Accepted: 09/18/2023] [Indexed: 09/21/2023] Open
Abstract
Advances in functional MRI (fMRI) allow mapping an individual's brain function in vivo. Task fMRI can localize domain-specific regions of cognitive processing or functional regions of interest (fROIs) within an individual. Moreover, data from resting state (no task) fMRI can be used to define an individual's connectome, which can characterize that individual's functional organization via connectivity-based parcellations. However, can connectivity-based parcellations alone predict an individual's fROIs? Here, we describe an approach to compute individualized rs-fROIs (i.e., regions that correspond to given fROI constructed using only resting state data) for motor control, working memory, high-level vision, and language comprehension. The rs-fROIs were computed and validated using a large sample of young adults (n = 1,018) with resting state and task fMRI from the Human Connectome Project. First, resting state parcellations were defined across a sequence of resolutions from broadscale to fine-grained networks in a training group of 500 individuals. Second, 21 rs-fROIs were defined from the training group by identifying the rs network that most closely matched task-defined fROIs across all individuals. Third, the selectivity of rs-fROIs was investigated in a training set of the remaining 518 individuals. All computed rs-fROIs were indeed selective for their preferred category. Critically, the rs-fROIs had higher selectivity than probabilistic atlas parcels for nearly all fROIs. In conclusion, we present a potential approach to define selective fROIs on an individual-level circumventing the need for multiple task-based localizers.NEW & NOTEWORTHY We compute individualized resting state parcels that identify an individual's own functional regions of interest (fROIs) for high-level vision, language comprehension, motor control, and working memory, using only their functional connectome. This approach demonstrates a rapid and powerful alternative for finding a large set of fROIs in an individual, using only their unique connectivity pattern, which does not require the costly acquisition of multiple fMRI localizer tasks.
Collapse
Affiliation(s)
- M. Fiona Molloy
- Department of Psychology, The Ohio State University, Columbus, Ohio, United States
- Department of Psychiatry, University of Michigan, Ann Arbor, Michigan, United States
| | - David E. Osher
- Department of Psychology, The Ohio State University, Columbus, Ohio, United States
| |
Collapse
|
26
|
Kamps FS, Rennert RJ, Radwan SF, Wahab S, Pincus JE, Dilks DD. Dissociable Cognitive Systems for Recognizing Places and Navigating through Them: Developmental and Neuropsychological Evidence. J Neurosci 2023; 43:6320-6329. [PMID: 37580121 PMCID: PMC10490455 DOI: 10.1523/jneurosci.0153-23.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2023] [Revised: 07/11/2023] [Accepted: 08/03/2023] [Indexed: 08/16/2023] Open
Abstract
Recent neural evidence suggests that the human brain contains dissociable systems for "scene categorization" (i.e., recognizing a place as a particular kind of place, for example, a kitchen), including the parahippocampal place area, and "visually guided navigation" (e.g., finding our way through a kitchen, not running into the kitchen walls or banging into the kitchen table), including the occipital place area. However, converging behavioral data - for instance, whether scene categorization and visually guided navigation abilities develop along different timelines and whether there is differential breakdown under neurologic deficit - would provide even stronger support for this two-scene-systems hypothesis. Thus, here we tested scene categorization and visually guided navigation abilities in 131 typically developing children between 4 and 9 years of age, as well as 46 adults with Williams syndrome, a developmental disorder with known impairment on "action" tasks, yet relative sparing on "perception" tasks, in object processing. We found that (1) visually guided navigation is later to develop than scene categorization, and (2) Williams syndrome adults are impaired in visually guided navigation, but not scene categorization, relative to mental age-matched children. Together, these findings provide the first developmental and neuropsychological evidence for dissociable cognitive systems for recognizing places and navigating through them.SIGNIFICANCE STATEMENT Two decades ago, Milner and Goodale showed us that identifying objects and manipulating them involve distinct cognitive and neural systems. Recent neural evidence suggests that the same may be true of our interactions with our environment: identifying places and navigating through them are dissociable systems. Here we provide converging behavioral evidence supporting this two-scene-systems hypothesis - finding both differential development and breakdown of "scene categorization" and "visually guided navigation." This finding suggests that the division of labor between perception and action systems is a general organizing principle for the visual system, not just a principle of the object processing system in particular.
Collapse
Affiliation(s)
- Frederik S Kamps
- Department of Psychology, Emory University, Atlanta, Georgia 30322
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139
| | | | - Samaher F Radwan
- Department of Psychology, Emory University, Atlanta, Georgia 30322
| | - Stephanie Wahab
- Department of Psychology, Emory University, Atlanta, Georgia 30322
| | - Jordan E Pincus
- Department of Psychology, Emory University, Atlanta, Georgia 30322
| | - Daniel D Dilks
- Department of Psychology, Emory University, Atlanta, Georgia 30322
| |
Collapse
|
27
|
Steel A, Garcia BD, Goyal K, Mynick A, Robertson CE. Scene Perception and Visuospatial Memory Converge at the Anterior Edge of Visually Responsive Cortex. J Neurosci 2023; 43:5723-5737. [PMID: 37474310 PMCID: PMC10401646 DOI: 10.1523/jneurosci.2043-22.2023] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2022] [Revised: 07/10/2023] [Accepted: 07/14/2023] [Indexed: 07/22/2023] Open
Abstract
To fluidly engage with the world, our brains must simultaneously represent both the scene in front of us and our memory of the immediate surrounding environment (i.e., local visuospatial context). How does the brain's functional architecture enable sensory and mnemonic representations to closely interface while also avoiding sensory-mnemonic interference? Here, we asked this question using first-person, head-mounted virtual reality and fMRI. Using virtual reality, human participants of both sexes learned a set of immersive, real-world visuospatial environments in which we systematically manipulated the extent of visuospatial context associated with a scene image in memory across three learning conditions, spanning from a single FOV to a city street. We used individualized, within-subject fMRI to determine which brain areas support memory of the visuospatial context associated with a scene during recall (Experiment 1) and recognition (Experiment 2). Across the whole brain, activity in three patches of cortex was modulated by the amount of known visuospatial context, each located immediately anterior to one of the three scene perception areas of high-level visual cortex. Individual subject analyses revealed that these anterior patches corresponded to three functionally defined place memory areas, which selectively respond when visually recalling personally familiar places. In addition to showing activity levels that were modulated by the amount of visuospatial context, multivariate analyses showed that these anterior areas represented the identity of the specific environment being recalled. Together, these results suggest a convergence zone for scene perception and memory of the local visuospatial context at the anterior edge of high-level visual cortex.SIGNIFICANCE STATEMENT As we move through the world, the visual scene around us is integrated with our memory of the wider visuospatial context. Here, we sought to understand how the functional architecture of the brain enables coexisting representations of the current visual scene and memory of the surrounding environment. Using a combination of immersive virtual reality and fMRI, we show that memory of visuospatial context outside the current FOV is represented in a distinct set of brain areas immediately anterior and adjacent to the perceptually oriented scene-selective areas of high-level visual cortex. This functional architecture would allow efficient interaction between immediately adjacent mnemonic and perceptual areas while also minimizing interference between mnemonic and perceptual representations.
Collapse
Affiliation(s)
- Adam Steel
- Department of Psychological & Brain Sciences, Dartmouth College, Hanover, New Hampshire 03755
| | - Brenda D Garcia
- Department of Psychological & Brain Sciences, Dartmouth College, Hanover, New Hampshire 03755
| | - Kala Goyal
- Department of Psychological & Brain Sciences, Dartmouth College, Hanover, New Hampshire 03755
| | - Anna Mynick
- Department of Psychological & Brain Sciences, Dartmouth College, Hanover, New Hampshire 03755
| | - Caroline E Robertson
- Department of Psychological & Brain Sciences, Dartmouth College, Hanover, New Hampshire 03755
| |
Collapse
|
28
|
Jones CM, Byland J, Dilks DD. The occipital place area represents visual information about walking, not crawling. Cereb Cortex 2023; 33:7500-7505. [PMID: 36918999 PMCID: PMC10267618 DOI: 10.1093/cercor/bhad055] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2022] [Revised: 02/06/2023] [Accepted: 02/07/2023] [Indexed: 03/16/2023] Open
Abstract
Recent work has shown that the occipital place area (OPA)-a scene-selective region in adult humans-supports "visually guided navigation" (i.e. moving about the local visual environment and avoiding boundaries/obstacles). But what is the precise role of OPA in visually guided navigation? Considering humans move about their local environments beginning with crawling followed by walking, 1 possibility is that OPA is involved in both modes of locomotion. Another possibility is that OPA is specialized for walking only, since walking and crawling are different kinds of locomotion. To test these possibilities, we measured the responses in OPA to first-person perspective videos from both "walking" and "crawling" perspectives as well as for 2 conditions by which humans do not navigate ("flying" and "scrambled"). We found that OPA responded more to walking videos than to any of the others, including crawling, and did not respond more to crawling videos than to flying or scrambled ones. These results (i) reveal that OPA represents visual information only from a walking (not crawling) perspective, (ii) suggest crawling is processed by a different neural system, and (iii) raise questions for how OPA develops; namely, OPA may have never supported crawling, which is consistent with the hypothesis that OPA undergoes protracted development.
Collapse
Affiliation(s)
- Christopher M Jones
- Department of Psychology, Emory University, Atlanta, GA 30322, United States
| | - Joshua Byland
- Department of Psychology, Emory University, Atlanta, GA 30322, United States
| | - Daniel D Dilks
- Department of Psychology, Emory University, Atlanta, GA 30322, United States
| |
Collapse
|
29
|
Kay K, Bonnen K, Denison RN, Arcaro MJ, Barack DL. Tasks and their role in visual neuroscience. Neuron 2023; 111:1697-1713. [PMID: 37040765 DOI: 10.1016/j.neuron.2023.03.022] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2023] [Revised: 03/13/2023] [Accepted: 03/15/2023] [Indexed: 04/13/2023]
Abstract
Vision is widely used as a model system to gain insights into how sensory inputs are processed and interpreted by the brain. Historically, careful quantification and control of visual stimuli have served as the backbone of visual neuroscience. There has been less emphasis, however, on how an observer's task influences the processing of sensory inputs. Motivated by diverse observations of task-dependent activity in the visual system, we propose a framework for thinking about tasks, their role in sensory processing, and how we might formally incorporate tasks into our models of vision.
Collapse
Affiliation(s)
- Kendrick Kay
- Center for Magnetic Resonance Research, Department of Radiology, University of Minnesota, Minneapolis, MN 55455, USA.
| | - Kathryn Bonnen
- School of Optometry, Indiana University, Bloomington, IN 47405, USA
| | - Rachel N Denison
- Department of Psychological and Brain Sciences, Boston University, Boston, MA 02215, USA
| | - Mike J Arcaro
- Department of Psychology, University of Pennsylvania, Philadelphia, PA 19146, USA
| | - David L Barack
- Departments of Neuroscience and Philosophy, University of Pennsylvania, Philadelphia, PA 19146, USA
| |
Collapse
|
30
|
Liang Q, Liao J, Li J, Zheng S, Jiang X, Huang R. The role of the parahippocampal cortex in landmark-based distance estimation based on the contextual hypothesis. Hum Brain Mapp 2023; 44:131-141. [PMID: 36066186 PMCID: PMC9783420 DOI: 10.1002/hbm.26069] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2021] [Revised: 07/30/2022] [Accepted: 08/13/2022] [Indexed: 02/05/2023] Open
Abstract
Parahippocampal cortex (PHC) is a vital neural bases in spatial navigation. However, its functional role is still unclear. "Contextual hypothesis," which assumes that the PHC participates in processing the spatial association between the landmark and destination, provides a potential answer to the question. Nevertheless, the hypothesis was previously tested using the picture categorization task, which is indirectly related to spatial navigation. By now, study is still needed for testing the hypothesis with a navigation-related paradigm. In the current study, we tested the hypothesis by an fMRI experiment in which participants performed a distance estimation task in a virtual environment under three different conditions: landmark free (LF), stable landmark (SL), and ambiguous landmark (AL). By analyzing the behavioral data, we found that the presence of an SL improved the participants' performance in distance estimation. Comparing the brain activity in SL-versus-LF contrast as well as AL-versus-LF contrast, we found that the PHC was activated by the SL rather than by AL when encoding the distance. This indicates that the PHC is elicited by strongly associated context and encodes the landmark reference for distance perception. Furthermore, accessing the representational similarity with the activity of the PHC across conditions, we observed a high similarity within the same condition but low similarity between conditions. This result indicated that the PHC sustains the contextual information for discriminating between scenes. Our findings provided insights into the neural correlates of the landmark information processing from the perspective of contextual hypothesis.
Collapse
Affiliation(s)
- Qunjun Liang
- School of Psychology, Center for Studies of Psychological Application, Guangdong Key Laboratory of Mental Health and Cognitive Science, Ministry of Education Key Laboratory of Brain Cognition and Educational ScienceSouth China Normal UniversityGuangzhouGuangdongChina
| | - Jiajun Liao
- School of Psychology, Center for Studies of Psychological Application, Guangdong Key Laboratory of Mental Health and Cognitive Science, Ministry of Education Key Laboratory of Brain Cognition and Educational ScienceSouth China Normal UniversityGuangzhouGuangdongChina
| | - Jinhui Li
- School of Psychology, Center for Studies of Psychological Application, Guangdong Key Laboratory of Mental Health and Cognitive Science, Ministry of Education Key Laboratory of Brain Cognition and Educational ScienceSouth China Normal UniversityGuangzhouGuangdongChina
| | - Senning Zheng
- School of Psychology, Center for Studies of Psychological Application, Guangdong Key Laboratory of Mental Health and Cognitive Science, Ministry of Education Key Laboratory of Brain Cognition and Educational ScienceSouth China Normal UniversityGuangzhouGuangdongChina
| | - Xiaoqian Jiang
- School of Psychology, Center for Studies of Psychological Application, Guangdong Key Laboratory of Mental Health and Cognitive Science, Ministry of Education Key Laboratory of Brain Cognition and Educational ScienceSouth China Normal UniversityGuangzhouGuangdongChina
| | - Ruiwang Huang
- School of Psychology, Center for Studies of Psychological Application, Guangdong Key Laboratory of Mental Health and Cognitive Science, Ministry of Education Key Laboratory of Brain Cognition and Educational ScienceSouth China Normal UniversityGuangzhouGuangdongChina
| |
Collapse
|
31
|
Park J, Josephs E, Konkle T. Ramp-shaped neural tuning supports graded population-level representation of the object-to-scene continuum. Sci Rep 2022; 12:18081. [PMID: 36302932 PMCID: PMC9613906 DOI: 10.1038/s41598-022-21768-2] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/05/2022] [Accepted: 09/30/2022] [Indexed: 01/24/2023] Open
Abstract
We can easily perceive the spatial scale depicted in a picture, regardless of whether it is a small space (e.g., a close-up view of a chair) or a much larger space (e.g., an entire class room). How does the human visual system encode this continuous dimension? Here, we investigated the underlying neural coding of depicted spatial scale, by examining the voxel tuning and topographic organization of brain responses. We created naturalistic yet carefully-controlled stimuli by constructing virtual indoor environments, and rendered a series of snapshots to smoothly sample between a close-up view of the central object and far-scale view of the full environment (object-to-scene continuum). Human brain responses were measured to each position using functional magnetic resonance imaging. We did not find evidence for a smooth topographic mapping for the object-to-scene continuum on the cortex. Instead, we observed large swaths of cortex with opposing ramp-shaped profiles, with highest responses to one end of the object-to-scene continuum or the other, and a small region showing a weak tuning to intermediate scale views. However, when we considered the population code of the entire ventral occipito-temporal cortex, we found smooth and linear representation of the object-to-scene continuum. Our results together suggest that depicted spatial scale information is encoded parametrically in large-scale population codes across the entire ventral occipito-temporal cortex.
Collapse
Affiliation(s)
- Jeongho Park
- Department of Psychology, Harvard University, Cambridge, USA.
| | - Emilie Josephs
- Computer Science & Artificial Intelligence Lab, Massachusetts Institute of Technology, Cambridge, USA
| | - Talia Konkle
- Department of Psychology, Harvard University, Cambridge, USA
| |
Collapse
|
32
|
Tang K, Chin M, Chun M, Xu Y. The contribution of object identity and configuration to scene representation in convolutional neural networks. PLoS One 2022; 17:e0270667. [PMID: 35763531 PMCID: PMC9239439 DOI: 10.1371/journal.pone.0270667] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2022] [Accepted: 06/14/2022] [Indexed: 11/23/2022] Open
Abstract
Scene perception involves extracting the identities of the objects comprising a scene in conjunction with their configuration (the spatial layout of the objects in the scene). How object identity and configuration information is weighted during scene processing and how this weighting evolves over the course of scene processing however, is not fully understood. Recent developments in convolutional neural networks (CNNs) have demonstrated their aptitude at scene processing tasks and identified correlations between processing in CNNs and in the human brain. Here we examined four CNN architectures (Alexnet, Resnet18, Resnet50, Densenet161) and their sensitivity to changes in object and configuration information over the course of scene processing. Despite differences among the four CNN architectures, across all CNNs, we observed a common pattern in the CNN's response to object identity and configuration changes. Each CNN demonstrated greater sensitivity to configuration changes in early stages of processing and stronger sensitivity to object identity changes in later stages. This pattern persists regardless of the spatial structure present in the image background, the accuracy of the CNN in classifying the scene, and even the task used to train the CNN. Importantly, CNNs' sensitivity to a configuration change is not the same as their sensitivity to any type of position change, such as that induced by a uniform translation of the objects without a configuration change. These results provide one of the first documentations of how object identity and configuration information are weighted in CNNs during scene processing.
Collapse
Affiliation(s)
- Kevin Tang
- Department of Psychology, Yale University, New Haven, CT, United States of America
| | - Matthew Chin
- Department of Psychology, Yale University, New Haven, CT, United States of America
| | - Marvin Chun
- Department of Psychology, Yale University, New Haven, CT, United States of America
| | - Yaoda Xu
- Department of Psychology, Yale University, New Haven, CT, United States of America
- * E-mail:
| |
Collapse
|