51
|
Cheng A, Walther DB, Park S, Dilks DD. Concavity as a diagnostic feature of visual scenes. Neuroimage 2021; 232:117920. [PMID: 33652147 PMCID: PMC8256888 DOI: 10.1016/j.neuroimage.2021.117920] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2020] [Revised: 02/07/2021] [Accepted: 02/21/2021] [Indexed: 11/29/2022] Open
Abstract
Despite over two decades of research on the neural mechanisms underlying human visual scene, or place, processing, it remains unknown what exactly a “scene” is. Intuitively, we are always inside a scene, while interacting with the outside of objects. Hence, we hypothesize that one diagnostic feature of a scene may be concavity, portraying “inside”, and predict that if concavity is a scene-diagnostic feature, then: 1) images that depict concavity, even non-scene images (e.g., the “inside” of an object – or concave object), will be behaviorally categorized as scenes more often than those that depict convexity, and 2) the cortical scene-processing system will respond more to concave images than to convex images. As predicted, participants categorized concave objects as scenes more often than convex objects, and, using functional magnetic resonance imaging (fMRI), two scene-selective cortical regions (the parahippocampal place area, PPA, and the occipital place area, OPA) responded significantly more to concave than convex objects. Surprisingly, we found no behavioral or neural differences between images of concave versus convex buildings. However, in a follow-up experiment, using tightly-controlled images, we unmasked a selective sensitivity to concavity over convexity of scene boundaries (i.e., walls) in PPA and OPA. Furthermore, we found that even highly impoverished line drawings of concave shapes are behaviorally categorized as scenes more often than convex shapes. Together, these results provide converging behavioral and neural evidence that concavity is a diagnostic feature of visual scenes.
Collapse
Affiliation(s)
- Annie Cheng
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
| | - Dirk B Walther
- Department of Psychology, University of Toronto, Toronto, ON, Canada
| | - Soojin Park
- Department of Psychology, Yonsei University, Seoul, Republic of Korea.
| | - Daniel D Dilks
- Department of Psychology, Emory University, Atlanta, GA 30322, USA.
| |
Collapse
|
52
|
Abstract
Rapid visual perception is often viewed as a bottom-up process. Category-preferred neural regions are often characterized as automatic, default processing mechanisms for visual inputs of their categorical preference. To explore the sensitivity of such regions to top-down information, we examined three scene-preferring brain regions, the occipital place area (OPA), the parahippocampal place area (PPA), and the retrosplenial complex (RSC) and tested whether the processing of outdoor scenes is influenced by the functional contexts in which they are seen. Context was manipulated by presenting real-world landscape images as if being viewed through a window or within a picture frame-manipulations that do not affect scene content but do affect one's functional knowledge regarding the scene. This manipulation influences neural scene processing (as measured by fMRI): The OPA and the PPA exhibited greater neural activity when participants viewed images as if through a window as compared with within a picture frame, whereas the RSC did not show this difference. In a separate behavioral experiment, functional context affected scene memory in predictable directions (boundary extension). Our interpretation is that the window context denotes three dimensionality, therefore rendering the perceptual experience of viewing landscapes as more realistic. Conversely, the frame context denotes a 2-D image. As such, more spatially biased scene representations in the OPA and the PPA are influenced by differences in top-down, perceptual expectations generated from context. In contrast, more semantically biased scene representations in the RSC are likely to be less affected by top-down signals that carry information about the physical layout of a scene.
Collapse
|
53
|
Altomare EC, Committeri G, Di Matteo R, Capotosto P, Tosoni A. Automatic coding of environmental distance for walking-related locomotion in the foot-related sensory-motor system: A TMS study on macro-affordances. Neuropsychologia 2020; 150:107696. [PMID: 33253691 DOI: 10.1016/j.neuropsychologia.2020.107696] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2020] [Revised: 11/17/2020] [Accepted: 11/25/2020] [Indexed: 11/25/2022]
Abstract
We have recently described a facilitation effect for the execution of a walking-related action in response to distant objects/locations in the extrapersonal space. Based on the parallelism with the well-known effect of "micro-affordance", observed during the execution of functionally appropriate hand-related actions towards manipulable objects, we have referred to this effect in terms of "macro-affordance". Here we used transcranical magnetic stimulation (TMS) to investigate whether a foot-related region located in the human dorsal precuneate cortex plays a causal role in the generation and maintenance of such behavioral effect. This question was addressed by comparing the magnitude of the facilitation effect during an incidental go/no-go task, i.e. advantage for walking-related actions to pictures framing an environment from a far vs. near distance, during three different TMS conditions. The three TMS conditions were collected in all subjects in a randomized order and included stimulation of: i. a foot-related region in the anterior precuneus, ii. a control region in the middle intraparietal sulcus (mIPS), and iii. a sham condition. Enrollment in the TMS protocol was based on analysis of individual performance during a preliminary session conducted using a sham stimulation. TMS was administered at a low frequency range before the beginning of each condition. The results showed that stimulation of the foot-related region in the anterior precuneus produced a significant reduction of the walking-related facilitation effect as compared to both stimulation of the active-control region and the non-active sham stimulation. These findings suggest that the foot-related sensory-motor system directly participates in the process of extraction of the spatial features (i.e. distance) from an environmental scene that are useful for locomotion. More in general, these findings support an automatic coding of environmental affordance or "macro-affordances" in the walking-related sensory-motor system.
Collapse
Affiliation(s)
- Emanuele Cosimo Altomare
- Department of Neuroscience, Imaging and Clinical Sciences, Institute for Advanced Biomedical Technologies (ITAB), University G. D'Annunzio, Chieti-Pescara, Italy
| | - Giorgia Committeri
- Department of Neuroscience, Imaging and Clinical Sciences, Institute for Advanced Biomedical Technologies (ITAB), University G. D'Annunzio, Chieti-Pescara, Italy
| | - Rosalia Di Matteo
- Department of Neuroscience, Imaging and Clinical Sciences, Institute for Advanced Biomedical Technologies (ITAB), University G. D'Annunzio, Chieti-Pescara, Italy
| | - Paolo Capotosto
- Department of Neuroscience, Imaging and Clinical Sciences, Institute for Advanced Biomedical Technologies (ITAB), University G. D'Annunzio, Chieti-Pescara, Italy
| | - Annalisa Tosoni
- Department of Neuroscience, Imaging and Clinical Sciences, Institute for Advanced Biomedical Technologies (ITAB), University G. D'Annunzio, Chieti-Pescara, Italy.
| |
Collapse
|
54
|
Josephs EL, Konkle T. Large-scale dissociations between views of objects, scenes, and reachable-scale environments in visual cortex. Proc Natl Acad Sci U S A 2020; 117:29354-29362. [PMID: 33229533 PMCID: PMC7703543 DOI: 10.1073/pnas.1912333117] [Citation(s) in RCA: 25] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Space-related processing recruits a network of brain regions separate from those recruited in object processing. This dissociation has largely been explored by contrasting views of navigable-scale spaces to views of close-up, isolated objects. However, in naturalistic visual experience, we encounter spaces intermediate to these extremes, like the tops of desks and kitchen counters, which are not navigable but typically contain multiple objects. How are such reachable-scale views represented in the brain? In three human functional neuroimaging experiments, we find evidence for a large-scale dissociation of reachable-scale views from both navigable scene views and close-up object views. Three brain regions were identified that showed a systematic response preference to reachable views, located in the posterior collateral sulcus, the inferior parietal sulcus, and superior parietal lobule. Subsequent analyses suggest that these three regions may be especially sensitive to the presence of multiple objects. Further, in all classic scene and object regions, reachable-scale views dissociated from both objects and scenes with an intermediate response magnitude. Taken together, these results establish that reachable-scale environments have a distinct representational signature from both scene and object views in visual cortex.
Collapse
Affiliation(s)
- Emilie L Josephs
- Department of Psychology, Harvard University, Cambridge, MA 02138
| | - Talia Konkle
- Department of Psychology, Harvard University, Cambridge, MA 02138
| |
Collapse
|
55
|
Nestor A, Lee ACH, Plaut DC, Behrmann M. The Face of Image Reconstruction: Progress, Pitfalls, Prospects. Trends Cogn Sci 2020; 24:747-759. [PMID: 32674958 PMCID: PMC7429291 DOI: 10.1016/j.tics.2020.06.006] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2020] [Revised: 05/27/2020] [Accepted: 06/15/2020] [Indexed: 10/23/2022]
Abstract
Recent research has demonstrated that neural and behavioral data acquired in response to viewing face images can be used to reconstruct the images themselves. However, the theoretical implications, promises, and challenges of this direction of research remain unclear. We evaluate the potential of this research for elucidating the visual representations underlying face recognition. Specifically, we outline complementary and converging accounts of the visual content, the representational structure, and the neural dynamics of face processing. We illustrate how this research addresses fundamental questions in the study of normal and impaired face recognition, and how image reconstruction provides a powerful framework for uncovering face representations, for unifying multiple types of empirical data, and for facilitating both theoretical and methodological progress.
Collapse
Affiliation(s)
- Adrian Nestor
- Department of Psychology at Scarborough, University of Toronto, Toronto, Ontario, Canada.
| | - Andy C H Lee
- Department of Psychology at Scarborough, University of Toronto, Toronto, Ontario, Canada; Rotman Research Institute, Baycrest Centre, Toronto, Ontario, Canada
| | - David C Plaut
- Department of Psychology, Carnegie Mellon University, Pittsburgh, PA, USA; Carnegie Mellon Neuroscience Institute, Pittsburgh, PA, USA
| | - Marlene Behrmann
- Department of Psychology, Carnegie Mellon University, Pittsburgh, PA, USA; Carnegie Mellon Neuroscience Institute, Pittsburgh, PA, USA
| |
Collapse
|
56
|
Nau M, Navarro Schröder T, Frey M, Doeller CF. Behavior-dependent directional tuning in the human visual-navigation network. Nat Commun 2020; 11:3247. [PMID: 32591544 PMCID: PMC7320013 DOI: 10.1038/s41467-020-17000-2] [Citation(s) in RCA: 23] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2019] [Accepted: 06/05/2020] [Indexed: 01/06/2023] Open
Abstract
The brain derives cognitive maps from sensory experience that guide memory formation and behavior. Despite extensive efforts, it still remains unclear how the underlying population activity unfolds during spatial navigation and how it relates to memory performance. To examine these processes, we combined 7T-fMRI with a kernel-based encoding model of virtual navigation to map world-centered directional tuning across the human cortex. First, we present an in-depth analysis of directional tuning in visual, retrosplenial, parahippocampal and medial temporal cortices. Second, we show that tuning strength, width and topology of this directional code during memory-guided navigation depend on successful encoding of the environment. Finally, we show that participants' locomotory state influences this tuning in sensory and mnemonic regions such as the hippocampus. We demonstrate a direct link between neural population tuning and human cognition, where high-level memory processing interacts with network-wide visuospatial coding in the service of behavior.
Collapse
Affiliation(s)
- Matthias Nau
- Kavli Institute for Systems Neuroscience, Centre for Neural Computation, The Egil and Pauline Braathen and Fred Kavli Centre for Cortical Microcircuits, NTNU, Trondheim, Norway.
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.
| | - Tobias Navarro Schröder
- Kavli Institute for Systems Neuroscience, Centre for Neural Computation, The Egil and Pauline Braathen and Fred Kavli Centre for Cortical Microcircuits, NTNU, Trondheim, Norway
| | - Markus Frey
- Kavli Institute for Systems Neuroscience, Centre for Neural Computation, The Egil and Pauline Braathen and Fred Kavli Centre for Cortical Microcircuits, NTNU, Trondheim, Norway
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| | - Christian F Doeller
- Kavli Institute for Systems Neuroscience, Centre for Neural Computation, The Egil and Pauline Braathen and Fred Kavli Centre for Cortical Microcircuits, NTNU, Trondheim, Norway.
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.
| |
Collapse
|
57
|
Horikawa T, Cowen AS, Keltner D, Kamitani Y. The Neural Representation of Visually Evoked Emotion Is High-Dimensional, Categorical, and Distributed across Transmodal Brain Regions. iScience 2020; 23:101060. [PMID: 32353765 PMCID: PMC7191651 DOI: 10.1016/j.isci.2020.101060] [Citation(s) in RCA: 36] [Impact Index Per Article: 7.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2019] [Revised: 03/11/2020] [Accepted: 04/09/2020] [Indexed: 12/12/2022] Open
Abstract
Central to our subjective lives is the experience of different emotions. Recent behavioral work mapping emotional responses to 2,185 videos found that people experience upward of 27 distinct emotions occupying a high-dimensional space, and that emotion categories, more so than affective dimensions (e.g., valence), organize self-reports of subjective experience. Here, we sought to identify the neural substrates of this high-dimensional space of emotional experience using fMRI responses to all 2,185 videos. Our analyses demonstrated that (1) dozens of video-evoked emotions were accurately predicted from fMRI patterns in multiple brain regions with different regional configurations for individual emotions; (2) emotion categories better predicted cortical and subcortical responses than affective dimensions, outperforming visual and semantic covariates in transmodal regions; and (3) emotion-related fMRI responses had a cluster-like organization efficiently characterized by distinct categories. These results support an emerging theory of the high-dimensional emotion space, illuminating its neural foundations distributed across transmodal regions.
Collapse
Affiliation(s)
- Tomoyasu Horikawa
- Department of Neuroinformatics, ATR Computational Neuroscience Laboratories, Hikaridai, Seika, Soraku, Kyoto, 619-0288, Japan.
| | - Alan S Cowen
- Department of Psychology, University of California, Berkeley, CA 94720-1500, USA
| | - Dacher Keltner
- Department of Psychology, University of California, Berkeley, CA 94720-1500, USA
| | - Yukiyasu Kamitani
- Department of Neuroinformatics, ATR Computational Neuroscience Laboratories, Hikaridai, Seika, Soraku, Kyoto, 619-0288, Japan; Graduate School of Informatics, Kyoto University, Yoshida-honmachi, Sakyo-ku, Kyoto, 606-8501, Japan.
| |
Collapse
|
58
|
Coding of Navigational Distance and Functional Constraint of Boundaries in the Human Scene-Selective Cortex. J Neurosci 2020; 40:3621-3630. [PMID: 32209608 DOI: 10.1523/jneurosci.1991-19.2020] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2019] [Revised: 02/28/2020] [Accepted: 03/05/2020] [Indexed: 11/21/2022] Open
Abstract
For visually guided navigation, the use of environmental cues is essential. Particularly, detecting local boundaries that impose limits to locomotion and estimating their location is crucial. In a series of three fMRI experiments, we investigated whether there is a neural coding of navigational distance in the human visual cortex (both female and male). We used virtual reality software to systematically manipulate the distance from a viewer perspective to different types of a boundary. Using a multivoxel pattern classification employing a linear support vector machine, we found that the occipital place area (OPA) is sensitive to the navigational distance restricted by the transparent glass wall. Further, the OPA was sensitive to a non-crossable boundary only, suggesting an importance of the functional constraint of a boundary. Together, we propose the OPA as a perceptual source of external environmental features relevant for navigation.SIGNIFICANCE STATEMENT One of major goals in cognitive neuroscience has been to understand the nature of visual scene representation in human ventral visual cortex. An aspect of scene perception that has been overlooked despite its ecological importance is the analysis of space for navigation. One of critical computation necessary for navigation is coding of distance to environmental boundaries that impose limit on navigator's movements. This paper reports the first empirical evidence for coding of navigational distance in the human visual cortex and its striking sensitivity to functional constraint of environmental boundaries. Such finding links the paper to previous neurological and behavioral works that emphasized the distance to boundaries as a crucial geometric property for reorientation behavior of children and other animal species.
Collapse
|
59
|
Chang TY, Doudlah R, Kim B, Sunkara A, Thompson LW, Lowe ME, Rosenberg A. Functional links between sensory representations, choice activity, and sensorimotor associations in parietal cortex. eLife 2020; 9:57968. [PMID: 33078705 PMCID: PMC7641584 DOI: 10.7554/elife.57968] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2020] [Accepted: 10/19/2020] [Indexed: 02/02/2023] Open
Abstract
Three-dimensional (3D) representations of the environment are often critical for selecting actions that achieve desired goals. The success of these goal-directed actions relies on 3D sensorimotor transformations that are experience-dependent. Here we investigated the relationships between the robustness of 3D visual representations, choice-related activity, and motor-related activity in parietal cortex. Macaque monkeys performed an eight-alternative 3D orientation discrimination task and a visually guided saccade task while we recorded from the caudal intraparietal area using laminar probes. We found that neurons with more robust 3D visual representations preferentially carried choice-related activity. Following the onset of choice-related activity, the robustness of the 3D representations further increased for those neurons. We additionally found that 3D orientation and saccade direction preferences aligned, particularly for neurons with choice-related activity, reflecting an experience-dependent sensorimotor association. These findings reveal previously unrecognized links between the fidelity of ecologically relevant object representations, choice-related activity, and motor-related activity.
Collapse
Affiliation(s)
- Ting-Yu Chang
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin–MadisonMadisonUnited States
| | - Raymond Doudlah
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin–MadisonMadisonUnited States
| | - Byounghoon Kim
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin–MadisonMadisonUnited States
| | | | - Lowell W Thompson
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin–MadisonMadisonUnited States
| | - Meghan E Lowe
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin–MadisonMadisonUnited States
| | - Ari Rosenberg
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin–MadisonMadisonUnited States
| |
Collapse
|
60
|
Sprague TC, Boynton GM, Serences JT. The Importance of Considering Model Choices When Interpreting Results in Computational Neuroimaging. eNeuro 2019; 6:ENEURO.0196-19.2019. [PMID: 31772033 PMCID: PMC6924997 DOI: 10.1523/eneuro.0196-19.2019] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2019] [Revised: 09/23/2019] [Accepted: 11/18/2019] [Indexed: 11/21/2022] Open
Abstract
Model-based analyses open exciting opportunities for understanding neural information processing. In a commentary published in eNeuro, Gardner and Liu (2019) discuss the role of model specification in interpreting results derived from complex models of neural data. As a case study, they suggest that one such analysis, the inverted encoding model (IEM), should not be used to assay properties of "stimulus representations" because the ability to apply linear transformations at various stages of the analysis procedure renders results "arbitrary." Here, we argue that the specification of all models is arbitrary to the extent that an experimenter makes choices based on current knowledge of the model system. However, the results derived from any given model, such as the reconstructed channel response profiles obtained from an IEM analysis, are uniquely defined and are arbitrary only in the sense that changes in the model can predictably change results. IEM-based channel response profiles should therefore not be considered arbitrary when the model is clearly specified and guided by our best understanding of neural population representations in the brain regions being analyzed. Intuitions derived from this case study are important to consider when interpreting results from all model-based analyses, which are similarly contingent upon the specification of the models used.
Collapse
Affiliation(s)
- Thomas C Sprague
- Department of Psychological and Brain Sciences, University of California, Santa Barbara, Santa Barbara, CA 93106-9660
| | - Geoffrey M Boynton
- Department of Psychology, University of Washington, Seattle, WA 98195-1525
| | - John T Serences
- Department of Psychology, University of California San Diego, La Jolla, CA 92093-0109
- Neurosciences Graduate Program, University of California San Diego, La Jolla, CA 92093-0109
- Kavli Foundation for the Brain and Mind, University of California San Diego, La Jolla, CA 92093-0126
| |
Collapse
|
61
|
Abstract
Three cortical brain regions are thought to underlie our remarkable ability to perceive and understand visual scenes. In this issue of Neuron, Lescroart and Gallant (2018) use quantitative models of scene processing to reveal 3D representations in these regions.
Collapse
|
62
|
Coggan DD, Giannakopoulou A, Ali S, Goz B, Watson DM, Hartley T, Baker DH, Andrews TJ. A data-driven approach to stimulus selection reveals an image-based representation of objects in high-level visual areas. Hum Brain Mapp 2019; 40:4716-4731. [PMID: 31338936 DOI: 10.1002/hbm.24732] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2019] [Revised: 06/25/2019] [Accepted: 06/26/2019] [Indexed: 11/07/2022] Open
Abstract
The ventral visual pathway is directly involved in the perception and recognition of objects. However, the extent to which the neural representation of objects in this region reflects low-level or high-level properties remains unresolved. A problem in resolving this issue is that only a small proportion of the objects experienced during natural viewing can be shown during a typical experiment. This can lead to an uneven sampling of objects that biases our understanding of how they are represented. To address this issue, we developed a data-driven approach to stimulus selection that involved describing a large number objects in terms of their image properties. In the first experiment, clusters of objects were evenly selected from this multi-dimensional image space. Although the clusters did not have any consistent semantic features, each elicited a distinct pattern of neural response. In the second experiment, we asked whether high-level, category-selective patterns of response could be elicited by objects from other categories, but with similar image properties. Object clusters were selected based on the similarity of their image properties to objects from five different categories (bottle, chair, face, house, and shoe). The pattern of response to each metameric object cluster was similar to the pattern elicited by objects from the corresponding category. For example, the pattern for bottles was similar to the pattern for objects with similar image properties to bottles. In both experiments, the patterns of response were consistent across participants providing evidence for common organising principles. This study provides a more ecological approach to understanding the perceptual representations of objects and reveals the importance of image properties.
Collapse
Affiliation(s)
| | | | - Sanah Ali
- Department of Psychology, University of York, York, UK
| | - Burcu Goz
- Department of Psychology, University of York, York, UK
| | - David M Watson
- School of Psychology, The University of Nottingham, Nottingham, UK
| | - Tom Hartley
- Department of Psychology, University of York, York, UK.,York Biomedical Research Institute, University of York, York, UK
| | - Daniel H Baker
- Department of Psychology, University of York, York, UK.,York Biomedical Research Institute, University of York, York, UK
| | | |
Collapse
|
63
|
Abstract
Humans are remarkably adept at perceiving and understanding complex real-world scenes. Uncovering the neural basis of this ability is an important goal of vision science. Neuroimaging studies have identified three cortical regions that respond selectively to scenes: parahippocampal place area, retrosplenial complex/medial place area, and occipital place area. Here, we review what is known about the visual and functional properties of these brain areas. Scene-selective regions exhibit retinotopic properties and sensitivity to low-level visual features that are characteristic of scenes. They also mediate higher-level representations of layout, objects, and surface properties that allow individual scenes to be recognized and their spatial structure ascertained. Challenges for the future include developing computational models of information processing in scene regions, investigating how these regions support scene perception under ecologically realistic conditions, and understanding how they operate in the context of larger brain networks.
Collapse
Affiliation(s)
- Russell A Epstein
- Department of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA;
| | - Chris I Baker
- Section on Learning and Plasticity, Laboratory of Brain and Cognition, National Institute of Mental Health, Bethesda, Maryland 20892, USA;
| |
Collapse
|
64
|
Henriksson L, Mur M, Kriegeskorte N. Rapid Invariant Encoding of Scene Layout in Human OPA. Neuron 2019; 103:161-171.e3. [PMID: 31097360 DOI: 10.1016/j.neuron.2019.04.014] [Citation(s) in RCA: 38] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2018] [Revised: 03/13/2019] [Accepted: 04/05/2019] [Indexed: 01/30/2023]
Abstract
Successful visual navigation requires a sense of the geometry of the local environment. How do our brains extract this information from retinal images? Here we visually presented scenes with all possible combinations of five scene-bounding elements (left, right, and back walls; ceiling; floor) to human subjects during functional magnetic resonance imaging (fMRI) and magnetoencephalography (MEG). The fMRI response patterns in the scene-responsive occipital place area (OPA) reflected scene layout with invariance to changes in surface texture. This result contrasted sharply with the primary visual cortex (V1), which reflected low-level image features of the stimuli, and the parahippocampal place area (PPA), which showed better texture than layout decoding. MEG indicated that the texture-invariant scene layout representation is computed from visual input within ∼100 ms, suggesting a rapid computational mechanism. Taken together, these results suggest that the cortical representation underlying our instant sense of the environmental geometry is located in the OPA.
Collapse
Affiliation(s)
- Linda Henriksson
- Department of Neuroscience and Biomedical Engineering, Aalto University, 02150 Espoo, Finland; AMI Centre, MEG Core, ABL, Aalto NeuroImaging, Aalto University, 02150 Espoo, Finland.
| | - Marieke Mur
- MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge CB2 7EF, UK; Department of Psychology, Brain and Mind Institute, Western University, London, ON N6A 3K7, Canada
| | - Nikolaus Kriegeskorte
- MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge CB2 7EF, UK; Department of Psychology, Department of Neuroscience, and Department of Electrical Engineering, Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY 10032, USA
| |
Collapse
|
65
|
Nag S, Berman D, Golomb JD. Category-selective areas in human visual cortex exhibit preferences for stimulus depth. Neuroimage 2019; 196:289-301. [PMID: 30978498 DOI: 10.1016/j.neuroimage.2019.04.025] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2018] [Revised: 03/21/2019] [Accepted: 04/07/2019] [Indexed: 12/01/2022] Open
Abstract
Multiple regions in the human brain are dedicated to accomplish the feat of object recognition; yet our brains must also compute the 2D and 3D locations of the objects we encounter in order to make sense of our visual environments. A number of studies have explored how various object category-selective regions are sensitive to and have preferences for specific 2D spatial locations in addition to processing their preferred-stimulus categories, but there is no survey of how these regions respond to depth information. In a blocked functional MRI experiment, subjects viewed a series of category-specific (i.e., faces, objects, scenes) and unspecific (e.g., random moving dots) stimuli with red/green anaglyph glasses. Critically, these stimuli were presented at different depth planes such that they appeared in front of, behind, or at the same (i.e., middle) depth plane as the fixation point (Experiment 1) or simultaneously in front of and behind fixation (i.e., mixed depth; Experiment 2). Comparisons of mean response magnitudes between back, middle, and front depth planes reveal that face and object regions OFA and LOC exhibit a preference for front depths, and motion area MT+ exhibits a strong linear preference for front, followed by middle, followed by back depth planes. In contrast, scene-selective regions PPA and OPA prefer front and/or back depth planes (relative to middle). Moreover, the occipital place area demonstrates a strong preference for "mixed" depth above and beyond back alone, raising potential implications about its particular role in scene perception. Crucially, the observed depth preferences in nearly all areas were evoked irrespective of the semantic stimulus category being viewed. These results reveal that the object category-selective regions may play a role in processing or incorporating depth information that is orthogonal to their primary processing of object category information.
Collapse
Affiliation(s)
- Samoni Nag
- Department of Psychology, Center for Cognitive & Brain Sciences, The Ohio State University, USA; Department of Psychology, The George Washington University, USA
| | - Daniel Berman
- Department of Psychology, Center for Cognitive & Brain Sciences, The Ohio State University, USA
| | - Julie D Golomb
- Department of Psychology, Center for Cognitive & Brain Sciences, The Ohio State University, USA.
| |
Collapse
|
66
|
Brady TF, Störmer VS, Shafer-Skelton A, Williams JR, Chapman AF, Schill HM. Scaling up visual attention and visual working memory to the real world. PSYCHOLOGY OF LEARNING AND MOTIVATION 2019. [DOI: 10.1016/bs.plm.2019.03.001] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/16/2023]
|