1
|
Wang R, Zhao B, Chen A. The visual representation of 3D orientation in macaque areas STPp and VPS. J Physiol 2025; 603:1541-1566. [PMID: 39949109 DOI: 10.1113/jp287309] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/15/2024] [Accepted: 01/23/2025] [Indexed: 03/15/2025] Open
Abstract
In the current study, we investigated the neural mechanisms underlying the representation of three-dimensional (3D) surface orientation within the posterior portion of the superior temporal polysensory area (STPp) and the visual posterior Sylvian area (VPS) in the macaque brain. Both areas are known for their integration of visual and vestibular signals, which are crucial for visual stability and spatial perception. However, it remains unclear how exactly these areas represent the orientation of 3D surfaces. To tackle this question, we used random dot stereograms (RDS) to present 3D planar stimuli defined by slant and tilt, with depth via binocular disparity. Through this method, we examined how STPp and VPS encode this information. Our results suggest that both regions encode the orientation and depth of 3D surfaces, with interactions among these parameters influencing neural responses. Additionally, we investigated how motion cues affect the perception of 3D surface orientation. STPp consistently encoded plane orientation information regardless of motion cue, whereas VPS responses showed less stability. These findings shed light on the distinct processing mechanisms for 3D spatial information in different cortical areas, offering insights into the neural basis of visual stability and spatial perception. KEY POINTS: Both STPp and VPS can encode 3D surface orientation. Slant is encoded independently from tilt and disparity in STPp and VPS areas. TDD neurons shift their depth preferences based on tilt in STPp and VPS areas. STPp maintains stable 3D orientation encoding under motion conditions, while VPS shows less stability with changes in tilt and disparity preferences.
Collapse
Affiliation(s)
- Rong Wang
- Key Laboratory of Brain Functional Genomics (Ministry of Education), East China Normal University, Shanghai, China
| | - Bin Zhao
- Key Laboratory of Brain Functional Genomics (Ministry of Education), East China Normal University, Shanghai, China
- Lingang Laboratory, Shanghai, China
| | - Aihua Chen
- Key Laboratory of Brain Functional Genomics (Ministry of Education), East China Normal University, Shanghai, China
- NYU-ECNU Institute of Brain and Cognitive Science, New York University Shanghai, Shanghai, China
| |
Collapse
|
2
|
Herrera-Esposito D, Burge J. Optimal Estimation of Local Motion-in-Depth with Naturalistic Stimuli. J Neurosci 2025; 45:e0490242024. [PMID: 39592236 PMCID: PMC11841760 DOI: 10.1523/jneurosci.0490-24.2024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2024] [Revised: 10/30/2024] [Accepted: 11/06/2024] [Indexed: 11/28/2024] Open
Abstract
Estimating the motion of objects in depth is important for behavior and is strongly supported by binocular visual cues. To understand both how the brain should estimate motion in depth and how natural constraints shape and limit performance in two local 3D motion tasks, we develop image-computable ideal observers from a large number of binocular video clips created from a dataset of natural images. The observers spatiotemporally filter the videos and nonlinearly decode 3D motion from the filter responses. The optimal filters and decoder are dictated by the task-relevant image statistics and are specific to each task. Multiple findings emerge. First, two distinct filter subpopulations are spontaneously learned for each task. For 3D speed estimation, filters emerge for processing either changing disparities over time or interocular velocity differences, cues that are used by humans. For 3D direction estimation, filters emerge for discriminating either left-right or toward-away motion. Second, the filter responses, conditioned on the latent variable, are well-described as jointly Gaussian, and the covariance of the filter responses carries the information about the task-relevant latent variable. Quadratic combination is thus necessary for optimal decoding, which can be implemented by biologically plausible neural computations. Finally, the ideal observer yields nonobvious-and in some cases counterintuitive-patterns of performance like those exhibited by humans. Important characteristics of human 3D motion processing and estimation may therefore result from optimal information processing in the early visual system.
Collapse
Affiliation(s)
| | - Johannes Burge
- Department of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania 19104
- Neuroscience Graduate Group, University of Pennsylvania, Philadelphia, Pennsylvania 19104
- Bioengineering Graduate Group, University of Pennsylvania, Philadelphia, Pennsylvania 19104
| |
Collapse
|
3
|
Nielsen KJ, Connor CE. How Shape Perception Works, in Two Dimensions and Three Dimensions. Annu Rev Vis Sci 2024; 10:47-68. [PMID: 38848596 DOI: 10.1146/annurev-vision-112823-031607] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/09/2024]
Abstract
The ventral visual pathway transforms retinal images into neural representations that support object understanding, including exquisite appreciation of precise 2D pattern shape and 3D volumetric shape. We articulate a framework for understanding the goals of this transformation and how they are achieved by neural coding at successive ventral pathway stages. The critical goals are (a) radical compression to make shape information communicable across axonal bundles and storable in memory, (b) explicit coding to make shape information easily readable by the rest of the brain and thus accessible for cognition and behavioral control, and (c) representational stability to maintain consistent perception across highly variable viewing conditions. We describe how each transformational step in ventral pathway vision serves one or more of these goals. This three-goal framework unifies discoveries about ventral shape processing into a neural explanation for our remarkable experience of shape as a vivid, richly detailed aspect of the natural world.
Collapse
Affiliation(s)
- Kristina J Nielsen
- Krieger Mind/Brain Institute and Department of Neuroscience, Johns Hopkins University, Baltimore, Maryland, USA; ,
| | - Charles E Connor
- Krieger Mind/Brain Institute and Department of Neuroscience, Johns Hopkins University, Baltimore, Maryland, USA; ,
| |
Collapse
|
4
|
Zhu Z, Kim B, Doudlah R, Chang TY, Rosenberg A. Differential clustering of visual and choice- and saccade-related activity in macaque V3A and CIP. J Neurophysiol 2024; 131:709-722. [PMID: 38478896 PMCID: PMC11305645 DOI: 10.1152/jn.00285.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2023] [Revised: 03/01/2024] [Accepted: 03/04/2024] [Indexed: 04/11/2024] Open
Abstract
Neurons in sensory and motor cortices tend to aggregate in clusters with similar functional properties. Within the primate dorsal ("where") pathway, an important interface between three-dimensional (3-D) visual processing and motor-related functions consists of two hierarchically organized areas: V3A and the caudal intraparietal (CIP) area. In these areas, 3-D visual information, choice-related activity, and saccade-related activity converge, often at the single-neuron level. Characterizing the clustering of functional properties in areas with mixed selectivity, such as these, may help reveal organizational principles that support sensorimotor transformations. Here we quantified the clustering of visual feature selectivity, choice-related activity, and saccade-related activity by performing correlational and parametric comparisons of the responses of well-isolated, simultaneously recorded neurons in macaque monkeys. Each functional domain showed statistically significant clustering in both areas. However, there were also domain-specific differences in the strength of clustering across the areas. Visual feature selectivity and saccade-related activity were more strongly clustered in V3A than in CIP. In contrast, choice-related activity was more strongly clustered in CIP than in V3A. These differences in clustering may reflect the areas' roles in sensorimotor processing. Stronger clustering of visual and saccade-related activity in V3A may reflect a greater role in within-domain processing, as opposed to cross-domain synthesis. In contrast, stronger clustering of choice-related activity in CIP may reflect a greater role in synthesizing information across functional domains to bridge perception and action.NEW & NOTEWORTHY The occipital and parietal cortices of macaque monkeys are bridged by hierarchically organized areas V3A and CIP. These areas support 3-D visual transformations, carry choice-related activity during 3-D perceptual tasks, and possess saccade-related activity. This study quantifies the functional clustering of neuronal response properties within V3A and CIP for each of these domains. The findings reveal domain-specific cross-area differences in clustering that may reflect the areas' roles in sensorimotor processing.
Collapse
Affiliation(s)
- Zikang Zhu
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin-Madison, Madison, Wisconsin, United States
| | - Byounghoon Kim
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin-Madison, Madison, Wisconsin, United States
| | - Raymond Doudlah
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin-Madison, Madison, Wisconsin, United States
| | - Ting-Yu Chang
- School of Medicine, National Defense Medical Center, Taipei, Taiwan
| | - Ari Rosenberg
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin-Madison, Madison, Wisconsin, United States
| |
Collapse
|
5
|
Thompson LW, Kim B, Rokers B, Rosenberg A. Hierarchical computation of 3D motion across macaque areas MT and FST. Cell Rep 2023; 42:113524. [PMID: 38064337 PMCID: PMC10791528 DOI: 10.1016/j.celrep.2023.113524] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/23/2023] [Revised: 10/25/2023] [Accepted: 11/15/2023] [Indexed: 12/30/2023] Open
Abstract
Computing behaviorally relevant representations of three-dimensional (3D) motion from two-dimensional (2D) retinal signals is critical for survival. To ascertain where and how the primate visual system performs this computation, we recorded from the macaque middle temporal (MT) area and its downstream target, the fundus of the superior temporal sulcus (area FST). Area MT is a key site of 2D motion processing, but its role in 3D motion processing is controversial. The functions of FST remain highly underexplored. To distinguish representations of 3D motion from those of 2D retinal motion, we contrast responses to multiple motion cues during a motion discrimination task. The results reveal a hierarchical transformation whereby many FST but not MT neurons are selective for 3D motion. Modeling results further show how generalized, cue-invariant representations of 3D motion in FST may be created by selectively integrating the output of 2D motion selective MT neurons.
Collapse
Affiliation(s)
- Lowell W Thompson
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin - Madison, Madison, WI 53705, USA
| | - Byounghoon Kim
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin - Madison, Madison, WI 53705, USA
| | - Bas Rokers
- Department of Psychology, New York University Abu Dhabi, Abu Dhabi, United Arab Emirates
| | - Ari Rosenberg
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin - Madison, Madison, WI 53705, USA.
| |
Collapse
|
6
|
Rosenberg A, Thompson LW, Doudlah R, Chang TY. Neuronal Representations Supporting Three-Dimensional Vision in Nonhuman Primates. Annu Rev Vis Sci 2023; 9:337-359. [PMID: 36944312 DOI: 10.1146/annurev-vision-111022-123857] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/23/2023]
Abstract
The visual system must reconstruct the dynamic, three-dimensional (3D) world from ambiguous two-dimensional (2D) retinal images. In this review, we synthesize current literature on how the visual system of nonhuman primates performs this transformation through multiple channels within the classically defined dorsal (where) and ventral (what) pathways. Each of these channels is specialized for processing different 3D features (e.g., the shape, orientation, or motion of objects, or the larger scene structure). Despite the common goal of 3D reconstruction, neurocomputational differences between the channels impose distinct information-limiting constraints on perception. Convergent evidence further points to the little-studied area V3A as a potential branchpoint from which multiple 3D-fugal processing channels diverge. We speculate that the expansion of V3A in humans may have supported the emergence of advanced 3D spatial reasoning skills. Lastly, we discuss future directions for exploring 3D information transmission across brain areas and experimental approaches that can further advance the understanding of 3D vision.
Collapse
Affiliation(s)
- Ari Rosenberg
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin-Madison, Madison, Wisconsin, USA;
| | - Lowell W Thompson
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin-Madison, Madison, Wisconsin, USA;
| | - Raymond Doudlah
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin-Madison, Madison, Wisconsin, USA;
| | - Ting-Yu Chang
- School of Medicine, National Defense Medical Center, Taipei, Taiwan
| |
Collapse
|
7
|
Burge J, Burge T. Shape, perspective, and what is and is not perceived: Comment on Morales, Bax, and Firestone (2020). Psychol Rev 2023; 130:1125-1136. [PMID: 35549319 PMCID: PMC11366222 DOI: 10.1037/rev0000363] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Psychology and philosophy have long reflected on the role of perspective in vision. Since the dawn of modern vision science-roughly, since Helmholtz in the late 1800s-scientific explanations in vision have focused on understanding the computations that transform the sensed retinal image into percepts of the three-dimensional environment. The standard view in the science is that distal properties-viewpoint-independent properties of the environment (object shape) and viewpoint-dependent relational properties (3D orientation relative to the viewer)-are perceptually represented and that properties of the proximal stimulus (in vision, the retinal image) are not. This view is woven into the nature of scientific explanation in perceptual psychology, and has guided impressive advances over the past 150 years. A recently published article suggests that in shape perception, the standard view must be revised. It argues, on the basis of new empirical data, that a new entity-perspectival shape-should be introduced into scientific explanations of shape perception. Specifically, the article's centrally advertised claim is that, in addition to distal shape, perspectival shape is perceived. We argue that this claim rests on a series of mistakes. Problems in experimental design entail that the article provides no empirical support for any claims regarding either perspective or the perception of shape. There are further problems in scientific reasoning and conceptual development. Detailing these criticisms and explaining how science treats these issues are meant to clarify method and theory, and to improve exchanges between the science and philosophy of perception. (PsycInfo Database Record (c) 2023 APA, all rights reserved).
Collapse
Affiliation(s)
- Johannes Burge
- Department of Psychology, University of Pennsylvania
- Neuroscience Graduate Group, University of Pennsylvania
- Bioengineering Graduate Group, University of Pennsylvania
| | - Tyler Burge
- Department of Philosophy, University of California, Los Angeles
| |
Collapse
|
8
|
Gao W, Lin Y, Shen J, Han J, Song X, Lu Y, Zhan H, Li Q, Ge H, Lin Z, Shi W, Drugowitsch J, Tang H, Chen X. Diverse effects of gaze direction on heading perception in humans. Cereb Cortex 2023:7024719. [PMID: 36734278 DOI: 10.1093/cercor/bhac541] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/07/2022] [Revised: 12/24/2022] [Accepted: 12/27/2022] [Indexed: 02/04/2023] Open
Abstract
Gaze change can misalign spatial reference frames encoding visual and vestibular signals in cortex, which may affect the heading discrimination. Here, by systematically manipulating the eye-in-head and head-on-body positions to change the gaze direction of subjects, the performance of heading discrimination was tested with visual, vestibular, and combined stimuli in a reaction-time task in which the reaction time is under the control of subjects. We found the gaze change induced substantial biases in perceived heading, increased the threshold of discrimination and reaction time of subjects in all stimulus conditions. For the visual stimulus, the gaze effects were induced by changing the eye-in-world position, and the perceived heading was biased in the opposite direction of gaze. In contrast, the vestibular gaze effects were induced by changing the eye-in-head position, and the perceived heading was biased in the same direction of gaze. Although the bias was reduced when the visual and vestibular stimuli were combined, integration of the 2 signals substantially deviated from predictions of an extended diffusion model that accumulates evidence optimally over time and across sensory modalities. These findings reveal diverse gaze effects on the heading discrimination and emphasize that the transformation of spatial reference frames may underlie the effects.
Collapse
Affiliation(s)
- Wei Gao
- Department of Neurology and Psychiatry of the Second Affiliated Hospital, College of Biomedical Engineering and Instrument Science, Interdisciplinary Institute of Neuroscience and Technology, School of Medicine, Zhejiang University, 268 Kaixuan Road, Jianggan District, Hangzhou 310029, China
| | - Yipeng Lin
- Department of Neurology and Psychiatry of the Second Affiliated Hospital, College of Biomedical Engineering and Instrument Science, Interdisciplinary Institute of Neuroscience and Technology, School of Medicine, Zhejiang University, 268 Kaixuan Road, Jianggan District, Hangzhou 310029, China
| | - Jiangrong Shen
- College of Computer Science and Technology, Zhejiang University, 38 Zheda Road, Xihu District, Hangzhou 310027, China
| | - Jianing Han
- College of Computer Science and Technology, Zhejiang University, 38 Zheda Road, Xihu District, Hangzhou 310027, China
| | - Xiaoxiao Song
- Department of Liberal Arts, School of Art Administration and Education, China Academy of Art, 218 Nanshan Road, Shangcheng District, Hangzhou 310002, China
| | - Yukun Lu
- Department of Neurology and Psychiatry of the Second Affiliated Hospital, College of Biomedical Engineering and Instrument Science, Interdisciplinary Institute of Neuroscience and Technology, School of Medicine, Zhejiang University, 268 Kaixuan Road, Jianggan District, Hangzhou 310029, China
| | - Huijia Zhan
- Department of Neurology and Psychiatry of the Second Affiliated Hospital, College of Biomedical Engineering and Instrument Science, Interdisciplinary Institute of Neuroscience and Technology, School of Medicine, Zhejiang University, 268 Kaixuan Road, Jianggan District, Hangzhou 310029, China
| | - Qianbing Li
- Department of Neurology and Psychiatry of the Second Affiliated Hospital, College of Biomedical Engineering and Instrument Science, Interdisciplinary Institute of Neuroscience and Technology, School of Medicine, Zhejiang University, 268 Kaixuan Road, Jianggan District, Hangzhou 310029, China
| | - Haoting Ge
- Department of Neurology and Psychiatry of the Second Affiliated Hospital, College of Biomedical Engineering and Instrument Science, Interdisciplinary Institute of Neuroscience and Technology, School of Medicine, Zhejiang University, 268 Kaixuan Road, Jianggan District, Hangzhou 310029, China
| | - Zheng Lin
- Department of Psychiatry, Second Affiliated Hospital, School of Medicine, Zhejiang University, 88 Jiefang Road, Shangcheng District, Hangzhou 310009, China
| | - Wenlei Shi
- Center for the Study of the History of Chinese Language and Center for the Study of Language and Cognition, Zhejiang University, 866 Yuhangtang Road, Xihu District, Hangzhou 310058, China
| | - Jan Drugowitsch
- Department of Neurobiology, Harvard Medical School, Longwood Avenue 220, Boston, MA 02116, United States
| | - Huajin Tang
- College of Computer Science and Technology, Zhejiang University, 38 Zheda Road, Xihu District, Hangzhou 310027, China
| | - Xiaodong Chen
- Department of Neurology and Psychiatry of the Second Affiliated Hospital, College of Biomedical Engineering and Instrument Science, Interdisciplinary Institute of Neuroscience and Technology, School of Medicine, Zhejiang University, 268 Kaixuan Road, Jianggan District, Hangzhou 310029, China
| |
Collapse
|
9
|
Kalou K, Sedda G, Gibaldi A, Sabatini SP. Learning bio-inspired head-centric representations of 3D shapes in an active fixation setting. Front Robot AI 2022; 9:994284. [PMID: 36329691 PMCID: PMC9623882 DOI: 10.3389/frobt.2022.994284] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/14/2022] [Accepted: 09/13/2022] [Indexed: 12/03/2022] Open
Abstract
When exploring the surrounding environment with the eyes, humans and primates need to interpret three-dimensional (3D) shapes in a fast and invariant way, exploiting a highly variant and gaze-dependent visual information. Since they have front-facing eyes, binocular disparity is a prominent cue for depth perception. Specifically, it serves as computational substrate for two ground mechanisms of binocular active vision: stereopsis and binocular coordination. To this aim, disparity information, which is expressed in a retinotopic reference frame, is combined along the visual cortical pathways with gaze information and transformed in a head-centric reference frame. Despite the importance of this mechanism, the underlying neural substrates still remain widely unknown. In this work, we investigate the capabilities of the human visual system to interpret the 3D scene exploiting disparity and gaze information. In a psychophysical experiment, human subjects were asked to judge the depth orientation of a planar surface either while fixating a target point or while freely exploring the surface. Moreover, we used the same stimuli to train a recurrent neural network to exploit the responses of a modelled population of cortical (V1) cells to interpret the 3D scene layout. The results for both human performance and from the model network show that integrating disparity information across gaze directions is crucial for a reliable and invariant interpretation of the 3D geometry of the scene.
Collapse
Affiliation(s)
- Katerina Kalou
- Department of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genoa, Italy
| | - Giulia Sedda
- Department of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genoa, Italy
| | - Agostino Gibaldi
- University of California Berkeley, School of Optometry, Berkeley, CA, United States
| | - Silvio P. Sabatini
- Department of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genoa, Italy
| |
Collapse
|
10
|
Doudlah R, Chang TY, Thompson LW, Kim B, Sunkara A, Rosenberg A. Parallel processing, hierarchical transformations, and sensorimotor associations along the 'where' pathway. eLife 2022; 11:78712. [PMID: 35950921 PMCID: PMC9439678 DOI: 10.7554/elife.78712] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2022] [Accepted: 08/10/2022] [Indexed: 11/13/2022] Open
Abstract
Visually guided behaviors require the brain to transform ambiguous retinal images into object-level spatial representations and implement sensorimotor transformations. These processes are supported by the dorsal ‘where’ pathway. However, the specific functional contributions of areas along this pathway remain elusive due in part to methodological differences across studies. We previously showed that macaque caudal intraparietal (CIP) area neurons possess robust 3D visual representations, carry choice- and saccade-related activity, and exhibit experience-dependent sensorimotor associations (Chang et al., 2020b). Here, we used a common experimental design to reveal parallel processing, hierarchical transformations, and the formation of sensorimotor associations along the ‘where’ pathway by extending the investigation to V3A, a major feedforward input to CIP. Higher-level 3D representations and choice-related activity were more prevalent in CIP than V3A. Both areas contained saccade-related activity that predicted the direction/timing of eye movements. Intriguingly, the time course of saccade-related activity in CIP aligned with the temporally integrated V3A output. Sensorimotor associations between 3D orientation and saccade direction preferences were stronger in CIP than V3A, and moderated by choice signals in both areas. Together, the results explicate parallel representations, hierarchical transformations, and functional associations of visual and saccade-related signals at a key juncture in the ‘where’ pathway.
Collapse
Affiliation(s)
- Raymond Doudlah
- Department of Neuroscience, University of Wisconsin-Madison, Madison, United States
| | - Ting-Yu Chang
- Department of Neuroscience, University of Wisconsin-Madison, Madison, United States
| | - Lowell W Thompson
- Department of Neuroscience, University of Wisconsin-Madison, Madison, United States
| | - Byounghoon Kim
- Department of Neuroscience, University of Wisconsin-Madison, Madison, United States
| | | | - Ari Rosenberg
- Department of Neuroscience, University of Wisconsin-Madison, Madison, United States
| |
Collapse
|
11
|
Thompson LW, Kim B, Zhu Z, Rokers B, Rosenberg A. Perspective Cues Make Eye-specific Contributions to 3-D Motion Perception. J Cogn Neurosci 2021; 34:192-208. [PMID: 34813655 PMCID: PMC8692976 DOI: 10.1162/jocn_a_01781] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/02/2023]
Abstract
Robust 3-D visual perception is achieved by integrating stereoscopic and perspective cues. The canonical model describing the integration of these cues assumes that perspective signals sensed by the left and right eyes are indiscriminately pooled into a single representation that contributes to perception. Here, we show that this model fails to account for 3-D motion perception. We measured the sensitivity of male macaque monkeys to 3-D motion signaled by left-eye perspective cues, right-eye perspective cues, stereoscopic cues, and all three cues combined. The monkeys exhibited idiosyncratic differences in their biases and sensitivities for each cue, including left- and right-eye perspective cues, suggesting that the signals undergo at least partially separate neural processing. Importantly, sensitivity to combined cue stimuli was greater than predicted by the canonical model, which previous studies found to account for the perception of 3-D orientation in both humans and monkeys. Instead, 3-D motion sensitivity was best explained by a model in which stereoscopic cues were integrated with left- and right-eye perspective cues whose representations were at least partially independent. These results indicate that the integration of perspective and stereoscopic cues is a shared computational strategy across 3-D processing domains. However, they also reveal a fundamental difference in how left- and right-eye perspective signals are represented for 3-D orientation versus motion perception. This difference results in more effective use of available sensory information in the processing of 3-D motion than orientation and may reflect the temporal urgency of avoiding and intercepting moving objects.
Collapse
|
12
|
Baeg E, Doudlah R, Swader R, Lee H, Han M, Kim SG, Rosenberg A, Kim B. MRI Compatible, Customizable, and 3D-Printable Microdrive for Neuroscience Research. eNeuro 2021; 8:ENEURO.0495-20.2021. [PMID: 33593730 PMCID: PMC7986532 DOI: 10.1523/eneuro.0495-20.2021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2020] [Revised: 01/26/2021] [Accepted: 01/30/2021] [Indexed: 02/02/2023] Open
Abstract
The effective connectivity of brain networks can be assessed using functional magnetic resonance imaging (fMRI) to quantify the effects of local electrical microstimulation (EM) on distributed neuronal activity. The delivery of EM to specific brain regions, particularly with layer specificity, requires MRI compatible equipment that provides fine control of a stimulating electrode's position within the brain while minimizing imaging artifacts. To this end, we developed a microdrive made entirely of MRI compatible materials. The microdrive uses an integrated penetration grid to guide electrodes and relies on a microdrilling technique to eliminate the need for large craniotomies, further reducing implant maintenance and image distortions. The penetration grid additionally serves as a built-in MRI marker, providing a visible fiducial reference for estimating probe trajectories. Following the initial implant procedure, these features allow for multiple electrodes to be inserted, removed, and repositioned with minimal effort, using a screw-type actuator. To validate the design of the microdrive, we conducted an EM-coupled fMRI study with a male macaque monkey. The results verified that the microdrive can be used to deliver EM during MRI procedures with minimal imaging artifacts, even within a 7 Tesla (7T) environment. Future applications of the microdrive include neuronal recordings and targeted drug delivery. We provide computer aided design (CAD) templates and a parts list for modifying and fabricating the microdrive for specific research needs. These designs provide a convenient, cost-effective approach to fabricating MRI compatible microdrives for neuroscience research.
Collapse
Affiliation(s)
- Eunha Baeg
- Center for Neuroscience Imaging Research, Institute for Basic Science, Suwon, Republic of Korea 16060
- Department of Biomedical Engineering, Sungkyunkwan University, Suwon, Republic of Korea 16419
| | - Raymond Doudlah
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin-Madison, Madison, WI 53705
| | | | - Hyowon Lee
- System Design Engineering, University of Waterloo, Waterloo, Ontario, Canada N2L 3G1
| | - Minjun Han
- Department of Biomedical Engineering, Sungkyunkwan University, Suwon, Republic of Korea 16419
| | - Seong-Gi Kim
- Center for Neuroscience Imaging Research, Institute for Basic Science, Suwon, Republic of Korea 16060
- Department of Biomedical Engineering, Sungkyunkwan University, Suwon, Republic of Korea 16419
| | - Ari Rosenberg
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin-Madison, Madison, WI 53705
| | - Byounghoon Kim
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin-Madison, Madison, WI 53705
| |
Collapse
|
13
|
Chauhan T, Héjja-Brichard Y, Cottereau BR. Modelling binocular disparity processing from statistics in natural scenes. Vision Res 2020; 176:27-39. [DOI: 10.1016/j.visres.2020.07.009] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2019] [Revised: 07/19/2020] [Accepted: 07/20/2020] [Indexed: 11/25/2022]
|
14
|
Rima S, Cottereau BR, Héjja-Brichard Y, Trotter Y, Durand JB. Wide-field retinotopy reveals a new visuotopic cluster in macaque posterior parietal cortex. Brain Struct Funct 2020; 225:2447-2461. [PMID: 32875354 PMCID: PMC7544618 DOI: 10.1007/s00429-020-02134-2] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2019] [Accepted: 08/22/2020] [Indexed: 12/31/2022]
Abstract
We investigated the visuotopic organization of macaque posterior parietal cortex (PPC) by combining functional imaging (fMRI) and wide-field retinotopic mapping in two macaque monkeys. Whole brain blood-oxygen-level-dependent (BOLD) signal was recorded while monkeys maintained central fixation during the presentation of large rotating wedges and expending/contracting annulus of a "shaking" fruit basket, designed to maximize the recruitment of PPC neurons. Results of the surface-based population receptive field (pRF) analysis reveal a new cluster of four visuotopic areas at the confluence of the parieto-occipital and intra-parietal sulci, in a location previously defined histologically and anatomically as the posterior intra-parietal (PIP) region. This PIP cluster groups together two recently described areas (CIP1/2) laterally and two newly identified ones (PIP1/2) medially, whose foveal representations merge in the fundus of the intra-parietal sulcus. The cluster shares borders with other visuotopic areas: V3d posteriorly, V3A/DP laterally, V6/V6A medially and LIP anteriorly. Together, these results show that monkey PPC is endowed with a dense set of visuotopic areas, as its human counterpart. The fact that fMRI and wide-field stimulation allows a functional parsing of monkey PPC offers a new framework for studying functional homologies with human PPC.
Collapse
Affiliation(s)
- Samy Rima
- Centre de Recherche Cerveau et Cognition, Université de Toulouse, Toulouse, France.
- Centre National de la Recherche Scientifique, Toulouse Cedex, France.
| | - Benoit R Cottereau
- Centre de Recherche Cerveau et Cognition, Université de Toulouse, Toulouse, France
- Centre National de la Recherche Scientifique, Toulouse Cedex, France
| | - Yseut Héjja-Brichard
- Centre de Recherche Cerveau et Cognition, Université de Toulouse, Toulouse, France
- Centre National de la Recherche Scientifique, Toulouse Cedex, France
| | - Yves Trotter
- Centre de Recherche Cerveau et Cognition, Université de Toulouse, Toulouse, France
- Centre National de la Recherche Scientifique, Toulouse Cedex, France
| | - Jean-Baptiste Durand
- Centre de Recherche Cerveau et Cognition, Université de Toulouse, Toulouse, France.
- Centre National de la Recherche Scientifique, Toulouse Cedex, France.
| |
Collapse
|
15
|
Effective Connectivity Reveals an Interconnected Inferotemporal Network for Three-Dimensional Structure Processing. J Neurosci 2020; 40:8501-8512. [PMID: 33028641 DOI: 10.1523/jneurosci.3024-19.2020] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2019] [Revised: 09/15/2020] [Accepted: 09/22/2020] [Indexed: 11/21/2022] Open
Abstract
Disparity-defined 3D shape is processed in both the ventral and the dorsal visual stream. The network of cortical areas that is activated during the processing of disparity-defined 3D shape includes, in addition to parietal and premotor areas, three clearly distinct regions in inferotemporal cortex (ITC). To investigate the connectivity of the latter regions, we combined electrical stimulation with fMRI in male macaque monkeys. Electrical stimulation of each of the 3D-structure nodes in ITC mainly elicited increased fMRI activations in the other 3D-structure nodes and more variably in other parts of ventral visual cortex. Importantly, no increased activation was found in parietal areas, nor in PFC, whereas microstimulation in posterior parietal cortex did activate the ITC. Our results indicate that 3D-structure nodes in ITC form a strongly interconnected network, receiving input from parietal areas implicated in 3D-structure processing.SIGNIFICANCE STATEMENT Previous studies combining electrical microstimulation with functional imaging showed an interconnected set of regions in the ventral stream processing faces or bodies, but is has been unclear whether the same is true for other visual categories. Here the authors show that there is a connected system of stereo-selective regions in inferotemporal cortex, receiving input from parietal areas in the dorsal stream.
Collapse
|
16
|
Kim S, Burge J. Natural scene statistics predict how humans pool information across space in surface tilt estimation. PLoS Comput Biol 2020; 16:e1007947. [PMID: 32579559 PMCID: PMC7340327 DOI: 10.1371/journal.pcbi.1007947] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2019] [Revised: 07/07/2020] [Accepted: 05/14/2020] [Indexed: 11/30/2022] Open
Abstract
Visual systems estimate the three-dimensional (3D) structure of scenes from information in two-dimensional (2D) retinal images. Visual systems use multiple sources of information to improve the accuracy of these estimates, including statistical knowledge of the probable spatial arrangements of natural scenes. Here, we examine how 3D surface tilts are spatially related in real-world scenes, and show that humans pool information across space when estimating surface tilt in accordance with these spatial relationships. We develop a hierarchical model of surface tilt estimation that is grounded in the statistics of tilt in natural scenes and images. The model computes a global tilt estimate by pooling local tilt estimates within an adaptive spatial neighborhood. The spatial neighborhood in which local estimates are pooled changes according to the value of the local estimate at a target location. The hierarchical model provides more accurate estimates of groundtruth tilt in natural scenes and provides a better account of human performance than the local estimates. Taken together, the results imply that the human visual system pools information about surface tilt across space in accordance with natural scene statistics. Visual systems estimate three-dimensional (3D) properties of scenes from two-dimensional images on the retinas. To solve this difficult problem as accurately as possible, visual systems use many available sources of information, including information about how the 3D properties of the world are spatially arranged. This manuscript reports a systematic analysis of 3D surface tilt in natural scenes, a model of surface tilt estimation that makes use of these scene statistics, and human psychophysical data on the estimation of surface tilt from natural images. The results show that the regularities present in the natural environment predict both how to maximize the accuracy of tilt estimation and how to maximize the prediction of human performance. This work contributes to a growing line of work that establishes links between rigorous measurements of natural scenes and the function of sensory and perceptual systems.
Collapse
Affiliation(s)
- Seha Kim
- Department of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America
- * E-mail:
| | - Johannes Burge
- Department of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America
- Neuroscience Graduate Group, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America
- Bioengineering Graduate Group, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America
| |
Collapse
|
17
|
Time Course of Sensory Substitution for Gravity Sensing in Visual Vertical Orientation Perception following Complete Vestibular Loss. eNeuro 2020; 7:ENEURO.0021-20.2020. [PMID: 32561572 PMCID: PMC7358335 DOI: 10.1523/eneuro.0021-20.2020] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2020] [Revised: 03/24/2020] [Accepted: 03/30/2020] [Indexed: 01/06/2023] Open
Abstract
Loss of vestibular function causes severe acute symptoms of dizziness and disorientation, yet the brain can adapt and regain near to normal locomotor and orientation function through sensory substitution. Animal studies quantifying functional recovery have yet been limited to reflexive eye movements. Here, we studied the interplay between vestibular and proprioceptive graviception in macaque monkeys trained in an earth-vertical visual orientation (subjective visual vertical; SVV) task and measured the time course of sensory substitution for gravity perception following complete bilateral vestibular loss (BVL). Graviceptive gain, defined as the ratio of perceived versus actual tilt angle, decreased to 20% immediately following labyrinthectomy, and recovered to nearly prelesion levels with a time constant of approximately three weeks of postsurgery testing. We conclude that proprioception accounts for up to 20% of gravity sensing in normal animals, and is re-weighted to substitute completely perceptual graviception after vestibular loss. We show that these results can be accounted for by an optimal sensory fusion model.
Collapse
|
18
|
Lanzilotto M, Ferroni CG, Livi A, Gerbella M, Maranesi M, Borra E, Passarelli L, Gamberini M, Fogassi L, Bonini L, Orban GA. Anterior Intraparietal Area: A Hub in the Observed Manipulative Action Network. Cereb Cortex 2020; 29:1816-1833. [PMID: 30766996 PMCID: PMC6418391 DOI: 10.1093/cercor/bhz011] [Citation(s) in RCA: 46] [Impact Index Per Article: 9.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2018] [Revised: 01/07/2019] [Accepted: 01/18/2019] [Indexed: 11/13/2022] Open
Abstract
Current knowledge regarding the processing of observed manipulative actions (OMAs) (e.g., grasping, dragging, or dropping) is limited to grasping and underlying neural circuitry remains controversial. Here, we addressed these issues by combining chronic neuronal recordings along the anteroposterior extent of monkeys’ anterior intraparietal (AIP) area with tracer injections into the recorded sites. We found robust neural selectivity for 7 distinct OMAs, particularly in the posterior part of AIP (pAIP), where it was associated with motor coding of grip type and own-hand visual feedback. This cluster of functional properties appears to be specifically grounded in stronger direct connections of pAIP with the temporal regions of the ventral visual stream and the prefrontal cortex, as connections with skeletomotor related areas and regions of the dorsal visual stream exhibited opposite or no rostrocaudal gradients. Temporal and prefrontal areas may provide visual and contextual information relevant for manipulative action processing. These results revise existing models of the action observation network, suggesting that pAIP constitutes a parietal hub for routing information about OMA identity to the other nodes of the network.
Collapse
Affiliation(s)
- Marco Lanzilotto
- Department of Medicine and Surgery, University of Parma, Via Volturno 39, Parma, Italy
| | | | - Alessandro Livi
- Department of Medicine and Surgery, University of Parma, Via Volturno 39, Parma, Italy
| | - Marzio Gerbella
- Department of Medicine and Surgery, University of Parma, Via Volturno 39, Parma, Italy
| | - Monica Maranesi
- Department of Medicine and Surgery, University of Parma, Via Volturno 39, Parma, Italy
| | - Elena Borra
- Department of Medicine and Surgery, University of Parma, Via Volturno 39, Parma, Italy
| | - Lauretta Passarelli
- Department of Biomedical and Neuromotor Sciences, University of Bologna, Piazza di Porta San Donato 2, Bologna, Italy
| | - Michela Gamberini
- Department of Biomedical and Neuromotor Sciences, University of Bologna, Piazza di Porta San Donato 2, Bologna, Italy
| | - Leonardo Fogassi
- Department of Medicine and Surgery, University of Parma, Via Volturno 39, Parma, Italy
| | - Luca Bonini
- Department of Medicine and Surgery, University of Parma, Via Volturno 39, Parma, Italy
| | - Guy A Orban
- Department of Medicine and Surgery, University of Parma, Via Volturno 39, Parma, Italy
| |
Collapse
|
19
|
Héjja-Brichard Y, Rima S, Rapha E, Durand JB, Cottereau BR. Stereomotion Processing in the Nonhuman Primate Brain. Cereb Cortex 2020; 30:4528-4543. [PMID: 32227117 DOI: 10.1093/cercor/bhaa055] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2019] [Revised: 01/22/2020] [Accepted: 02/14/2020] [Indexed: 12/21/2022] Open
Abstract
The cortical areas that process disparity-defined motion-in-depth (i.e., cyclopean stereomotion [CSM]) were characterized with functional magnetic resonance imaging (fMRI) in two awake, behaving macaques. The experimental protocol was similar to previous human neuroimaging studies. We contrasted the responses to dynamic random-dot patterns that continuously changed their binocular disparity over time with those to a control condition that shared the same properties, except that the temporal frames were shuffled. A whole-brain voxel-wise analysis revealed that in all four cortical hemispheres, three areas showed consistent sensitivity to CSM. Two of them were localized respectively in the lower bank of the superior temporal sulcus (CSMSTS) and on the neighboring infero-temporal gyrus (CSMITG). The third area was situated in the posterior parietal cortex (CSMPPC). Additional regions of interest-based analyses within retinotopic areas defined in both animals indicated weaker but significant responses to CSM within the MT cluster (most notably in areas MSTv and FST). Altogether, our results are in agreement with previous findings in both human and macaque and suggest that the cortical areas that process CSM are relatively well preserved between the two primate species.
Collapse
Affiliation(s)
- Yseult Héjja-Brichard
- Centre de Recherche Cerveau et Cognition, Université de Toulouse, 31052 Toulouse, France.,Centre National de la Recherche Scientifique, 31055 Toulouse, France
| | - Samy Rima
- Centre de Recherche Cerveau et Cognition, Université de Toulouse, 31052 Toulouse, France.,Centre National de la Recherche Scientifique, 31055 Toulouse, France
| | - Emilie Rapha
- Centre de Recherche Cerveau et Cognition, Université de Toulouse, 31052 Toulouse, France.,Centre National de la Recherche Scientifique, 31055 Toulouse, France
| | - Jean-Baptiste Durand
- Centre de Recherche Cerveau et Cognition, Université de Toulouse, 31052 Toulouse, France.,Centre National de la Recherche Scientifique, 31055 Toulouse, France
| | - Benoit R Cottereau
- Centre de Recherche Cerveau et Cognition, Université de Toulouse, 31052 Toulouse, France.,Centre National de la Recherche Scientifique, 31055 Toulouse, France
| |
Collapse
|
20
|
Computational Mechanisms for Perceptual Stability using Disparity and Motion Parallax. J Neurosci 2020; 40:996-1014. [PMID: 31699889 DOI: 10.1523/jneurosci.0036-19.2019] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2019] [Revised: 09/24/2019] [Accepted: 10/07/2019] [Indexed: 11/21/2022] Open
Abstract
Walking and other forms of self-motion create global motion patterns across our eyes. With the resulting stream of visual signals, how do we perceive ourselves as moving through a stable world? Although the neural mechanisms are largely unknown, human studies (Warren and Rushton, 2009) provide strong evidence that the visual system is capable of parsing the global motion into two components: one due to self-motion and the other due to independently moving objects. In the present study, we use computational modeling to investigate potential neural mechanisms for stabilizing visual perception during self-motion that build on neurophysiology of the middle temporal (MT) and medial superior temporal (MST) areas. One such mechanism leverages direction, speed, and disparity tuning of cells in dorsal MST (MSTd) to estimate the combined motion parallax and disparity signals attributed to the observer's self-motion. Feedback from the most active MSTd cell subpopulations suppresses motion signals in MT that locally match the preference of the MSTd cell in both parallax and disparity. This mechanism combined with local surround inhibition in MT allows the model to estimate self-motion while maintaining a sparse motion representation that is compatible with perceptual stability. A key consequence is that after signals compatible with the observer's self-motion are suppressed, the direction of independently moving objects is represented in a world-relative rather than observer-relative reference frame. Our analysis explicates how temporal dynamics and joint motion parallax-disparity tuning resolve the world-relative motion of moving objects and establish perceptual stability. Together, these mechanisms capture findings on the perception of object motion during self-motion.SIGNIFICANCE STATEMENT The image integrated by our eyes as we move through our environment undergoes constant flux as trees, buildings, and other surroundings stream by us. If our view can change so radically from one moment to the next, how do we perceive a stable world? Although progress has been made in understanding how this works, little is known about the underlying brain mechanisms. We propose a computational solution whereby multiple brain areas communicate to suppress the motion attributed to our movement relative to the stationary world, which is often responsible for a large proportion of the flux across the visual field. We simulated the proposed neural mechanisms and tested model estimates using data from human perceptual studies.
Collapse
|
21
|
Optimized but Not Maximized Cue Integration for 3D Visual Perception. eNeuro 2020; 7:ENEURO.0411-19.2019. [PMID: 31836597 PMCID: PMC6948924 DOI: 10.1523/eneuro.0411-19.2019] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2019] [Revised: 12/05/2019] [Accepted: 12/08/2019] [Indexed: 02/02/2023] Open
Abstract
Reconstructing three-dimensional (3D) scenes from two-dimensional (2D) retinal images is an ill-posed problem. Despite this, 3D perception of the world based on 2D retinal images is seemingly accurate and precise. The integration of distinct visual cues is essential for robust 3D perception in humans, but it is unclear whether this is true for non-human primates (NHPs). Here, we assessed 3D perception in macaque monkeys using a planar surface orientation discrimination task. Perception was accurate across a wide range of spatial poses (orientations and distances), but precision was highly dependent on the plane's pose. The monkeys achieved robust 3D perception by dynamically reweighting the integration of stereoscopic and perspective cues according to their pose-dependent reliabilities. Errors in performance could be explained by a prior resembling the 3D orientation statistics of natural scenes. We used neural network simulations based on 3D orientation-selective neurons recorded from the same monkeys to assess how neural computation might constrain perception. The perceptual data were consistent with a model in which the responses of two independent neuronal populations representing stereoscopic cues and perspective cues (with perspective signals from the two eyes combined using nonlinear canonical computations) were optimally integrated through linear summation. Perception of combined-cue stimuli was optimal given this architecture. However, an alternative architecture in which stereoscopic cues, left eye perspective cues, and right eye perspective cues were represented by three independent populations yielded two times greater precision than the monkeys. This result suggests that, due to canonical computations, cue integration for 3D perception is optimized but not maximized.
Collapse
|
22
|
Chang TY, Doudlah R, Kim B, Sunkara A, Thompson LW, Lowe ME, Rosenberg A. Functional links between sensory representations, choice activity, and sensorimotor associations in parietal cortex. eLife 2020; 9:57968. [PMID: 33078705 PMCID: PMC7641584 DOI: 10.7554/elife.57968] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2020] [Accepted: 10/19/2020] [Indexed: 02/02/2023] Open
Abstract
Three-dimensional (3D) representations of the environment are often critical for selecting actions that achieve desired goals. The success of these goal-directed actions relies on 3D sensorimotor transformations that are experience-dependent. Here we investigated the relationships between the robustness of 3D visual representations, choice-related activity, and motor-related activity in parietal cortex. Macaque monkeys performed an eight-alternative 3D orientation discrimination task and a visually guided saccade task while we recorded from the caudal intraparietal area using laminar probes. We found that neurons with more robust 3D visual representations preferentially carried choice-related activity. Following the onset of choice-related activity, the robustness of the 3D representations further increased for those neurons. We additionally found that 3D orientation and saccade direction preferences aligned, particularly for neurons with choice-related activity, reflecting an experience-dependent sensorimotor association. These findings reveal previously unrecognized links between the fidelity of ecologically relevant object representations, choice-related activity, and motor-related activity.
Collapse
Affiliation(s)
- Ting-Yu Chang
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin–MadisonMadisonUnited States
| | - Raymond Doudlah
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin–MadisonMadisonUnited States
| | - Byounghoon Kim
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin–MadisonMadisonUnited States
| | | | - Lowell W Thompson
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin–MadisonMadisonUnited States
| | - Meghan E Lowe
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin–MadisonMadisonUnited States
| | - Ari Rosenberg
- Department of Neuroscience, School of Medicine and Public Health, University of Wisconsin–MadisonMadisonUnited States
| |
Collapse
|
23
|
Li Z, Shigemasu H. Generalized Representation of Stereoscopic Surface Shape and Orientation in the Human Visual Cortex. Front Hum Neurosci 2019; 13:283. [PMID: 31481886 PMCID: PMC6710440 DOI: 10.3389/fnhum.2019.00283] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2019] [Accepted: 07/31/2019] [Indexed: 11/13/2022] Open
Abstract
The brain's ability to extract three-dimensional (3D) shape and orientation information from viewed objects is vital in daily life. Stereoscopic 3D surface perception relies on binocular disparity. Neurons selective to binocular disparity are widely distributed among visual areas, but the manner in these areas are involved in stereoscopic 3D surface representation is unclear. To address this, participants were instructed to observe random dot stereograms (RDS) depicting convex and concave curved surfaces and the blood oxygenation level-dependent (BOLD) signal of visual cortices was recorded. Two surface types were: (i) horizontally positioned surfaces defined by shear disparity; and (ii) vertically positioned surfaces defined by compression disparity. The surfaces were presented at different depth positions per trial. Functional magnetic resonance imaging (fMRI) data were classified from early visual areas to higher visual areas. We determined whether cortical areas were selective to shape and orientation by assessing same-type stimuli classification accuracies based on multi-voxel activity patterns per area. To identify whether some areas were related to a more generalized sign of curvature or orientation representation, transfer classification was used by training classifiers on one dataset type and testing classifiers on another type. Same-type stimuli classification results showed that most selected visual areas were selective to shape and all were selective to the orientation of disparity-defined 3D surfaces. Transfer classification results showed that in the dorsal visual area V3A, classification accuracies for the discriminate sign of surface curvature were higher than the baseline of statistical significance for all types of classifications, demonstrating that V3A is related to generalized shape representation. Classification accuracies for discriminating horizontal-vertical surfaces in higher dorsal areas V3A and V7 and ventral area lateral occipital complex (LOC) as well as in some areas of intraparietal sulcus (IPS) were higher than the baseline of statistical significance, indicating their relation to the generalized representation of 3D surface orientation.
Collapse
Affiliation(s)
- Zhen Li
- Graduate School of Engineering, Kochi University of Technology, Kochi, Japan
| | | |
Collapse
|
24
|
Uji M, Lingnau A, Cavin I, Vishwanath D. Identifying Cortical Substrates Underlying the Phenomenology of Stereopsis and Realness: A Pilot fMRI Study. Front Neurosci 2019; 13:646. [PMID: 31354404 PMCID: PMC6637755 DOI: 10.3389/fnins.2019.00646] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2019] [Accepted: 06/05/2019] [Indexed: 12/05/2022] Open
Abstract
Viewing a real scene or a stereoscopic image (e.g., 3D movies) with both eyes yields a vivid subjective impression of object solidity, tangibility, immersive negative space and sense of realness; something that is not experienced when viewing single pictures of 3D scenes normally with both eyes. This phenomenology, sometimes referred to as stereopsis, is conventionally ascribed to the derivation of depth from the differences in the two eye's images (binocular disparity). Here we report on a pilot study designed to explore if dissociable neural activity associated with the phenomenology of realness can be localized in the cortex. In order to dissociate subjective impression from disparity processing, we capitalized on the finding that the impression of realness associated with stereoscopic viewing can also be generated when viewing a single picture of a 3D scene with one eye through an aperture. Under a blocked fMRI design, subjects viewed intact and scrambled images of natural 3-D objects, and scenes under three viewing conditions: (1) single pictures viewed normally with both eyes (binocular); (2) single pictures viewed with one eye through an aperture (monocular-aperture); and (3) stereoscopic anaglyph images of the same scenes viewed with both eyes (binocular stereopsis). Fixed-effects GLM contrasts aimed at isolating the phenomenology of stereopsis demonstrated a selective recruitment of similar posterior parietal regions for both monocular and binocular stereopsis conditions. Our findings provide preliminary evidence that the cortical processing underlying the subjective impression of realness may be dissociable and distinct from the derivation of depth from disparity.
Collapse
Affiliation(s)
- Makoto Uji
- School of Psychology and Neuroscience, University of St Andrews, St Andrews, United Kingdom
| | - Angelika Lingnau
- Institute of Psychology, University of Regensburg, Regensburg, Germany
| | - Ian Cavin
- TAyside Medical Science Centre (TASC), NHS Tayside, Dundee, United Kingdom
| | - Dhanraj Vishwanath
- School of Psychology and Neuroscience, University of St Andrews, St Andrews, United Kingdom
| |
Collapse
|
25
|
Uji M, Jentzsch I, Redburn J, Vishwanath D. Dissociating neural activity associated with the subjective phenomenology of monocular stereopsis: An EEG study. Neuropsychologia 2019; 129:357-371. [PMID: 31034841 DOI: 10.1016/j.neuropsychologia.2019.04.017] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/10/2018] [Revised: 03/26/2019] [Accepted: 04/23/2019] [Indexed: 12/15/2022]
Abstract
The subjective phenomenology associated with stereopsis, of solid tangible objects separated by a palpable negative space, is conventionally thought to be a by-product of the derivation of depth from binocular disparity. However, the same qualitative impression has been reported in the absence of disparity, e.g., when viewing pictorial images monocularly through an aperture. Here we aimed to explore if we could identify dissociable neural activity associated with the qualitative impression of stereopsis in the absence of the processing of binocular disparities. We measured EEG activity while subjects viewed pictorial (non-stereoscopic) images of 2D and 3D geometric forms under four different viewing conditions (binocular, monocular, binocular aperture, monocular aperture). EEG activity was analysed by oscillatory source localization (beamformer technique) to examine power change in occipital and parietal regions across viewing and stimulus conditions in targeted frequency bands (alpha: 8-13 Hz & gamma: 60-90 Hz). We observed expected event-related gamma synchronization and alpha desynchronization in occipital cortex and predominant gamma synchronization in parietal cortex across viewing and stimulus conditions. However, only the viewing condition predicted to generate the strongest impression of stereopsis (monocular aperture) revealed significantly elevated gamma synchronization within the parietal cortex for the critical contrasts (3D vs. 2D form). These findings suggest dissociable neural processes specific to the qualitative impression of stereopsis as distinguished from disparity processing.
Collapse
Affiliation(s)
- Makoto Uji
- School of Psychology and Neuroscience, University of St Andrews, UK.
| | - Ines Jentzsch
- School of Psychology and Neuroscience, University of St Andrews, UK
| | - James Redburn
- School of Psychology and Neuroscience, University of St Andrews, UK
| | | |
Collapse
|
26
|
Choice-Related Activity during Visual Slant Discrimination in Macaque CIP But Not V3A. eNeuro 2019; 6:eN-NWR-0248-18. [PMID: 30923736 PMCID: PMC6437654 DOI: 10.1523/eneuro.0248-18.2019] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2018] [Revised: 02/18/2019] [Accepted: 02/26/2019] [Indexed: 02/02/2023] Open
Abstract
Creating three-dimensional (3D) representations of the world from two-dimensional retinal images is fundamental to visually guided behaviors including reaching and grasping. A critical component of this process is determining the 3D orientation of objects. Previous studies have shown that neurons in the caudal intraparietal area (CIP) of the macaque monkey represent 3D planar surface orientation (i.e., slant and tilt). Here we compare the responses of neurons in areas V3A (which is implicated in 3D visual processing and precedes CIP in the visual hierarchy) and CIP to 3D-oriented planar surfaces. We then examine whether activity in these areas correlates with perception during a fine slant discrimination task in which the monkeys report if the top of a surface is slanted toward or away from them. Although we find that V3A and CIP neurons show similar sensitivity to planar surface orientation, significant choice-related activity during the slant discrimination task is rare in V3A but prominent in CIP. These results implicate both V3A and CIP in the representation of 3D surface orientation, and suggest a functional dissociation between the areas based on slant-related choice signals.
Collapse
|
27
|
Kim B, Kenchappa SC, Sunkara A, Chang TY, Thompson L, Doudlah R, Rosenberg A. Real-time experimental control using network-based parallel processing. eLife 2019; 8:e40231. [PMID: 30730290 PMCID: PMC6366897 DOI: 10.7554/elife.40231] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2018] [Accepted: 01/29/2019] [Indexed: 02/02/2023] Open
Abstract
Modern neuroscience research often requires the coordination of multiple processes such as stimulus generation, real-time experimental control, as well as behavioral and neural measurements. The technical demands required to simultaneously manage these processes with high temporal fidelity is a barrier that limits the number of labs performing such work. Here we present an open-source, network-based parallel processing framework that lowers this barrier. The Real-Time Experimental Control with Graphical User Interface (REC-GUI) framework offers multiple advantages: (i) a modular design that is agnostic to coding language(s) and operating system(s) to maximize experimental flexibility and minimize researcher effort, (ii) simple interfacing to connect multiple measurement and recording devices, (iii) high temporal fidelity by dividing task demands across CPUs, and (iv) real-time control using a fully customizable and intuitive GUI. We present applications for human, non-human primate, and rodent studies which collectively demonstrate that the REC-GUI framework facilitates technically demanding, behavior-contingent neuroscience research. Editorial note This article has been through an editorial process in which the authors decide how to respond to the issues raised during peer review. The Reviewing Editor's assessment is that all the issues have been addressed (see decision letter).
Collapse
Affiliation(s)
- Byounghoon Kim
- Department of Neuroscience, School of Medicine and Public HealthUniversity of Wisconsin–MadisonMadisonUnited States
| | | | - Adhira Sunkara
- Department of SurgeryStanford University School of MedicineStanfordUnited States
| | - Ting-Yu Chang
- Department of Neuroscience, School of Medicine and Public HealthUniversity of Wisconsin–MadisonMadisonUnited States
| | - Lowell Thompson
- Department of Neuroscience, School of Medicine and Public HealthUniversity of Wisconsin–MadisonMadisonUnited States
| | - Raymond Doudlah
- Department of Neuroscience, School of Medicine and Public HealthUniversity of Wisconsin–MadisonMadisonUnited States
| | - Ari Rosenberg
- Department of Neuroscience, School of Medicine and Public HealthUniversity of Wisconsin–MadisonMadisonUnited States
| |
Collapse
|
28
|
Rideaux R, Welchman AE. Proscription supports robust perceptual integration by suppression in human visual cortex. Nat Commun 2018; 9:1502. [PMID: 29666361 PMCID: PMC5904115 DOI: 10.1038/s41467-018-03400-y] [Citation(s) in RCA: 24] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2017] [Accepted: 02/07/2018] [Indexed: 11/14/2022] Open
Abstract
Perception relies on integrating information within and between the senses, but how does the brain decide which pieces of information should be integrated and which kept separate? Here we demonstrate how proscription can be used to solve this problem: certain neurons respond best to unrealistic combinations of features to provide ‘what not’ information that drives suppression of unlikely perceptual interpretations. First, we present a model that captures both improved perception when signals are consistent (and thus should be integrated) and robust estimation when signals are conflicting. Second, we test for signatures of proscription in the human brain. We show that concentrations of inhibitory neurotransmitter GABA in a brain region intricately involved in integrating cues (V3B/KO) correlate with robust integration. Finally, we show that perturbing excitation/inhibition impairs integration. These results highlight the role of proscription in robust perception and demonstrate the functional purpose of ‘what not’ sensors in supporting sensory estimation. Perception relies on information integration but it is unclear how the brain decides which information to integrate and which to keep separate. Here, the authors develop and test a biologically inspired model of cue-integration, implicating a key role for GABAergic proscription in robust perception.
Collapse
Affiliation(s)
- Reuben Rideaux
- Department of Psychology, University of Cambridge, Downing Street, Cambridge, CB2 3EB, UK
| | - Andrew E Welchman
- Department of Psychology, University of Cambridge, Downing Street, Cambridge, CB2 3EB, UK.
| |
Collapse
|
29
|
Alizadeh AM, Van Dromme I, Verhoef BE, Janssen P. Caudal Intraparietal Sulcus and three-dimensional vision: A combined functional magnetic resonance imaging and single-cell study. Neuroimage 2018; 166:46-59. [DOI: 10.1016/j.neuroimage.2017.10.045] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2017] [Revised: 09/28/2017] [Accepted: 10/21/2017] [Indexed: 11/30/2022] Open
|
30
|
Kim S, Burge J. The lawful imprecision of human surface tilt estimation in natural scenes. eLife 2018; 7:31448. [PMID: 29384477 PMCID: PMC5844693 DOI: 10.7554/elife.31448] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2017] [Accepted: 01/29/2018] [Indexed: 01/03/2023] Open
Abstract
Estimating local surface orientation (slant and tilt) is fundamental to recovering the three-dimensional structure of the environment. It is unknown how well humans perform this task in natural scenes. Here, with a database of natural stereo-images having groundtruth surface orientation at each pixel, we find dramatic differences in human tilt estimation with natural and artificial stimuli. Estimates are precise and unbiased with artificial stimuli and imprecise and strongly biased with natural stimuli. An image-computable Bayes optimal model grounded in natural scene statistics predicts human bias, precision, and trial-by-trial errors without fitting parameters to the human data. The similarities between human and model performance suggest that the complex human performance patterns with natural stimuli are lawful, and that human visual systems have internalized local image and scene statistics to optimally infer the three-dimensional structure of the environment. These results generalize our understanding of vision from the lab to the real world.
Collapse
Affiliation(s)
- Seha Kim
- Department of Psychology, University of Pennsylvania, Philadelphia, United States
| | - Johannes Burge
- Department of Psychology, University of Pennsylvania, Philadelphia, United States
| |
Collapse
|
31
|
Abstract
The world has a complex, three-dimensional (3-D) spatial structure, but until recently the neural representation of space was studied primarily in planar horizontal environments. Here we review the emerging literature on allocentric spatial representations in 3-D and discuss the relations between 3-D spatial perception and the underlying neural codes. We suggest that the statistics of movements through space determine the topology and the dimensionality of the neural representation, across species and different behavioral modes. We argue that hippocampal place-cell maps are metric in all three dimensions, and might be composed of 2-D and 3-D fragments that are stitched together into a global 3-D metric representation via the 3-D head-direction cells. Finally, we propose that the hippocampal formation might implement a neural analogue of a Kalman filter, a standard engineering algorithm used for 3-D navigation.
Collapse
Affiliation(s)
- Arseny Finkelstein
- Department of Neurobiology, Weizmann Institute of Science, Rehovot 76100, Israel;
| | - Liora Las
- Department of Neurobiology, Weizmann Institute of Science, Rehovot 76100, Israel;
| | - Nachum Ulanovsky
- Department of Neurobiology, Weizmann Institute of Science, Rehovot 76100, Israel;
| |
Collapse
|
32
|
Abstract
In 1992, Goodale and Milner proposed a division of labor in the visual pathways of the primate cerebral cortex. According to their account, the ventral pathway, which projects to occipitotemporal cortex, constructs our visual percepts, while the dorsal pathway, which projects to posterior parietal cortex, mediates the visual control of action. Although the framing of the two-visual-system hypothesis has not been without controversy, it is clear that vision for action and vision for perception have distinct computational requirements, and significant support for the proposed neuroanatomic division has continued to emerge over the last two decades from human neuropsychology, neuroimaging, behavioral psychophysics, and monkey neurophysiology. In this chapter, we review much of this evidence, with a particular focus on recent findings from human neuroimaging and monkey neurophysiology, demonstrating a specialized role for parietal cortex in visually guided behavior. But even though the available evidence suggests that dedicated circuits mediate action and perception, in order to produce adaptive goal-directed behavior there must be a close coupling and seamless integration of information processing across these two systems. We discuss such ventral-dorsal-stream interactions and argue that the two pathways play different, yet complementary, roles in the production of skilled behavior.
Collapse
Affiliation(s)
- Jason P Gallivan
- Department of Psychology, Queen's University, Kingston, Ontario, Canada; Department of Biomedical and Molecular Sciences, Queen's University, Kingston, Ontario, Canada; Centre for Neuroscience Studies, Queen's University, Kingston, Ontario, Canada
| | - Melvyn A Goodale
- Department of Psychology, University of Western Ontario, London, Ontario, Canada; Brain and Mind Institute, University of Western Ontario, London, Ontario, Canada.
| |
Collapse
|
33
|
Abstract
Surface orientation is an important visual primitive that can be estimated from monocular or binocular (stereoscopic) signals. Changes in motor planning occur within about 200 ms after either type of signal is perturbed, but the time it takes for apparent (perceived) slant to develop from stereoscopic cues is not known. Apparent slant sometimes develops very slowly (Gillam, Chambers, & Russo, 1988; van Ee & Erkelens, 1996). However, these long durations could reflect the time it takes for the visual system to resolve conflicts between slant cues that inevitably specify different slants in laboratory displays (Allison & Howard, 2000). We used a speed–accuracy tradeoff analysis to measure the time it takes to discriminate slant, allowing us to report psychometric functions as a function of response time. Observers reported which side of a slanted surface was farther, with a temporal deadline for responding that varied block-to-block. Stereoscopic slant discrimination rose above chance starting at 200 ms after stimulus onset. Unexpectedly, observers discriminated slant from binocular disparity faster than texture, and for stereoscopic whole-field stimuli faster than stereoscopic slant contrast stimuli. However, performance after the initial deviation from chance increased more rapidly for slant-contrast stimuli than whole-field stimuli. Discrimination latencies were similar for slants about the horizontal and vertical axes, but performance increased faster for slants about the vertical axis. Finally, slant from vertical disparity was somewhat slower than slant from horizontal disparity, which may reflect cue conflict. These results demonstrate, in contradiction with the previous literature, that the perception of slant from disparity happens very quickly—in fact, more quickly than the perception of slant from texture—and in comparable time to the simple perception of brightness from luminance.
Collapse
Affiliation(s)
- Baptiste Caziot
- Graduate Center for Vision Research, SUNY College of Optometry, New York, NY, USA.,SUNY Eye Institute, New York, NY, USA.,Department of Neuroscience, Baylor College of Medicine, Houston, TX, USA
| | - Benjamin T Backus
- Graduate Center for Vision Research, SUNY College of Optometry, New York, NY, USA.,SUNY Eye Institute, New York, NY, USA
| | - Esther Lin
- Southern California College of Optometry, Ketchum University, Fullerton, CA, USA
| |
Collapse
|
34
|
Canessa A, Gibaldi A, Chessa M, Fato M, Solari F, Sabatini SP. A dataset of stereoscopic images and ground-truth disparity mimicking human fixations in peripersonal space. Sci Data 2017; 4:170034. [PMID: 28350382 PMCID: PMC5369322 DOI: 10.1038/sdata.2017.34] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2016] [Accepted: 01/13/2017] [Indexed: 01/17/2023] Open
Abstract
Binocular stereopsis is the ability of a visual system, belonging to a live being or a machine, to interpret the different visual information deriving from two eyes/cameras for depth perception. From this perspective, the ground-truth information about three-dimensional visual space, which is hardly available, is an ideal tool both for evaluating human performance and for benchmarking machine vision algorithms. In the present work, we implemented a rendering methodology in which the camera pose mimics realistic eye pose for a fixating observer, thus including convergent eye geometry and cyclotorsion. The virtual environment we developed relies on highly accurate 3D virtual models, and its full controllability allows us to obtain the stereoscopic pairs together with the ground-truth depth and camera pose information. We thus created a stereoscopic dataset: GENUA PESTO-GENoa hUman Active fixation database: PEripersonal space STereoscopic images and grOund truth disparity. The dataset aims to provide a unified framework useful for a number of problems relevant to human and computer vision, from scene exploration and eye movement studies to 3D scene reconstruction.
Collapse
Affiliation(s)
| | | | | | - Marco Fato
- DIBRIS—University of Genoa, Genoa, GE 16145, Italy
| | - Fabio Solari
- DIBRIS—University of Genoa, Genoa, GE 16145, Italy
| | | |
Collapse
|
35
|
Janssen P, Verhoef BE, Premereur E. Functional interactions between the macaque dorsal and ventral visual pathways during three-dimensional object vision. Cortex 2017; 98:218-227. [PMID: 28258716 DOI: 10.1016/j.cortex.2017.01.021] [Citation(s) in RCA: 30] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2016] [Revised: 01/23/2017] [Accepted: 01/25/2017] [Indexed: 11/18/2022]
Abstract
The division of labor between the dorsal and the ventral visual stream in the primate brain has inspired numerous studies on the visual system in humans and in nonhuman primates. However, how and under which circumstances the two visual streams interact is still poorly understood. Here we review evidence from anatomy, modelling, electrophysiology, electrical microstimulation (EM), reversible inactivation and functional imaging in the macaque monkey aimed at clarifying at which levels in the hierarchy of visual areas the two streams interact, and what type of information might be exchanged between the two streams during three-dimensional (3D) object viewing. Neurons in both streams encode 3D structure from binocular disparity, synchronized activity between parietal and inferotemporal areas is present during 3D structure categorization, and clusters of 3D structure-selective neurons in parietal cortex are anatomically connected to ventral stream areas. In addition, caudal intraparietal cortex exerts a causal influence on 3D-structure related activations in more anterior parietal cortex and in inferotemporal cortex. Thus, both anatomical and functional evidence indicates that the dorsal and the ventral visual stream interact during 3D object viewing.
Collapse
Affiliation(s)
- Peter Janssen
- Laboratorium voor Neuro- en Psychofysiologie, KU Leuven, Leuven, Belgium.
| | - Bram-Ernst Verhoef
- Laboratorium voor Neuro- en Psychofysiologie, KU Leuven, Leuven, Belgium; Department of Neurobiology, University of Chicago, Chicago, IL 60637, USA
| | - Elsie Premereur
- Laboratorium voor Neuro- en Psychofysiologie, KU Leuven, Leuven, Belgium
| |
Collapse
|
36
|
Kastner S, Chen Q, Jeong SK, Mruczek REB. A brief comparative review of primate posterior parietal cortex: A novel hypothesis on the human toolmaker. Neuropsychologia 2017; 105:123-134. [PMID: 28159617 DOI: 10.1016/j.neuropsychologia.2017.01.034] [Citation(s) in RCA: 52] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2016] [Revised: 01/26/2017] [Accepted: 01/30/2017] [Indexed: 10/20/2022]
Abstract
The primate visual system contains two major cortical pathways: a ventral-temporal pathway that has been associated with object processing and recognition, and a dorsal-parietal pathway that has been associated with spatial processing and action guidance. Our understanding of the role of the dorsal pathway, in particular, has greatly evolved within the framework of the two-pathway hypothesis since its original conception. Here, we present a comparative review of the primate dorsal pathway in humans and monkeys based on electrophysiological, neuroimaging, neuropsychological, and neuroanatomical studies. We consider similarities and differences across species in terms of the topographic representation of visual space; specificity for eye, reaching, or grasping movements; multi-modal response properties; and the representation of objects and tools. We also review the relative anatomical location of functionally- and topographically-defined regions of the posterior parietal cortex. An emerging theme from this comparative analysis is that non-spatial information is represented to a greater degree, and with increased complexity, in the human dorsal visual system. We propose that non-spatial information in the primate parietal cortex contributes to the perception-to-action system aimed at manipulating objects in peripersonal space. In humans, this network has expanded in multiple ways, including the development of a dorsal object vision system mirroring the complexity of the ventral stream, the integration of object information with parietal working memory systems, and the emergence of tool-specific object representations in the anterior intraparietal sulcus and regions of the inferior parietal lobe. We propose that these evolutionary changes have enabled the emergence of human-specific behaviors, such as the sophisticated use of tools.
Collapse
Affiliation(s)
- S Kastner
- Department of Psychology, USA; Princeton Neuroscience Institute, Princeton University, Princeton, NJ 08544, USA.
| | - Q Chen
- Department of Psychology, USA; School of Psychology, South China Normal University, Guangzhou 510631, China
| | - S K Jeong
- Department of Psychology, USA; Korea Brain Research Institute, Daegu, South Korea
| | - R E B Mruczek
- Department of Psychology, Worcester State University, Worcester, MA 01520, USA
| |
Collapse
|
37
|
Affiliation(s)
- Andrew E. Welchman
- Department of Psychology, University of Cambridge, Cambridge CB2 3EB, United Kingdom;
| |
Collapse
|
38
|
Burge J, McCann BC, Geisler WS. Estimating 3D tilt from local image cues in natural scenes. J Vis 2016; 16:2. [PMID: 27738702 PMCID: PMC5066913 DOI: 10.1167/16.13.2] [Citation(s) in RCA: 25] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2015] [Accepted: 08/15/2016] [Indexed: 11/24/2022] Open
Abstract
Estimating three-dimensional (3D) surface orientation (slant and tilt) is an important first step toward estimating 3D shape. Here, we examine how three local image cues from the same location (disparity gradient, luminance gradient, and dominant texture orientation) should be combined to estimate 3D tilt in natural scenes. We collected a database of natural stereoscopic images with precisely co-registered range images that provide the ground-truth distance at each pixel location. We then analyzed the relationship between ground-truth tilt and image cue values. Our analysis is free of assumptions about the joint probability distributions and yields the Bayes optimal estimates of tilt, given the cue values. Rich results emerge: (a) typical tilt estimates are only moderately accurate and strongly influenced by the cardinal bias in the prior probability distribution; (b) when cue values are similar, or when slant is greater than 40°, estimates are substantially more accurate; (c) when luminance and texture cues agree, they often veto the disparity cue, and when they disagree, they have little effect; and (d) simplifying assumptions common in the cue combination literature is often justified for estimating tilt in natural scenes. The fact that tilt estimates are typically not very accurate is consistent with subjective impressions from viewing small patches of natural scene. The fact that estimates are substantially more accurate for a subset of image locations is also consistent with subjective impressions and with the hypothesis that perceived surface orientation, at more global scales, is achieved by interpolation or extrapolation from estimates at key locations.
Collapse
Affiliation(s)
- Johannes Burge
- Department of Psychology, University of Pennsylvania, Philadelphia, PA,
| | - Brian C McCann
- Texas Advanced Computing Center, University of Texas at Austin, Austin, TX, USA
| | - Wilson S Geisler
- Center for Perceptual Systems and Department of Psychology, University of Texas at Austin, Austin, TX, USA
| |
Collapse
|
39
|
Van Dromme IC, Premereur E, Verhoef BE, Vanduffel W, Janssen P. Posterior Parietal Cortex Drives Inferotemporal Activations During Three-Dimensional Object Vision. PLoS Biol 2016; 14:e1002445. [PMID: 27082854 PMCID: PMC4833303 DOI: 10.1371/journal.pbio.1002445] [Citation(s) in RCA: 72] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2015] [Accepted: 03/18/2016] [Indexed: 11/18/2022] Open
Abstract
The primate visual system consists of a ventral stream, specialized for object recognition, and a dorsal visual stream, which is crucial for spatial vision and actions. However, little is known about the interactions and information flow between these two streams. We investigated these interactions within the network processing three-dimensional (3D) object information, comprising both the dorsal and ventral stream. Reversible inactivation of the macaque caudal intraparietal area (CIP) during functional magnetic resonance imaging (fMRI) reduced fMRI activations in posterior parietal cortex in the dorsal stream and, surprisingly, also in the inferotemporal cortex (ITC) in the ventral visual stream. Moreover, CIP inactivation caused a perceptual deficit in a depth-structure categorization task. CIP-microstimulation during fMRI further suggests that CIP projects via posterior parietal areas to the ITC in the ventral stream. To our knowledge, these results provide the first causal evidence for the flow of visual 3D information from the dorsal stream to the ventral stream, and identify CIP as a key area for depth-structure processing. Thus, combining reversible inactivation and electrical microstimulation during fMRI provides a detailed view of the functional interactions between the two visual processing streams.
Collapse
Affiliation(s)
- Ilse C. Van Dromme
- KU Leuven, Laboratorium voor Neuro- en Psychofysiologie, Leuven, Belgium
| | - Elsie Premereur
- KU Leuven, Laboratorium voor Neuro- en Psychofysiologie, Leuven, Belgium
| | - Bram-Ernst Verhoef
- KU Leuven, Laboratorium voor Neuro- en Psychofysiologie, Leuven, Belgium
- Department of Neurobiology, University of Chicago, Chicago, Illinois, United States of America
| | - Wim Vanduffel
- KU Leuven, Laboratorium voor Neuro- en Psychofysiologie, Leuven, Belgium
- Harvard Medical School, Boston, Massachusetts, United States of America
- MGH Martinos Center for Biomedical Imaging, Charlestown, Massachusetts, United States of America
| | - Peter Janssen
- KU Leuven, Laboratorium voor Neuro- en Psychofysiologie, Leuven, Belgium
- * E-mail:
| |
Collapse
|
40
|
Abstract
The play of light on the retina contains multiple sources of information about the three-dimensional (3D) structure of the world. Some of the best information is derived from differencing operations that act on the images that result from the two eyes’ laterally displaced vantage points. Other information is available in systematic retinal patterns of local texture and motion cues. This article describes what is currently known about the development of sensitivity to these binocular and monocular cues for depth in human infants, and it places the results in the context of what is known about the underlying neural mechanisms from work in nonhuman primates and human neuroimaging studies.
Collapse
Affiliation(s)
- Anthony M. Norcia
- Department of Psychology, Stanford University, Stanford, California 94305;,
| | - Holly E. Gerhard
- Department of Psychology, Stanford University, Stanford, California 94305;,
| |
Collapse
|
41
|
Abstract
The brain's skill in estimating the 3-D orientation of viewed surfaces supports a range of behaviors, from placing an object on a nearby table, to planning the best route when hill walking. This ability relies on integrating depth signals across extensive regions of space that exceed the receptive fields of early sensory neurons. Although hierarchical selection and pooling is central to understanding of the ventral visual pathway, the successive operations in the dorsal stream are poorly understood. Here we use computational modeling of human fMRI signals to probe the computations that extract 3-D surface orientation from binocular disparity. To understand how representations evolve across the hierarchy, we developed an inference approach using a series of generative models to explain the empirical fMRI data in different cortical areas. Specifically, we simulated the responses of candidate visual processing algorithms and tested how well they explained fMRI responses. Thereby we demonstrate a hierarchical refinement of visual representations moving from the representation of edges and figure-ground segmentation (V1, V2) to spatially extensive disparity gradients in V3A. We show that responses in V3A are little affected by low-level image covariates, and have a partial tolerance to the overall depth position. Finally, we show that responses in V3A parallel perceptual judgments of slant. This reveals a relatively short computational hierarchy that captures key information about the 3-D structure of nearby surfaces, and more generally demonstrates an analysis approach that may be of merit in a diverse range of brain imaging domains.
Collapse
|
42
|
Breveglieri R, Galletti C, Bosco A, Gamberini M, Fattori P. Object Affordance Modulates Visual Responses in the Macaque Medial Posterior Parietal Cortex. J Cogn Neurosci 2015; 27:1447-55. [DOI: 10.1162/jocn_a_00793] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Abstract
Area V6A is a visuomotor area of the dorsomedial visual stream that contains cells modulated by object observation and by grip formation. As different objects have different shapes but also evoke different grips, the response selectivity during object presentation could reflect either the coding of object geometry or object affordances. To clarify this point, we here investigate neural responses of V6A cells when monkeys observed two objects with similar visual features but different contextual information, such as the evoked grip type. We demonstrate that many V6A cells respond to the visual presentation of objects and about 30% of them by the object affordance. Given that area V6A is an early stage in the visuomotor processes underlying grasping, these data suggest that V6A may participate in the computation of object affordances. These results add some elements in the recent literature about the role of the dorsal visual stream areas in object representation and contribute in elucidating the neural correlates of the extraction of action-relevant information from general object properties, in agreement with recent neuroimaging studies on humans showing that vision of graspable objects activates action coding in the dorsomedial visual steam.
Collapse
|
43
|
Caminiti R, Innocenti GM, Battaglia-Mayer A. Organization and evolution of parieto-frontal processing streams in macaque monkeys and humans. Neurosci Biobehav Rev 2015; 56:73-96. [PMID: 26112130 DOI: 10.1016/j.neubiorev.2015.06.014] [Citation(s) in RCA: 59] [Impact Index Per Article: 5.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2015] [Revised: 05/08/2015] [Accepted: 06/09/2015] [Indexed: 01/01/2023]
Abstract
The functional organization of the parieto-frontal system is crucial for understanding cognitive-motor behavior and provides the basis for interpreting the consequences of parietal lesions in humans from a neurobiological perspective. The parieto-frontal connectivity defines some main information streams that, rather than being devoted to restricted functions, underlie a rich behavioral repertoire. Surprisingly, from macaque to humans, evolution has added only a few, new functional streams, increasing however their complexity and encoding power. In fact, the characterization of the conduction times of parietal and frontal areas to different target structures has recently opened a new window on cortical dynamics, suggesting that evolution has amplified the probability of dynamic interactions between the nodes of the network, thanks to communication patterns based on temporally-dispersed conduction delays. This might allow the representation of sensory-motor signals within multiple neural assemblies and reference frames, as to optimize sensory-motor remapping within an action space characterized by different and more complex demands across evolution.
Collapse
Affiliation(s)
- Roberto Caminiti
- Department of Physiology and Pharmacology, University of Rome SAPIENZA, P.le Aldo Moro 5, 00185 Rome, Italy.
| | - Giorgio M Innocenti
- Department of Neuroscience, Karolinska Institutet, Stockholm, Sweden; Brain and Mind Institute, Federal Institute of Technology, EPFL, Lausanne, Switzerland
| | - Alexandra Battaglia-Mayer
- Department of Physiology and Pharmacology, University of Rome SAPIENZA, P.le Aldo Moro 5, 00185 Rome, Italy
| |
Collapse
|
44
|
Theys T, Romero MC, van Loon J, Janssen P. Shape representations in the primate dorsal visual stream. Front Comput Neurosci 2015; 9:43. [PMID: 25954189 PMCID: PMC4406065 DOI: 10.3389/fncom.2015.00043] [Citation(s) in RCA: 24] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2014] [Accepted: 03/20/2015] [Indexed: 11/13/2022] Open
Abstract
The primate visual system extracts object shape information for object recognition in the ventral visual stream. Recent research has demonstrated that object shape is also processed in the dorsal visual stream, which is specialized for spatial vision and the planning of actions. A number of studies have investigated the coding of 2D shape in the anterior intraparietal area (AIP), one of the end-stage areas of the dorsal stream which has been implicated in the extraction of affordances for the purpose of grasping. These findings challenge the current understanding of area AIP as a critical stage in the dorsal stream for the extraction of object affordances. The representation of three-dimensional (3D) shape has been studied in two interconnected areas known to be critical for object grasping: area AIP and area F5a in the ventral premotor cortex (PMv), to which AIP projects. In both areas neurons respond selectively to 3D shape defined by binocular disparity, but the latency of the neural selectivity is approximately 10 ms longer in F5a compared to AIP, consistent with its higher position in the hierarchy of cortical areas. Furthermore, F5a neurons were more sensitive to small amplitudes of 3D curvature and could detect subtle differences in 3D structure more reliably than AIP neurons. In both areas, 3D-shape selective neurons were co-localized with neurons showing motor-related activity during object grasping in the dark, indicating a close convergence of visual and motor information on the same clusters of neurons.
Collapse
Affiliation(s)
- Tom Theys
- Laboratorium voor Neuro- en Psychofysiologie, Katholieke Universiteit Leuven Leuven, Belgium ; Afdeling Experimentele Neurochirurgie en Neuroanatomie, Katholieke Universiteit Leuven Leuven, Belgium
| | - Maria C Romero
- Laboratorium voor Neuro- en Psychofysiologie, Katholieke Universiteit Leuven Leuven, Belgium
| | - Johannes van Loon
- Afdeling Experimentele Neurochirurgie en Neuroanatomie, Katholieke Universiteit Leuven Leuven, Belgium
| | - Peter Janssen
- Laboratorium voor Neuro- en Psychofysiologie, Katholieke Universiteit Leuven Leuven, Belgium
| |
Collapse
|
45
|
Abstract
Sensory systems encode the environment in egocentric (e.g., eye, head, or body) reference frames, creating inherently unstable representations that shift and rotate as we move. However, it is widely speculated that the brain transforms these signals into an allocentric, gravity-centered representation of the world that is stable and independent of the observer's spatial pose. Where and how this representation may be achieved is currently unknown. Here we demonstrate that a subpopulation of neurons in the macaque caudal intraparietal area (CIP) visually encodes object tilt in nonegocentric coordinates defined relative to the gravitational vector. Neuronal responses to the tilt of a visually presented planar surface were measured with the monkey in different spatial orientations (upright and rolled left/right ear down) and then compared. This revealed a continuum of representations in which planar tilt was encoded in a gravity-centered reference frame in approximately one-tenth of the comparisons, intermediate reference frames ranging between gravity-centered and egocentric in approximately two-tenths of the comparisons, and in an egocentric reference frame in less than half of the comparisons. Altogether, almost half of the comparisons revealed a shift in the preferred tilt and/or a gain change consistent with encoding object orientation in nonegocentric coordinates. Through neural network modeling, we further show that a purely gravity-centered representation of object tilt can be achieved directly from the population activity of CIP-like units. These results suggest that area CIP may play a key role in creating a stable, allocentric representation of the environment defined relative to an "earth-vertical" direction.
Collapse
|
46
|
Reliability-dependent contributions of visual orientation cues in parietal cortex. Proc Natl Acad Sci U S A 2014; 111:18043-8. [PMID: 25427796 DOI: 10.1073/pnas.1421131111] [Citation(s) in RCA: 24] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Creating accurate 3D representations of the world from 2D retinal images is a fundamental task for the visual system. However, the reliability of different 3D visual signals depends inherently on viewing geometry, such as how much an object is slanted in depth. Human perceptual studies have correspondingly shown that texture and binocular disparity cues for object orientation are combined according to their slant-dependent reliabilities. Where and how this cue combination occurs in the brain is currently unknown. Here, we search for neural correlates of this property in the macaque caudal intraparietal area (CIP) by measuring slant tuning curves using mixed-cue (texture + disparity) and cue-isolated (texture or disparity) planar stimuli. We find that texture cues contribute more to the mixed-cue responses of CIP neurons that prefer larger slants, consistent with theoretical and psychophysical results showing that the reliability of texture relative to disparity cues increases with slant angle. By analyzing responses to binocularly viewed texture stimuli with conflicting texture and disparity information, some cells that are sensitive to both cues when presented in isolation are found to disregard one of the cues during cue conflict. Additionally, the similarity between texture and mixed-cue responses is found to be greater when this cue conflict is eliminated by presenting the texture stimuli monocularly. The present findings demonstrate reliability-dependent contributions of visual orientation cues at the level of the CIP, thus revealing a neural correlate of this property of human visual perception.
Collapse
|
47
|
Rezai O, Kleinhans A, Matallanas E, Selby B, Tripp BP. Modeling the shape hierarchy for visually guided grasping. Front Comput Neurosci 2014; 8:132. [PMID: 25386134 PMCID: PMC4209868 DOI: 10.3389/fncom.2014.00132] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2014] [Accepted: 09/26/2014] [Indexed: 11/25/2022] Open
Abstract
The monkey anterior intraparietal area (AIP) encodes visual information about three-dimensional object shape that is used to shape the hand for grasping. We modeled shape tuning in visual AIP neurons and its relationship with curvature and gradient information from the caudal intraparietal area (CIP). The main goal was to gain insight into the kinds of shape parameterizations that can account for AIP tuning and that are consistent with both the inputs to AIP and the role of AIP in grasping. We first experimented with superquadric shape parameters. We considered superquadrics because they occupy a role in robotics that is similar to AIP, in that superquadric fits are derived from visual input and used for grasp planning. We also experimented with an alternative shape parameterization that was based on an Isomap dimension reduction of spatial derivatives of depth (i.e., distance from the observer to the object surface). We considered an Isomap-based model because its parameters lacked discontinuities between similar shapes. When we matched the dimension of the Isomap to the number of superquadric parameters, the superquadric model fit the AIP data somewhat more closely. However, higher-dimensional Isomaps provided excellent fits. Also, we found that the Isomap parameters could be approximated much more accurately than superquadric parameters by feedforward neural networks with CIP-like inputs. We conclude that Isomaps, or perhaps alternative dimension reductions of visual inputs to AIP, provide a promising model of AIP electrophysiology data. Further work is needed to test whether such shape parameterizations actually provide an effective basis for grasp control.
Collapse
Affiliation(s)
- Omid Rezai
- Department of Systems Design Engineering, Centre for Theoretical Neuroscience, University of Waterloo Waterloo, ON, Canada
| | - Ashley Kleinhans
- Mobile Intelligent Autonomous Systems, Council for Scientific and Industrial Research Pretoria, South Africa ; School of Mechanical and Industrial Engineering, University of Johannesburg Johannesburg, South Africa
| | | | - Ben Selby
- Department of Systems Design Engineering, Centre for Theoretical Neuroscience, University of Waterloo Waterloo, ON, Canada
| | - Bryan P Tripp
- Department of Systems Design Engineering, Centre for Theoretical Neuroscience, University of Waterloo Waterloo, ON, Canada
| |
Collapse
|
48
|
Seilheimer RL, Rosenberg A, Angelaki DE. Models and processes of multisensory cue combination. Curr Opin Neurobiol 2013; 25:38-46. [PMID: 24709599 DOI: 10.1016/j.conb.2013.11.008] [Citation(s) in RCA: 69] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2013] [Revised: 09/26/2013] [Accepted: 11/18/2013] [Indexed: 01/13/2023]
Abstract
Fundamental to our perception of a unified and stable environment is the capacity to combine information across the senses. Although this process appears seamless as an adult, the brain's ability to successfully perform multisensory cue combination takes years to develop and relies on a number of complex processes including cue integration, cue calibration, causal inference, and reference frame transformations. Further complexities exist because multisensory cue combination is implemented across time by populations of noisy neurons. In this review, we discuss recent behavioral studies exploring how the brain combines information from different sensory systems, neurophysiological studies relating behavior to neuronal activity, and a theory of neural sensory encoding that can account for many of these experimental findings.
Collapse
Affiliation(s)
| | - Ari Rosenberg
- Baylor College of Medicine, Houston, TX, United States
| | | |
Collapse
|