1
|
Moreno-Sánchez M, Matsushima EH, Aznar-Casanova JA. Exploring Binocular Visual Attention by Presenting Rapid Dichoptic and Dioptic Series. Brain Sci 2024; 14:518. [PMID: 38790496 PMCID: PMC11119880 DOI: 10.3390/brainsci14050518] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2024] [Revised: 05/12/2024] [Accepted: 05/16/2024] [Indexed: 05/26/2024] Open
Abstract
This study addresses an issue in attentional distribution in a binocular visual system using RSVP tasks under Attentional Blink (AB) experimental protocols. In Experiment 1, we employed dichoptic RSVP to verify whether, under interocular competition, attention may be captured by a monocular channel. Experiment 2 was a control experiment, where a monoptic RSVP assessed by both or only one eye determines whether Experiment 1 monocular condition results were due to an allocation of attention to one eye. Experiment 3 was also a control experiment designed to determine whether Experiment 1 results were due to the effect of interocular competition or to a diminished visual contrast. Results from Experiment 1 revealed that dichoptic presentations caused a delay in the type stage of the Wyble's eSTST model, postponing the subsequent tokenization process. The delay in monocular conditions may be further explained by a visual attenuation, due to fusion of target and an empty frame. Experiment 2 evidenced the attentional allocation to monocular channels when forced by eye occlusion. Experiment 3 disclosed that monocular performance in Experiment 1 differs significantly from conditions with interocular competition. While both experiments revealed similar performance in monocular conditions, rivalry conditions exhibit lower detection rates, suggesting that competing stimuli was not responsible for Experiment 1 results. These findings highlight the differences between dichoptic and monoptic presentations of stimuli, particularly on the AB effect, which appears attenuated or absent in dichoptic settings. Furthermore, results suggest that monoptic presentation and binocular fusion stages were a necessary condition for the attentional allocation.
Collapse
Affiliation(s)
- Manuel Moreno-Sánchez
- Facultad de Psicologia, Department Section of Cognitive Processes, Universitat de Barcelona, 08035 Barcelona, Spain;
- Institut of Neuroscience, Universitat de Barcelona, 08028 Barcelona, Spain
| | - Elton H. Matsushima
- Programa de Pós-Graduação em Medicina (Neurologia and Eurociências), Niterói 24020-140, RJ, Brazil;
- Institute of Psychology, Universidade Federal Fluminense, Niterói 24020-140, RJ, Brazil
| | - Jose Antonio Aznar-Casanova
- Facultad de Psicologia, Department Section of Cognitive Processes, Universitat de Barcelona, 08035 Barcelona, Spain;
- Institut of Neuroscience, Universitat de Barcelona, 08028 Barcelona, Spain
| |
Collapse
|
2
|
Sadeghi R, Ressmeyer R, Yates J, Otero-Millan J. Open Iris - An Open Source Framework for Video-Based Eye-Tracking Research and Development. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.02.27.582401. [PMID: 38463977 PMCID: PMC10925248 DOI: 10.1101/2024.02.27.582401] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/12/2024]
Abstract
Eye-tracking is an essential tool in many fields, yet existing solutions are often limited for customized applications due to cost or lack of flexibility. We present OpenIris, an adaptable and user-friendly open-source framework for video-based eye-tracking. OpenIris is developed in C# with modular design that allows further extension and customization through plugins for different hardware systems, tracking, and calibration pipelines. It can be remotely controlled via a network interface from other devices or programs. Eye movements can be recorded online from camera stream or offline post-processing recorded videos. Example plugins have been developed to track eye motion in 3-D, including torsion. Currently implemented binocular pupil tracking pipelines can achieve frame rates of more than 500Hz. With the OpenIris framework, we aim to fill a gap in the research tools available for high-precision and high-speed eye-tracking, especially in environments that require custom solutions that are not currently well-served by commercial eye-trackers.
Collapse
Affiliation(s)
- Roksana Sadeghi
- Herbert Wertheim School of Optometry and Vision Science, University of California, Berkeley, California, USA
| | - Ryan Ressmeyer
- Bioengineering, University of Washington, Seattle, Washington, USA
| | - Jacob Yates
- Herbert Wertheim School of Optometry and Vision Science, University of California, Berkeley, California, USA
| | - Jorge Otero-Millan
- Herbert Wertheim School of Optometry and Vision Science, University of California, Berkeley, California, USA
- Department of Neurology, Johns Hopkins University, Baltimore, Maryland, USA
| |
Collapse
|
3
|
DuTell V, Gibaldi A, Focarelli G, Olshausen BA, Banks MS. High-fidelity eye, head, body, and world tracking with a wearable device. Behav Res Methods 2024; 56:32-42. [PMID: 35879503 PMCID: PMC10794349 DOI: 10.3758/s13428-022-01888-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 05/09/2022] [Indexed: 12/13/2022]
Abstract
We describe the design and performance of a high-fidelity wearable head-, body-, and eye-tracking system that offers significant improvement over previous such devices. This device's sensors include a binocular eye tracker, an RGB-D scene camera, a high-frame-rate scene camera, and two visual odometry sensors, for a total of ten cameras, which we synchronize and record from with a data rate of over 700 MB/s. The sensors are operated by a mini-PC optimized for fast data collection, and powered by a small battery pack. The device records a subject's eye, head, and body positions, simultaneously with RGB and depth data from the subject's visual environment, measured with high spatial and temporal resolution. The headset weighs only 1.4 kg, and the backpack with batteries 3.9 kg. The device can be comfortably worn by the subject, allowing a high degree of mobility. Together, this system overcomes many limitations of previous such systems, allowing high-fidelity characterization of the dynamics of natural vision.
Collapse
Affiliation(s)
- Vasha DuTell
- Wertheim School of Optometry and Vision Science, UC Berkeley, Minor Hall, Berkeley, CA, USA.
- Redwood Center for Theoretical Neuroscience, UC Berkeley, Evans Hall, Berkeley, CA, USA.
| | - Agostino Gibaldi
- Wertheim School of Optometry and Vision Science, UC Berkeley, Minor Hall, Berkeley, CA, USA
| | - Giulia Focarelli
- Wertheim School of Optometry and Vision Science, UC Berkeley, Minor Hall, Berkeley, CA, USA
| | - Bruno A Olshausen
- Wertheim School of Optometry and Vision Science, UC Berkeley, Minor Hall, Berkeley, CA, USA
- Redwood Center for Theoretical Neuroscience, UC Berkeley, Evans Hall, Berkeley, CA, USA
| | - Martin S Banks
- Wertheim School of Optometry and Vision Science, UC Berkeley, Minor Hall, Berkeley, CA, USA
| |
Collapse
|
4
|
Kim H, Koike Y, Choi W, Lee J. The effect of different depth planes during a manual tracking task in three-dimensional virtual reality space. Sci Rep 2023; 13:21499. [PMID: 38057361 PMCID: PMC10700492 DOI: 10.1038/s41598-023-48869-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/20/2023] [Accepted: 11/30/2023] [Indexed: 12/08/2023] Open
Abstract
Unlike ballistic arm movements such as reaching, the contribution of depth information to the performance of manual tracking movements is unclear. Thus, to understand how the brain handles information, we investigated how a required movement along the depth axis would affect behavioral tracking performance, postulating that it would be affected by the amount of depth movement. We designed a visually guided planar tracking task that requires movement on three planes with different depths: a fronto-parallel plane called ROT (0), a sagittal plane called ROT (90), and a plane rotated by 45° with respect to the sagittal plane called ROT (45). Fifteen participants performed a circular manual tracking task under binocular and monocular visions in a three-dimensional (3D) virtual reality space. As a result, under binocular vision, ROT (90), which required the largest depth movement among the tasks, showed the greatest error in 3D. Similarly, the errors (deviation from the target path) on the depth axis revealed significant differences among the tasks. Under monocular vision, significant differences in errors were observed only on the lateral axis. Moreover, we observed that the errors in the lateral and depth axes were proportional to the required movement on these axes under binocular vision and confirmed that the required depth movement under binocular vision determined depth error independent of the other axes. This finding implies that the brain may independently process binocular vision information on each axis. Meanwhile, the required depth movement under monocular vision was independent of performance along the depth axis, indicating an intractable behavior. Our findings highlight the importance of handling depth movement, especially when a virtual reality situation, involving tracking tasks, is generated.
Collapse
Affiliation(s)
- Hyeonseok Kim
- Swartz Center for Computational Neuroscience, Institute for Neural Computation, University of California San Diego, La Jolla, CA, 92093, USA
| | - Yasuharu Koike
- Institute of Innovative Research, Tokyo Institute of Technology, Yokohama, 226-8503, Japan
| | - Woong Choi
- College of ICT Construction & Welfare Convergence, Kangnam University, Yongin, 16979, Republic of Korea.
| | - Jongho Lee
- Department of Clinical Engineering, Komatsu University, Komatsu, 923-0961, Japan.
| |
Collapse
|
5
|
Hooge ITC, Niehorster DC, Hessels RS, Benjamins JS, Nyström M. How robust are wearable eye trackers to slow and fast head and body movements? Behav Res Methods 2023; 55:4128-4142. [PMID: 36326998 PMCID: PMC10700439 DOI: 10.3758/s13428-022-02010-3] [Citation(s) in RCA: 9] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 10/11/2022] [Indexed: 06/16/2023]
Abstract
How well can modern wearable eye trackers cope with head and body movement? To investigate this question, we asked four participants to stand still, walk, skip, and jump while fixating a static physical target in space. We did this for six different eye trackers. All the eye trackers were capable of recording gaze during the most dynamic episodes (skipping and jumping). The accuracy became worse as movement got wilder. During skipping and jumping, the biggest error was 5.8∘. However, most errors were smaller than 3∘. We discuss the implications of decreased accuracy in the context of different research scenarios.
Collapse
Affiliation(s)
- Ignace T C Hooge
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, The Netherlands.
| | - Diederick C Niehorster
- Lund University Humanities Lab and Department of Psychology, Lund University, Lund, Sweden
| | - Roy S Hessels
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, The Netherlands
| | - Jeroen S Benjamins
- Experimental Psychology, Helmholtz Institute, and Social, Health and Organisational Psychology, Utrecht University, Utrecht, The Netherlands
| | - Marcus Nyström
- Lund University Humanities Lab, Lund University, Lund, Sweden
| |
Collapse
|
6
|
Talluri BC, Kang I, Lazere A, Quinn KR, Kaliss N, Yates JL, Butts DA, Nienborg H. Activity in primate visual cortex is minimally driven by spontaneous movements. Nat Neurosci 2023; 26:1953-1959. [PMID: 37828227 PMCID: PMC10620084 DOI: 10.1038/s41593-023-01459-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2022] [Accepted: 09/08/2023] [Indexed: 10/14/2023]
Abstract
Organisms process sensory information in the context of their own moving bodies, an idea referred to as embodiment. This idea is important for developmental neuroscience, robotics and systems neuroscience. The mechanisms supporting embodiment are unknown, but a manifestation could be the observation in mice of brain-wide neuromodulation, including in the primary visual cortex, driven by task-irrelevant spontaneous body movements. We tested this hypothesis in macaque monkeys (Macaca mulatta), a primate model for human vision, by simultaneously recording visual cortex activity and facial and body movements. We also sought a direct comparison using an analogous approach to those used in mouse studies. Here we found that activity in the primate visual cortex (V1, V2 and V3/V3A) was associated with the animals' own movements, but this modulation was largely explained by the impact of the movements on the retinal image, that is, by changes in visual input. These results indicate that visual cortex in primates is minimally driven by spontaneous movements and may reflect species-specific sensorimotor strategies.
Collapse
Affiliation(s)
- Bharath Chandra Talluri
- Laboratory of Sensorimotor Research, National Eye Institute, National Institutes of Health, Bethesda, MD, USA
| | - Incheol Kang
- Laboratory of Sensorimotor Research, National Eye Institute, National Institutes of Health, Bethesda, MD, USA
| | - Adam Lazere
- Laboratory of Sensorimotor Research, National Eye Institute, National Institutes of Health, Bethesda, MD, USA
| | - Katrina R Quinn
- Center for Integrative Neuroscience, University of Tübingen, Tübingen, Germany
| | - Nicholas Kaliss
- Laboratory of Sensorimotor Research, National Eye Institute, National Institutes of Health, Bethesda, MD, USA
| | - Jacob L Yates
- Herbert Wertheim School of Optometry & Vision Science, University of California, Berkeley, Berkeley, CA, USA
- Department of Biology and Program in Neuroscience and Cognitive Science, University of Maryland, College Park, MD, USA
| | - Daniel A Butts
- Department of Biology and Program in Neuroscience and Cognitive Science, University of Maryland, College Park, MD, USA
| | - Hendrikje Nienborg
- Laboratory of Sensorimotor Research, National Eye Institute, National Institutes of Health, Bethesda, MD, USA.
| |
Collapse
|
7
|
Abstract
The superior colliculus (SC) is a subcortical brain structure that is relevant for sensation, cognition, and action. In nonhuman primates, a rich history of studies has provided unprecedented detail about this structure's role in controlling orienting behaviors; as a result, the primate SC has become primarily regarded as a motor control structure. However, as in other species, the primate SC is also a highly visual structure: A fraction of its inputs is retinal and complemented by inputs from visual cortical areas, including the primary visual cortex. Motivated by this, recent investigations are revealing the rich visual pattern analysis capabilities of the primate SC, placing this structure in an ideal position to guide orienting movements. The anatomical proximity of the primate SC to both early visual inputs and final motor control apparatuses, as well as its ascending feedback projections to the cortex, affirms an important role for this structure in active perception.
Collapse
Affiliation(s)
- Ziad M Hafed
- Werner Reichardt Centre for Integrative Neuroscience, University of Tübingen, Tübingen, Germany;
- Hertie Institute for Clinical Brain Research, University of Tübingen, Tübingen, Germany
| | | | - Chih-Yang Chen
- Institute for the Advanced Study of Human Biology, Kyoto University, Kyoto, Japan;
| | - Amarender R Bogadhi
- Central Nervous System Diseases Research, Boehringer Ingelheim Pharma GmbH & Co. KG, Biberach, Germany;
| |
Collapse
|
8
|
Sili D, De Giorgi C, Pizzuti A, Spezialetti M, de Pasquale F, Betti V. The spatio-temporal architecture of everyday manual behavior. Sci Rep 2023; 13:9451. [PMID: 37296243 PMCID: PMC10256758 DOI: 10.1038/s41598-023-36280-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2023] [Accepted: 05/31/2023] [Indexed: 06/12/2023] Open
Abstract
In everyday activities, humans move alike to manipulate objects. Prior works suggest that hand movements are built by a limited set of basic building blocks consisting of a set of common postures. However, how the low dimensionality of hand movements supports the adaptability and flexibility of natural behavior is unknown. Through a sensorized glove, we collected kinematics data from thirty-six participants preparing and having breakfast in naturalistic conditions. By means of an unbiased analysis, we identified a repertoire of hand states. Then, we tracked their transitions over time. We found that manual behavior can be described in space through a complex organization of basic configurations. These, even in an unconstrained experiment, recurred across subjects. A specific temporal structure, highly consistent within the sample, seems to integrate such identified hand shapes to realize skilled movements. These findings suggest that the simplification of the motor commands unravels in the temporal dimension more than in the spatial one.
Collapse
Affiliation(s)
- Daniele Sili
- Department of Psychology, Sapienza University of Rome, Roma, Italy
- IRCCS Fondazione Santa Lucia, Roma, Italy
| | - Chiara De Giorgi
- Department of Psychology, Sapienza University of Rome, Roma, Italy
- IRCCS Fondazione Santa Lucia, Roma, Italy
| | - Alessandra Pizzuti
- Department of Psychology, Sapienza University of Rome, Roma, Italy
- IRCCS Fondazione Santa Lucia, Roma, Italy
| | - Matteo Spezialetti
- Department of Psychology, Sapienza University of Rome, Roma, Italy
- IRCCS Fondazione Santa Lucia, Roma, Italy
| | | | - Viviana Betti
- Department of Psychology, Sapienza University of Rome, Roma, Italy.
- IRCCS Fondazione Santa Lucia, Roma, Italy.
| |
Collapse
|
9
|
Krauze L, Delesa-Velina M, Pladere T, Krumina G. Why 2D layout in 3D images matters: evidence from visual search and eyetracking. J Eye Mov Res 2023; 16:10.16910/jemr.16.1.4. [PMID: 37965285 PMCID: PMC10643048 DOI: 10.16910/jemr.16.1.4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2023] Open
Abstract
Precise perception of three-dimensional (3D) images is crucial for a rewarding experience when using novel displays. However, the capability of the human visual system to perceive binocular disparities varies across the visual field meaning that depth perception might be affected by the two-dimensional (2D) layout of items on the screen. Nevertheless, potential difficulties in perceiving 3D images during free viewing have received only a little attention so far, limiting opportunities to enhance visual effectiveness of information presentation. The aim of this study was to elucidate how the 2D layout of items in 3D images impacts visual search and distribution of maintaining attention based on the analysis of the viewer's gaze. Participants were searching for a target which was projected one plane closer to the viewer compared to distractors on a multi-plane display. The 2D layout of items was manipulated by changing the item distance from the center of the display plane from 2° to 8°. As a result, the targets were identified correctly when the items were displayed close to the center of the display plane, however, the number of errors grew with an increase in distance. Moreover, correct responses were given more often when subjects paid more attention to targets compared to other items on the screen. However, a more balanced distribution of attention over time across all items was characteristic of the incorrectly completed trials. Thus, our results suggest that items should be displayed close to each other in a 2D layout to facilitate precise perception of 3D images and considering distribution of attention maintenance based on eye-tracking might be useful in the objective assessment of user experience for novel displays.
Collapse
|
10
|
Aizenman AM, Koulieris GA, Gibaldi A, Sehgal V, Levi DM, Banks MS. The Statistics of Eye Movements and Binocular Disparities during VR Gaming: Implications for Headset Design. ACM TRANSACTIONS ON GRAPHICS 2023; 42:7. [PMID: 37122317 PMCID: PMC10139447 DOI: 10.1145/3549529] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 05/03/2023]
Abstract
The human visual system evolved in environments with statistical regularities. Binocular vision is adapted to these such that depth perception and eye movements are more precise, faster, and performed comfortably in environments consistent with the regularities. We measured the statistics of eye movements and binocular disparities in virtual-reality (VR) - gaming environments and found that they are quite different from those in the natural environment. Fixation distance and direction are more restricted in VR, and fixation distance is farther. The pattern of disparity across the visual field is less regular in VR and does not conform to a prominent property of naturally occurring disparities. From this we predict that double vision is more likely in VR than in the natural environment. We also determined the optimal screen distance to minimize discomfort due to the vergence-accommodation conflict, and the optimal nasal-temporal positioning of head-mounted display (HMD) screens to maximize binocular field of view. Finally, in a user study we investigated how VR content affects comfort and performance. Content that is more consistent with the statistics of the natural world yields less discomfort than content that is not. Furthermore, consistent content yields slightly better performance than inconsistent content.
Collapse
|
11
|
Aguilera-Francisco A, Serrano-Pedraza I. The Stereoscopic Anisotropy Is Smaller in Elderly Population. Invest Ophthalmol Vis Sci 2022; 63:26. [DOI: 10.1167/iovs.63.12.26] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Affiliation(s)
- Aracelis Aguilera-Francisco
- Department of Experimental Psychology, Faculty of Psychology, Universidad Complutense de Madrid, Campus de Somosaguas, Madrid, Spain
| | - Ignacio Serrano-Pedraza
- Department of Experimental Psychology, Faculty of Psychology, Universidad Complutense de Madrid, Campus de Somosaguas, Madrid, Spain
| |
Collapse
|
12
|
Abstract
Stereopsis provides us with a vivid impression of the depth and distance of objects in our 3- dimensional world. Stereopsis is important for a number of everyday visual tasks, including (but not limited to) reaching and grasping, fine visuo-motor control, and navigating in our world. This review briefly discusses the neural substrate for normal binocular vision and stereopsis and its development in primates; outlines some of the issues and limitations of stereopsis tests and examines some of the factors that limit the typical development of stereopsis and the causes and consequences of stereo-deficiency and stereo-blindness. Finally, we review several approaches to improving or recovering stereopsis in both neurotypical individuals and those with stereo-deficiency and stereo-blindness and outline some emerging strategies for improving stereopsis.
Collapse
|
13
|
Kalou K, Sedda G, Gibaldi A, Sabatini SP. Learning bio-inspired head-centric representations of 3D shapes in an active fixation setting. Front Robot AI 2022; 9:994284. [PMID: 36329691 PMCID: PMC9623882 DOI: 10.3389/frobt.2022.994284] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/14/2022] [Accepted: 09/13/2022] [Indexed: 12/03/2022] Open
Abstract
When exploring the surrounding environment with the eyes, humans and primates need to interpret three-dimensional (3D) shapes in a fast and invariant way, exploiting a highly variant and gaze-dependent visual information. Since they have front-facing eyes, binocular disparity is a prominent cue for depth perception. Specifically, it serves as computational substrate for two ground mechanisms of binocular active vision: stereopsis and binocular coordination. To this aim, disparity information, which is expressed in a retinotopic reference frame, is combined along the visual cortical pathways with gaze information and transformed in a head-centric reference frame. Despite the importance of this mechanism, the underlying neural substrates still remain widely unknown. In this work, we investigate the capabilities of the human visual system to interpret the 3D scene exploiting disparity and gaze information. In a psychophysical experiment, human subjects were asked to judge the depth orientation of a planar surface either while fixating a target point or while freely exploring the surface. Moreover, we used the same stimuli to train a recurrent neural network to exploit the responses of a modelled population of cortical (V1) cells to interpret the 3D scene layout. The results for both human performance and from the model network show that integrating disparity information across gaze directions is crucial for a reliable and invariant interpretation of the 3D geometry of the scene.
Collapse
Affiliation(s)
- Katerina Kalou
- Department of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genoa, Italy
- *Correspondence: Katerina Kalou,
| | - Giulia Sedda
- Department of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genoa, Italy
| | - Agostino Gibaldi
- University of California Berkeley, School of Optometry, Berkeley, CA, United States
| | - Silvio P. Sabatini
- Department of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genoa, Italy
| |
Collapse
|
14
|
Wagner P, Ho A, Kim J. Estimating 3D spatiotemporal point of regard: a device evaluation. JOURNAL OF THE OPTICAL SOCIETY OF AMERICA. A, OPTICS, IMAGE SCIENCE, AND VISION 2022; 39:1343-1351. [PMID: 36215577 DOI: 10.1364/josaa.457663] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/18/2022] [Accepted: 06/10/2022] [Indexed: 06/16/2023]
Abstract
This paper presents and evaluates a system and method that record spatiotemporal scene information and location of the center of visual attention, i.e., spatiotemporal point of regard (PoR) in ecological environments. A primary research application of the proposed system and method is for enhancing current 2D visual attention models. Current eye-tracking approaches collapse a scene's depth structures to a 2D image, omitting visual cues that trigger important functions of the human visual system (e.g., accommodation and vergence). We combined head-mounted eye-tracking with a miniature time-of-flight camera to produce a system that could be used to estimate the spatiotemporal location of the PoR-the point of highest visual attention-within 3D scene layouts. Maintaining calibration accuracy is a primary challenge for gaze mapping; hence, we measured accuracy repeatedly by matching the PoR to fixated targets arranged within a range of working distances in depth. Accuracy was estimated as the deviation from estimated PoR relative to known locations of scene targets. We found that estimates of 3D PoR had an overall accuracy of approximately 2° omnidirectional mean average error (OMAE) with variation over a 1 h recording maintained within 3.6° OMAE. This method can be used to determine accommodation and vergence cues of the human visual system continuously within habitual environments, including everyday applications (e.g., use of hand-held devices).
Collapse
|
15
|
Large-scale cortico-cerebellar computations for horizontal and vertical vergence in humans. Sci Rep 2022; 12:11672. [PMID: 35803967 PMCID: PMC9270479 DOI: 10.1038/s41598-022-15780-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/23/2022] [Accepted: 06/29/2022] [Indexed: 11/12/2022] Open
Abstract
Horizontal and vertical vergence eye movements play a central role in binocular coordination. Neurophysiological studies suggest that cortical and subcortical regions in animals and humans are involved in horizontal vergence. However, little is known about the extent to which the neural mechanism underlying vertical vergence overlaps with that of horizontal vergence. In this study, to explore neural computation for horizontal and vertical vergence, we simultaneously recorded electrooculography (EOG) and whole-head magnetoencephalography (MEG) while presenting large-field stereograms for 29 healthy human adults. The stereograms were designed to produce vergence responses by manipulating horizontal and vertical binocular disparities. A model-based approach was used to assess neural sensitivity to horizontal and vertical disparities via MEG source estimation and the theta-band (4 Hz) coherence between brain activity and EOG vergence velocity. We found similar time-locked neural responses to horizontal and vertical disparity in cortical and cerebellar areas at around 100–250 ms after stimulus onset. In contrast, the low-frequency oscillatory neural activity associated with the execution of vertical vergence differed from that of horizontal vergence. These findings indicate that horizontal and vertical vergence involve partially shared but distinct computations in large-scale cortico-cerebellar networks.
Collapse
|
16
|
Hartle B, Wilcox LM. Stereoscopic depth constancy for physical objects and their virtual counterparts. J Vis 2022; 22:9. [PMID: 35315875 PMCID: PMC8944385 DOI: 10.1167/jov.22.4.9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Stereopsis plays an important role in depth perception; if so, disparity-defined depth should not vary with distance. However, studies of stereoscopic depth constancy often report systematic distortions in depth judgments over distance, particularly for virtual stimuli. Our aim was to understand how depth estimation is impacted by viewing distance and display-based cue conflicts by replicating physical objects in virtual counterparts. To this end, we measured perceived depth using virtual textured half-cylinders and identical three-dimensional (3D) printed versions at two viewing distances under monocular and binocular conditions. Virtual stimuli were viewed using a mirror stereoscope and an Oculus Rift head-mounted display (HMD), while physical stimuli were viewed in a controlled test environment. Depth judgments were similar in both virtual apparatuses, which suggests that variations in the viewing geometry and optics of the HMD have little impact on perceived depth. When viewing physical stimuli binocularly, judgments were accurate and exhibited stereoscopic depth constancy. However, in all cases, depth was underestimated for virtual stimuli and failed to achieve depth constancy. It is clear that depth constancy is only complete for cue-rich physical stimuli and that the failure of constancy in virtual stimuli is due to the presence of the vergence-accommodation conflict. Further, our post hoc analysis revealed that prior experience with virtual and physical environments had a strong effect on depth judgments. That is, performance in virtual environments was enhanced by limited exposure to a related task using physical objects.
Collapse
Affiliation(s)
- Brittney Hartle
- Department of Psychology and Centre for Vision Research, York University, Toronto, Ontario, Canada.,
| | - Laurie M Wilcox
- Department of Psychology and Centre for Vision Research, York University, Toronto, Ontario, Canada.,
| |
Collapse
|
17
|
Gayet S, Peelen MV. Preparatory attention incorporates contextual expectations. Curr Biol 2021; 32:687-692.e6. [PMID: 34919809 DOI: 10.1016/j.cub.2021.11.062] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/24/2021] [Revised: 10/18/2021] [Accepted: 11/25/2021] [Indexed: 10/19/2022]
Abstract
Humans are remarkably proficient at finding objects within complex visual scenes. According to current theories of attention,1-3 visual processing of an object of interest is favored through the preparatory activation of object-specific representations in visual cortex.4-15 One key problem that is inherent to real-world visual search but is not accounted for by current theories is that a given object will produce a dramatically different retinal image depending on its location, which is unknown in advance. For instance, the color of the retinal image depends on the illumination on the object, its shape depends on the viewpoint, and (most critically) its size can vary by several orders of magnitude, depending on the distance to the observer. In order to benefit search, preparatory activity thus needs to incorporate contextual expectations. In the current study, we measured fMRI blood-oxygen-level-dependent (BOLD) activity in human observers while they prepared to search for objects at different distances in indoor-scene photographs. First, we established that observers instantiated preparatory object representations: activity patterns in object-selective cortex evoked during search preparation (while no objects were presented) resembled activity patterns evoked by viewing those objects in isolation. Second, we demonstrated that these preparatory object representations were systematically modulated by expectations derived from scene context: activity patterns reflected the predicted retinal image of the object at each distance (i.e., distant search evoking smaller object representations and nearby search evoking larger object representations). These findings reconcile current theories of attentional selection with the challenges of real-world vision.
Collapse
Affiliation(s)
- Surya Gayet
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, 6525 GD Nijmegen, the Netherlands; Helmholtz Institute, Experimental Psychology, Utrecht University, 3584 CS Utrecht, the Netherlands.
| | - Marius V Peelen
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, 6525 GD Nijmegen, the Netherlands
| |
Collapse
|
18
|
Binocular vision and the control of foot placement during walking in natural terrain. Sci Rep 2021; 11:20881. [PMID: 34686759 PMCID: PMC8536664 DOI: 10.1038/s41598-021-99846-0] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2021] [Accepted: 09/27/2021] [Indexed: 11/19/2022] Open
Abstract
Coordination between visual and motor processes is critical for the selection of stable footholds when walking in uneven terrains. While recent work (Matthis et al. in Curr Biol 8(28):1224–1233, 2018) demonstrates a tight link between gaze (visual) and gait (motor), it remains unclear which aspects of visual information play a role in this visuomotor control loop, and how the loss of this information affects that relationship. Here we examine the role of binocular information in the visuomotor control of walking over complex terrain. We recorded eye and body movements while normally-sighted participants walked over terrains of varying difficulty, with intact vision or with vision in one eye blurred to disrupt binocular vision. Gaze strategy was highly sensitive to the complexity of the terrain, with more fixations dedicated to foothold selection as the terrain became more difficult. The primary effect of increased sensory uncertainty due to disrupted binocular vision was a small bias in gaze towards closer footholds, indicating greater pressure on the visuomotor control process. Participants with binocular vision losses due to developmental disorders (i.e., amblyopia, strabismus), who have had the opportunity to develop alternative strategies, also biased their gaze towards closer footholds. Across all participants, we observed a relationship between an individual’s typical level of binocular visual function and the degree to which gaze is shifted toward the body. Thus the gaze–gait relationship is sensitive to the level of sensory uncertainty, and deficits in binocular visual function (whether transient or long-standing) have systematic effects on gaze strategy in complex terrains. We conclude that binocular vision provides useful information for locating footholds during locomotion. Furthermore, we have demonstrated that combined eye/body tracking in natural environments can be used to provide a more detailed understanding of the impact of a type of vision loss on the visuomotor control process of walking, a vital everyday task.
Collapse
|
19
|
Candy TR, Cormack LK. Recent understanding of binocular vision in the natural environment with clinical implications. Prog Retin Eye Res 2021; 88:101014. [PMID: 34624515 PMCID: PMC8983798 DOI: 10.1016/j.preteyeres.2021.101014] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2021] [Revised: 09/26/2021] [Accepted: 09/29/2021] [Indexed: 10/20/2022]
Abstract
Technological advances in recent decades have allowed us to measure both the information available to the visual system in the natural environment and the rich array of behaviors that the visual system supports. This review highlights the tasks undertaken by the binocular visual system in particular and how, for much of human activity, these tasks differ from those considered when an observer fixates a static target on the midline. The everyday motor and perceptual challenges involved in generating a stable, useful binocular percept of the environment are discussed, together with how these challenges are but minimally addressed by much of current clinical interpretation of binocular function. The implications for new technology, such as virtual reality, are also highlighted in terms of clinical and basic research application.
Collapse
Affiliation(s)
- T Rowan Candy
- School of Optometry, Programs in Vision Science, Neuroscience and Cognitive Science, Indiana University, 800 East Atwater Avenue, Bloomington, IN, 47405, USA.
| | - Lawrence K Cormack
- Department of Psychology, Institute for Neuroscience, and Center for Perceptual Systems, The University of Texas at Austin, Austin, TX, 78712, USA.
| |
Collapse
|
20
|
Reiniger JL, Domdei N, Holz FG, Harmening WM. Human gaze is systematically offset from the center of cone topography. Curr Biol 2021; 31:4188-4193.e3. [PMID: 34343479 DOI: 10.1016/j.cub.2021.07.005] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2021] [Revised: 05/28/2021] [Accepted: 07/02/2021] [Indexed: 10/20/2022]
Abstract
The small physical depression of the human retina, the fovea, is the retinal locus of prime visual resolution, achieved by a peaking topography of the light-sensitive cone photoreceptor outer segments1-3 and a post-receptor wiring scheme preserving high-density sampling.4,5 Humans dynamically direct their gaze such that the retinal images of objects of interest fall onto the foveola, the central one-degree diameter of the fovea,6-8 but it is yet unclear whether a relationship between the individual photoreceptor topography at this location and visual fixation behavior exists.9,10 By employing adaptive optics in vivo imaging and micro-stimulation,11-13 we created topographical maps of the complete foveolar cone mosaics in both eyes of 20 healthy participants while simultaneously recording the retinal location of a fixated visual object in a psychophysical experiment with cellular resolution. We found that the locus of fixation was systematically shifted away from the topographical center toward a naso-superior quadrant on the retina, about 5 min of arc of visual angle on average, with a mirror symmetrical trend between fellow eyes. In cyclopean view, the topographical centers were superior to the fixated target, corresponding to areas in the visual field usually more distant14,15 and thus containing higher spatial frequencies. Given the large variability in foveal topography between individuals, and the surprising precision with which fixation is repeatedly directed to just a small bouquet of cones in the foveola, these findings demonstrate a finely tuned, functionally relevant link between the development of the cellular mosaic of photoreceptors and visual behavior.
Collapse
Affiliation(s)
- Jenny L Reiniger
- Rheinische Friedrich-Wilhelms-Universität Bonn, Department of Ophthalmology, Ernst-Abbe-Str. 2, Bonn 53127, Germany
| | - Niklas Domdei
- Rheinische Friedrich-Wilhelms-Universität Bonn, Department of Ophthalmology, Ernst-Abbe-Str. 2, Bonn 53127, Germany
| | - Frank G Holz
- Rheinische Friedrich-Wilhelms-Universität Bonn, Department of Ophthalmology, Ernst-Abbe-Str. 2, Bonn 53127, Germany
| | - Wolf M Harmening
- Rheinische Friedrich-Wilhelms-Universität Bonn, Department of Ophthalmology, Ernst-Abbe-Str. 2, Bonn 53127, Germany.
| |
Collapse
|
21
|
Crossed-uncrossed projections from primate retina are adapted to disparities of natural scenes. Proc Natl Acad Sci U S A 2021; 118:2015651118. [PMID: 33574061 DOI: 10.1073/pnas.2015651118] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/26/2023] Open
Abstract
In mammals with frontal eyes, optic-nerve fibers from nasal retina project to the contralateral hemisphere of the brain, and fibers from temporal retina project ipsilaterally. The division between crossed and uncrossed projections occurs at or near the vertical meridian. If the division was precise, a problem would arise. Small objects near midline, but nearer or farther than current fixation, would produce signals that travel to opposite hemispheres, making the binocular disparity of those objects difficult to compute. However, in species that have been studied, the division is not precise. Rather, there are overlapping crossed and uncrossed projections such that some fibers from nasal retina project ipsilaterally as well as contralaterally and some from temporal retina project contralaterally as well as ipsilaterally. This increases the probability that signals from an object near vertical midline travel to the same hemisphere, thereby aiding disparity estimation. We investigated whether there is a deficit in binocular vision near the vertical meridian in humans and found no evidence for one. We also investigated the effectiveness of the observed decussation pattern, quantified from anatomical data in monkeys and humans. We used measurements of naturally occurring disparities in humans to determine disparity distributions across the visual field. We then used those distributions to calculate the probability of natural disparities transmitting to the same hemisphere, thereby aiding disparity computation. We found that the pattern of overlapping projections is quite effective. Thus, crossed and uncrossed projections from the retinas are well designed for aiding disparity estimation and stereopsis.
Collapse
|
22
|
Gibaldi A, Labhishetty V, Thibos LN, Banks MS. The blur horopter: Retinal conjugate surface in binocular viewing. J Vis 2021; 21:8. [PMID: 33661280 PMCID: PMC7938023 DOI: 10.1167/jov.21.3.8] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
From measurements of wavefront aberrations in 16 emmetropic eyes, we calculated where objects in the world create best-focused images across the central 27∘ (diameter) of the retina. This is the retinal conjugate surface. We calculated how the surface changes as the eye accommodates from near to far and found that it mostly maintains its shape. The conjugate surface is pitched top-back, meaning that the upper visual field is relatively hyperopic compared to the lower field. We extended the measurements of best image quality into the binocular domain by considering how the retinal conjugate surfaces for the two eyes overlap in binocular viewing. We call this binocular extension the blur horopter. We show that in combining the two images with possibly different sharpness, the visual system creates a larger depth of field of apparently sharp images than occurs with monocular viewing. We examined similarities between the blur horopter and its analog in binocular vision: the binocular horopter. We compared these horopters to the statistics of the natural visual environment. The binocular horopter and scene statistics are strikingly similar. The blur horopter and natural statistics are qualitatively, but not quantitatively, similar. Finally, we used the measurements to refine what is commonly referred to as the zone of clear single binocular vision.
Collapse
Affiliation(s)
- Agostino Gibaldi
- School of Optometry, University of California at Berkeley, Berkeley, CA, USA., https://vision.berkeley.edu/people/agostino-gibaldi-phd
| | - Vivek Labhishetty
- School of Optometry, University of California at Berkeley, Berkeley, CA, USA., https://vision.berkeley.edu/people/vivek-labhishetty-phd/
| | - Larry N Thibos
- School of Optometry, Indiana University, Bloomington, IN, USA, USA., https://optometry.iu.edu/people-directory/thibos-larry.html
| | - Martin S Banks
- School of Optometry, Vision Science Program, University of California at Berkeley, Berkeley, CA, USA., http://bankslab.berkeley.edu/
| |
Collapse
|
23
|
Aghasi A, Heshmat B, Wei L, Tian M. Optimal allocation of quantized human eye depth perception for multi-focal 3D display design. OPTICS EXPRESS 2021; 29:9878-9896. [PMID: 33820153 DOI: 10.1364/oe.412373] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/12/2020] [Accepted: 02/02/2021] [Indexed: 06/12/2023]
Abstract
Creating immersive 3D stereoscopic, autostereoscopic, and lightfield experiences are becoming the center point of optical design of future head mounted displays and lightfield displays. However, despite the advancement in 3D and light field displays, there is no consensus on what are the necessary quantized depth levels for such emerging displays at stereoscopic or monocular modalities. Here we start from psychophysical theories and work toward defining and prioritizing quantized levels of depth that would saturate the human depth perception. We propose a general optimization framework, which locates the depth levels in a globally optimal way for band limited displays. While the original problem is computationally intractable, we manage to find a tractable reformulation as maximally covering a region of interest with a selection of hypographs corresponding to the monocular depth of field profiles. The results indicate that on average 1731 stereoscopic and 7 monocular depth levels (distributed optimally from 25 cm to infinity) would saturate the visual depth perception. Such that adding further depth levels adds negligible improvement. Also the first 3 depth levels should be allocated at (148), then (83, 170), then (53, 90, 170) distances respectively from the face plane to minimize the monocular error in the entire population. The study further discusses the 3D spatial profile of the quantized stereoscopic and monocular depth levels. The study provides fundamental guidelines for designing optimal near eye displays, light-field monitors, and 3D screens.
Collapse
|
24
|
Abstract
Saccades are rapid ballistic eye movements that humans make to direct the fovea to an object of interest. Their kinematics is well defined, showing regular relationships between amplitude, duration, and velocity: the saccadic 'main sequence'. Deviations of eye movements from the main sequence can be used as markers of specific neurological disorders. Despite its significance, there is no general methodological consensus for reliable and repeatable measurements of the main sequence. In this work, we propose a novel approach for standard indicators of oculomotor performance. The obtained measurements are characterized by high repeatability, allowing for fine assessments of inter- and intra-subject variability, and inter-ocular differences. The designed experimental procedure is natural and non-fatiguing, thus it is well suited for fragile or non-collaborative subjects like neurological patients and infants. The method has been released as a software toolbox for public use. This framework lays the foundation for a normative dataset of healthy oculomotor performance for the assessment of oculomotor dysfunctions.
Collapse
Affiliation(s)
- Agostino Gibaldi
- School of Optometry and Vision Science, University of California at Berkeley, 380, Minor Lane, CA, USA.
| | - Silvio P Sabatini
- Department of Informatics, Bioengineering, Robotics and System Engineering, University of Genoa, Via All'Opera Pia, 13, Genoa, 16145, Italy
| |
Collapse
|
25
|
Stein N. A Comparison of Eye Tracking Latencies Among Several Commercial Head-Mounted Displays. Iperception 2021; 12:2041669520983338. [PMID: 33628410 PMCID: PMC7883159 DOI: 10.1177/2041669520983338] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/01/2020] [Accepted: 11/16/2020] [Indexed: 11/15/2022] Open
Abstract
A number of virtual reality head-mounted displays (HMDs) with integrated eye trackers have recently become commercially available. If their eye tracking latency is low and reliable enough for gaze-contingent rendering, this may open up many interesting opportunities for researchers. We measured eye tracking latencies for the Fove-0, the Varjo VR-1, and the High Tech Computer Corporation (HTC) Vive Pro Eye using simultaneous electrooculography measurements. We determined the time from the occurrence of an eye position change to its availability as a data sample from the eye tracker (delay) and the time from an eye position change to the earliest possible change of the display content (latency). For each test and each device, participants performed 60 saccades between two targets 20° of visual angle apart. The targets were continuously visible in the HMD, and the saccades were instructed by an auditory cue. Data collection and eye tracking calibration were done using the recommended scripts for each device in Unity3D. The Vive Pro Eye was recorded twice, once using the SteamVR SDK and once using the Tobii XR SDK. Our results show clear differences between the HMDs. Delays ranged from 15 ms to 52 ms, and the latencies ranged from 45 ms to 81 ms. The Fove-0 appears to be the fastest device and best suited for gaze-contingent rendering.
Collapse
Affiliation(s)
- Niklas Stein
- Institute for Psychology, University of Muenster, Muenster, Germany
| |
Collapse
|
26
|
Turski J. A Geometric Theory Integrating Human Binocular Vision With Eye Movement. Front Neurosci 2020; 14:555965. [PMID: 33364918 PMCID: PMC7750472 DOI: 10.3389/fnins.2020.555965] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2020] [Accepted: 09/18/2020] [Indexed: 11/25/2022] Open
Abstract
A theory of the binocular system with asymmetric eyes (AEs) is developed in the framework of bicentric perspective projections. The AE accounts for the eyeball's global asymmetry produced by the foveal displacement from the posterior pole, the main source of the eye's optical aberrations, and the crystalline lens' tilt countering some of these aberrations. In this theory, the horopter curves, which specify retinal correspondence of binocular single vision, are conic sections resembling empirical horopters. This advances the classic model of empirical horopters as conic sections introduced in an ad hoc way by Ogle in 1932. In contrast to Ogle's theory, here, anatomically supported horopteric conics vary with the AEs' position in the visual plane of bifoveal fixations and their transformations are visualized in a computer simulation. Integrating horopteric conics with eye movements can help design algorithms for maintaining a stable perceptual world from visual information captured by a mobile robot's camera head. Further, this paper proposes a neurophysiologically meaningful definition for the eyes' primary position, a concept which has remained elusive despite its theoretical importance to oculomotor research. Finally, because the horopteric conic's shape is dependent on the AE's parameters, this theory allows for changes in retinal correspondence, which is usually considered preformed and stable.
Collapse
|
27
|
Neural control of rapid binocular eye movements: Saccade-vergence burst neurons. Proc Natl Acad Sci U S A 2020; 117:29123-29132. [PMID: 33139553 DOI: 10.1073/pnas.2015318117] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/20/2022] Open
Abstract
During normal viewing, we direct our eyes between objects in three-dimensional (3D) space many times a minute. To accurately fixate these objects, which are usually located in different directions and at different distances, we must generate eye movements with appropriate versional and vergence components. These combined saccade-vergence eye movements result in disjunctive saccades with a vergence component that is much faster than that generated during smooth, symmetric vergence eye movements. The neural control of disjunctive saccades is still poorly understood. Recent anatomical studies suggested that the central mesencephalic reticular formation (cMRF), located lateral to the oculomotor nucleus, contains premotor neurons potentially involved in the neural control of these eye movements. We have therefore investigated the role of the cMRF in the control of disjunctive saccades in trained rhesus monkeys. Here, we describe a unique population of cMRF neurons that, during disjunctive saccades, display a burst of spikes that are highly correlated with vergence velocity. Importantly, these neurons show no increase in activity for either conjugate saccades or symmetric vergence. These neurons are termed saccade-vergence burst neurons (SVBNs) to maintain consistency with modeling studies that proposed that such a class of neuron exists to generate the enhanced vergence velocities observed during disjunctive saccades. Our results demonstrate the existence and characteristics of SVBNs whose activity is correlated solely with the vergence component of disjunctive saccades and, based on modeling studies, are critically involved in the generation of the disjunctive saccades required to view objects in our 3D world.
Collapse
|
28
|
Chauhan T, Héjja-Brichard Y, Cottereau BR. Modelling binocular disparity processing from statistics in natural scenes. Vision Res 2020; 176:27-39. [DOI: 10.1016/j.visres.2020.07.009] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2019] [Revised: 07/19/2020] [Accepted: 07/20/2020] [Indexed: 11/25/2022]
|
29
|
A dichoptic feedback-based oculomotor training method to manipulate interocular alignment. Sci Rep 2020; 10:15634. [PMID: 32973252 PMCID: PMC7515870 DOI: 10.1038/s41598-020-72561-y] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/11/2020] [Accepted: 08/24/2020] [Indexed: 11/15/2022] Open
Abstract
Strabismus is a prevalent impairment of binocular alignment that is associated with a spectrum of perceptual deficits and social disadvantages. Current treatments for strabismus involve ocular alignment through surgical or optical methods and may include vision therapy exercises. In the present study, we explore the potential of real-time dichoptic visual feedback that may be used to quantify and manipulate interocular alignment. A gaze-contingent ring was presented independently to each eye of 11 normally-sighted observers as they fixated a target dot presented only to their dominant eye. Their task was to center the rings within 2° of the target for at least 1 s, with feedback provided by the sizes of the rings. By offsetting the ring in the non-dominant eye temporally or nasally, this task required convergence or divergence, respectively, of the non-dominant eye. Eight of 11 observers attained 5° asymmetric convergence and 3 of 11 attained 3° asymmetric divergence. The results suggest that real-time gaze-contingent feedback may be used to quantify and transiently simulate strabismus and holds promise as a method to augment existing therapies for oculomotor alignment disorders.
Collapse
|
30
|
Kim S, Burge J. Natural scene statistics predict how humans pool information across space in surface tilt estimation. PLoS Comput Biol 2020; 16:e1007947. [PMID: 32579559 PMCID: PMC7340327 DOI: 10.1371/journal.pcbi.1007947] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2019] [Revised: 07/07/2020] [Accepted: 05/14/2020] [Indexed: 11/30/2022] Open
Abstract
Visual systems estimate the three-dimensional (3D) structure of scenes from information in two-dimensional (2D) retinal images. Visual systems use multiple sources of information to improve the accuracy of these estimates, including statistical knowledge of the probable spatial arrangements of natural scenes. Here, we examine how 3D surface tilts are spatially related in real-world scenes, and show that humans pool information across space when estimating surface tilt in accordance with these spatial relationships. We develop a hierarchical model of surface tilt estimation that is grounded in the statistics of tilt in natural scenes and images. The model computes a global tilt estimate by pooling local tilt estimates within an adaptive spatial neighborhood. The spatial neighborhood in which local estimates are pooled changes according to the value of the local estimate at a target location. The hierarchical model provides more accurate estimates of groundtruth tilt in natural scenes and provides a better account of human performance than the local estimates. Taken together, the results imply that the human visual system pools information about surface tilt across space in accordance with natural scene statistics. Visual systems estimate three-dimensional (3D) properties of scenes from two-dimensional images on the retinas. To solve this difficult problem as accurately as possible, visual systems use many available sources of information, including information about how the 3D properties of the world are spatially arranged. This manuscript reports a systematic analysis of 3D surface tilt in natural scenes, a model of surface tilt estimation that makes use of these scene statistics, and human psychophysical data on the estimation of surface tilt from natural images. The results show that the regularities present in the natural environment predict both how to maximize the accuracy of tilt estimation and how to maximize the prediction of human performance. This work contributes to a growing line of work that establishes links between rigorous measurements of natural scenes and the function of sensory and perceptual systems.
Collapse
Affiliation(s)
- Seha Kim
- Department of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America
- * E-mail:
| | - Johannes Burge
- Department of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America
- Neuroscience Graduate Group, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America
- Bioengineering Graduate Group, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America
| |
Collapse
|
31
|
Near-optimal combination of disparity across a log-polar scaled visual field. PLoS Comput Biol 2020; 16:e1007699. [PMID: 32275711 PMCID: PMC7176150 DOI: 10.1371/journal.pcbi.1007699] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/26/2019] [Revised: 04/22/2020] [Accepted: 01/30/2020] [Indexed: 01/06/2023] Open
Abstract
The human visual system is foveated: we can see fine spatial details in central vision, whereas resolution is poor in our peripheral visual field, and this loss of resolution follows an approximately logarithmic decrease. Additionally, our brain organizes visual input in polar coordinates. Therefore, the image projection occurring between retina and primary visual cortex can be mathematically described by the log-polar transform. Here, we test and model how this space-variant visual processing affects how we process binocular disparity, a key component of human depth perception. We observe that the fovea preferentially processes disparities at fine spatial scales, whereas the visual periphery is tuned for coarse spatial scales, in line with the naturally occurring distributions of depths and disparities in the real-world. We further show that the visual system integrates disparity information across the visual field, in a near-optimal fashion. We develop a foveated, log-polar model that mimics the processing of depth information in primary visual cortex and that can process disparity directly in the cortical domain representation. This model takes real images as input and recreates the observed topography of human disparity sensitivity. Our findings support the notion that our foveated, binocular visual system has been moulded by the statistics of our visual environment. We investigate how humans perceive depth from binocular disparity at different spatial scales and across different regions of the visual field. We show that small changes in disparity-defined depth are detected best in central vision, whereas peripheral vision best captures the coarser structure of the environment. We also demonstrate that depth information extracted from different regions of the visual field is combined into a unified depth percept. We then construct an image-computable model of disparity processing that takes into account how our brain organizes the visual input at our retinae. The model operates directly in cortical image space, and neatly accounts for human depth perception across the visual field.
Collapse
|
32
|
Chang Y, Zhou H, Wang X, Shen L, Hu T. Cross-Drone Binocular Coordination for Ground Moving Target Tracking in Occlusion-Rich Scenarios. IEEE Robot Autom Lett 2020. [DOI: 10.1109/lra.2020.2975713] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
33
|
Wu H, Li X, Tang Y, Xu Q, Zhang X, Zhou L, Lan W, Zhang B, Yang Z. Optimal Stereoacuity Reveals More Than Critical Time in Patients With Intermittent Exotropia. Front Neurosci 2020; 14:133. [PMID: 32132901 PMCID: PMC7040177 DOI: 10.3389/fnins.2020.00133] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2019] [Accepted: 02/03/2020] [Indexed: 01/06/2023] Open
Abstract
Synopsis Both optimal stereoacuity and integration time to achieve that are impaired in patients with intermittent exotropia. The deterioration of stereoacuity is more revealing since it correlates well with exotropia control score. Background Despite the periodic misalignment of two eyes, some intermittent exotropia (IXT) patients exhibit normal stereoacuity, particularly when evaluated with static tests. It is not clear if the temporal integration process of stereopsis is altered in IXT patients, thus warranting further research. Methods IXT patients (n = 29) and age-matched normal controls (n = 36) were recruited. Static stereopsis was measured with the Titmus stereoacuity test. In computer-generated random dots tests, stereoacuity was measured with a stimuli presentation duration varying from 100 to 1,200 ms. And the relationship between stereoacuity and stimuli duration was fitted into a quadratic model. Optimal stereoacuity was achieved when fitted curve flattened and the critical integration time was the duration needed to achieve optimal stereoacuity. Results IXT patients were not found to differ significantly from control subjects under the Titmus test, while the Random Dots stereotest showed significantly worse optimal stereoacuity and significantly longer critical integration time. Multiple regression analysis showed that age (R = −4.83; P = 0.04) had statistically significant negative correlation on the critical integration time, age (R = −6.45; P = 0.047) and exotropia control scores (R = 60.71; P = 0.007) had statistically significant effects on optimal stereoacuity. Conclusion The temporal integration for stereopsis is impaired in IXT patients, requiring longer critical integration time to achieve elevated optimal stereoacuity.
Collapse
Affiliation(s)
- Haoran Wu
- Aier School of Ophthalmology, Central South University, Changsha, China.,Aier Institute of Optometry and Vision Science, Changsha, China
| | - Xiaoning Li
- Aier Institute of Optometry and Vision Science, Changsha, China.,Aier School of Optometry and Vision Science, Hubei University of Science and Technology, Xianning, China
| | - Yao Tang
- Aier Institute of Optometry and Vision Science, Changsha, China
| | - Qinglin Xu
- Aier Institute of Optometry and Vision Science, Changsha, China
| | - Xuhong Zhang
- College of Medicine, Zhejiang University, Hangzhou, China
| | - Lu Zhou
- Department of Ophthalmology, Children's Hospital of Nanjing Medical University, Nanjing, China
| | - Weizhong Lan
- Aier School of Ophthalmology, Central South University, Changsha, China.,Aier School of Optometry and Vision Science, Hubei University of Science and Technology, Xianning, China
| | - Bin Zhang
- College of Optometry, Nova Southeastern University, Davie, FL, United States
| | - Zhikuan Yang
- Aier School of Ophthalmology, Central South University, Changsha, China.,Aier School of Optometry and Vision Science, Hubei University of Science and Technology, Xianning, China
| |
Collapse
|
34
|
Mice Discriminate Stereoscopic Surfaces Without Fixating in Depth. J Neurosci 2019; 39:8024-8037. [PMID: 31462533 DOI: 10.1523/jneurosci.0895-19.2019] [Citation(s) in RCA: 23] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/18/2019] [Revised: 07/26/2019] [Accepted: 07/30/2019] [Indexed: 11/21/2022] Open
Abstract
Stereopsis is a ubiquitous feature of primate mammalian vision, but little is known about if and how rodents such as mice use stereoscopic vision. We used random dot stereograms to test for stereopsis in male and female mice, and they were able to discriminate near from far surfaces over a range of disparities, with diminishing performance for small and large binocular disparities. Based on two-photon measurements of disparity tuning, the range of disparities represented in the visual cortex aligns with the behavior and covers a broad range of disparities. When we examined their binocular eye movements, we found that, unlike primates, mice did not systematically vary relative eye positions or use vergence eye movements when presented with different disparities. Nonetheless, the representation of disparity tuning was wide enough to capture stereoscopic information over a range of potential vergence angles. Although mice share fundamental characteristics of stereoscopic vision with primates and carnivores, their lack of disparity-dependent vergence eye movements and wide neuronal representation suggests that they may use a distinct strategy for stereopsis.SIGNIFICANCE STATEMENT Binocular vision allows us to derive depth information by comparing right and left eye information. We characterized binocular integration in mice because tools exist in these animals to dissect the underlying neural circuitry for binocular vision. Using random dot stereograms, we find that behavior and disparity tuning in the visual cortex share fundamental characteristics with primates, but we did not observe any evidence of disparity-dependent changes in vergence angle. We propose that mice use a distinct strategy of stereopsis compared with primates by using a broad range of disparities to encode depth over a large field of view and to compensate for nonstereoscopic changes in vergence angle that arise during natural behavior.
Collapse
|