1
|
Segraves MA. Using Natural Scenes to Enhance our Understanding of the Cerebral Cortex's Role in Visual Search. Annu Rev Vis Sci 2023; 9:435-454. [PMID: 37164028 DOI: 10.1146/annurev-vision-100720-124033] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/12/2023]
Abstract
Using natural scenes is an approach to studying the visual and eye movement systems approximating how these systems function in everyday life. This review examines the results from behavioral and neurophysiological studies using natural scene viewing in humans and monkeys. The use of natural scenes for the study of cerebral cortical activity is relatively new and presents challenges for data analysis. Methods and results from the use of natural scenes for the study of the visual and eye movement cortex are presented, with emphasis on new insights that this method provides enhancing what is known about these cortical regions from the use of conventional methods.
Collapse
Affiliation(s)
- Mark A Segraves
- Department of Neurobiology, Northwestern University, Evanston, Illinois, USA;
| |
Collapse
|
2
|
Dimension of visual information interacts with working memory in monkeys and humans. Sci Rep 2022; 12:5335. [PMID: 35351948 PMCID: PMC8964748 DOI: 10.1038/s41598-022-09367-7] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2021] [Accepted: 03/16/2022] [Indexed: 12/16/2022] Open
Abstract
Humans demonstrate behavioural advantages (biases) towards particular dimensions (colour or shape of visual objects), but such biases are significantly altered in neuropsychological disorders. Recent studies have shown that lesions in the prefrontal cortex do not abolish dimensional biases, and therefore suggest that such biases might not depend on top-down prefrontal-mediated attention and instead emerge as bottom-up processing advantages. We hypothesised that if dimensional biases merely emerge from an enhancement of object features, the presence of visual objects would be necessary for the manifestation of dimensional biases. In a specifically-designed working memory task, in which macaque monkeys and humans performed matching based on the object memory rather than the actual object, we found significant dimensional biases in both species, which appeared as a shorter response time and higher accuracy in the preferred dimension (colour and shape dimension in humans and monkeys, respectively). Moreover, the mnemonic demands of the task influenced the magnitude of dimensional bias. Our findings in two primate species indicate that the dichotomy of top-down and bottom-up processing does not fully explain the emergence of dimensional biases. Instead, dimensional biases may emerge when processed information regarding visual object features interact with mnemonic and executive functions to guide goal-directed behaviour.
Collapse
|
3
|
Unsupervised foveal vision neural architecture with top-down attention. Neural Netw 2021; 141:145-159. [PMID: 33901879 DOI: 10.1016/j.neunet.2021.03.003] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2020] [Revised: 02/23/2021] [Accepted: 03/02/2021] [Indexed: 11/20/2022]
Abstract
Deep learning architectures are an extremely powerful tool for recognizing and classifying images. However, they require supervised learning and normally work on vectors of the size of image pixels and produce the best results when trained on millions of object images. To help mitigate these issues, we propose an end-to-end architecture that fuses bottom-up saliency and top-down attention with an object recognition module to focus on relevant data and learn important features that can later be fine-tuned for a specific task, employing only unsupervised learning. In addition, by utilizing a virtual fovea that focuses on relevant portions of the data, the training speed can be greatly improved. We test the performance of the proposed Gamma saliency technique on the Toronto and CAT 2000 databases, and the foveated vision in the large Street View House Numbers (SVHN) database. The results with foveated vision show that Gamma saliency performs at the same level as the best alternative algorithms while being computationally faster. The results in SVHN show that our unsupervised cognitive architecture is comparable to fully supervised methods and that saliency also improves CNN performance if desired. Finally, we develop and test a top-down attention mechanism based on the Gamma saliency applied to the top layer of CNNs to facilitate scene understanding in multi-object cluttered images. We show that the extra information from top-down saliency is capable of speeding up the extraction of digits in the cluttered multidigit MNIST data set, corroborating the important role of top down attention.
Collapse
|
4
|
Mansouri FA, Buckley MJ, Fehring DJ, Tanaka K. The Role of Primate Prefrontal Cortex in Bias and Shift Between Visual Dimensions. Cereb Cortex 2021; 30:85-99. [PMID: 31220222 PMCID: PMC7029686 DOI: 10.1093/cercor/bhz072] [Citation(s) in RCA: 21] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2018] [Revised: 03/11/2019] [Accepted: 03/11/2019] [Indexed: 12/23/2022] Open
Abstract
Imaging and neural activity recording studies have shown activation in the primate prefrontal cortex when shifting attention between visual dimensions is necessary to achieve goals. A fundamental unanswered question is whether representations of these dimensions emerge from top-down attentional processes mediated by prefrontal regions or from bottom-up processes within visual cortical regions. We hypothesized a causative link between prefrontal cortical regions and dimension-based behavior. In large cohorts of humans and macaque monkeys, performing the same attention shifting task, we found that both species successfully shifted between visual dimensions, but both species also showed a significant behavioral advantage/bias to a particular dimension; however, these biases were in opposite directions in humans (bias to color) versus monkeys (bias to shape). Monkeys' bias remained after selective bilateral lesions within the anterior cingulate cortex (ACC), frontopolar cortex, dorsolateral prefrontal cortex (DLPFC), orbitofrontal cortex (OFC), or superior, lateral prefrontal cortex. However, lesions within certain regions (ACC, DLPFC, or OFC) impaired monkeys' ability to shift between these dimensions. We conclude that goal-directed processing of a particular dimension for the executive control of behavior depends on the integrity of prefrontal cortex; however, representation of competing dimensions and bias toward them does not depend on top-down prefrontal-mediated processes.
Collapse
Affiliation(s)
- Farshad A Mansouri
- Cognitive Neuroscience Laboratory, Department of Physiology, Monash Biomedicine Discovery Institute, Monash University, Victoria, Australia.,ARC Centre of Excellence for Integrative Brain Function, Monash University, Victoria, Australia
| | - Mark J Buckley
- Department of Experimental Psychology, Oxford University, Oxford, UK
| | - Daniel J Fehring
- Cognitive Neuroscience Laboratory, Department of Physiology, Monash Biomedicine Discovery Institute, Monash University, Victoria, Australia.,ARC Centre of Excellence for Integrative Brain Function, Monash University, Victoria, Australia
| | - Keiji Tanaka
- Cognitive Brain Mapping Laboratory, RIKEN Center for Brain Science, Wako, Saitama, Japan
| |
Collapse
|
5
|
Dimigen O, Ehinger BV. Regression-based analysis of combined EEG and eye-tracking data: Theory and applications. J Vis 2021; 21:3. [PMID: 33410892 PMCID: PMC7804566 DOI: 10.1167/jov.21.1.3] [Citation(s) in RCA: 34] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2019] [Accepted: 08/14/2020] [Indexed: 12/27/2022] Open
Abstract
Fixation-related potentials (FRPs), neural responses aligned to the end of saccades, are a promising tool for studying the dynamics of attention and cognition under natural viewing conditions. In the past, four methodological problems have complicated the analysis of such combined eye-tracking/electroencephalogram experiments: (1) the synchronization of data streams, (2) the removal of ocular artifacts, (3) the condition-specific temporal overlap between the brain responses evoked by consecutive fixations, and (4) the fact that numerous low-level stimulus and saccade properties also influence the postsaccadic neural responses. Although effective solutions exist for the first two problems, the latter two are only beginning to be addressed. In the current paper, we present and review a unified regression-based framework for FRP analysis that allows us to deconvolve overlapping potentials while also controlling for both linear and nonlinear confounds on the FRP waveform. An open software implementation is provided for all procedures. We then demonstrate the advantages of this proposed (non)linear deconvolution modeling approach for data from three commonly studied paradigms: face perception, scene viewing, and reading. First, for a traditional event-related potential (ERP) face recognition experiment, we show how this technique can separate stimulus ERPs from overlapping muscle and brain potentials produced by small (micro)saccades on the face. Second, in natural scene viewing, we model and isolate multiple nonlinear effects of saccade parameters on the FRP. Finally, for a natural sentence reading experiment using the boundary paradigm, we show how it is possible to study the neural correlates of parafoveal preview after removing spurious overlap effects caused by the associated difference in average fixation time. Our results suggest a principal way of measuring reliable eye movement-related brain activity during natural vision.
Collapse
Affiliation(s)
- Olaf Dimigen
- Department of Psychology, Humboldt-Universität zu Berlin, Berlin, Germany
| | - Benedikt V Ehinger
- Institute of Cognitive Science, Universität Osnabrück, Osnabrück, Germany
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
| |
Collapse
|
6
|
Uejima T, Niebur E, Etienne-Cummings R. Proto-Object Based Saliency Model With Texture Detection Channel. Front Comput Neurosci 2020; 14:541581. [PMID: 33071766 PMCID: PMC7541834 DOI: 10.3389/fncom.2020.541581] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2020] [Accepted: 08/14/2020] [Indexed: 11/13/2022] Open
Abstract
The amount of visual information projected from the retina to the brain exceeds the information processing capacity of the latter. Attention, therefore, functions as a filter to highlight important information at multiple stages of the visual pathway that requires further and more detailed analysis. Among other functions, this determines where to fixate since only the fovea allows for high resolution imaging. Visual saliency modeling, i.e. understanding how the brain selects important information to analyze further and to determine where to fixate next, is an important research topic in computational neuroscience and computer vision. Most existing bottom-up saliency models use low-level features such as intensity and color, while some models employ high-level features, like faces. However, little consideration has been given to mid-level features, such as texture, for visual saliency models. In this paper, we extend a biologically plausible proto-object based saliency model by adding simple texture channels which employ nonlinear operations that mimic the processing performed by primate visual cortex. The extended model shows statistically significant improved performance in predicting human fixations compared to the previous model. Comparing the performance of our model with others on publicly available benchmarking datasets, we find that our biologically plausible model matches the performance of other models, even though those were designed entirely for maximal performance with little regard to biological realism.
Collapse
Affiliation(s)
- Takeshi Uejima
- The Department of Electrical and Computer Engineering, The Johns Hopkins University, Baltimore, MD, United States
| | - Ernst Niebur
- The Solomon Snyder Department of Neuroscience and the Zanvyl Krieger Mind/Brain Institute, The Johns Hopkins University, Baltimore, MD, United States
| | - Ralph Etienne-Cummings
- The Department of Electrical and Computer Engineering, The Johns Hopkins University, Baltimore, MD, United States
| |
Collapse
|
7
|
Maylott SE, Paukner A, Ahn YA, Simpson EA. Human and monkey infant attention to dynamic social and nonsocial stimuli. Dev Psychobiol 2020; 62:841-857. [PMID: 32424813 PMCID: PMC7944642 DOI: 10.1002/dev.21979] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2019] [Revised: 03/23/2020] [Accepted: 03/31/2020] [Indexed: 12/14/2022]
Abstract
The present study explored behavioral norms for infant social attention in typically developing human and nonhuman primate infants. We examined the normative development of attention to dynamic social and nonsocial stimuli longitudinally in macaques (Macaca mulatta) at 1, 3, and 5 months of age (N = 75) and humans at 2, 4, 6, 8, and 13 months of age (N = 69) using eye tracking. All infants viewed concurrently played silent videos-one social video and one nonsocial video. Both macaque and human infants were faster to look to the social than the nonsocial stimulus, and both species grew faster to orient to the social stimulus with age. Further, macaque infants' social attention increased linearly from 1 to 5 months. In contrast, human infants displayed a nonlinear pattern of social interest, with initially greater attention to the social stimulus, followed by a period of greater interest in the nonsocial stimulus, and then a rise in social interest from 6 to 13 months. Overall, human infants looked longer than macaque infants, suggesting humans have more sustained attention in the first year of life. These findings highlight potential species similarities and differences, and reflect a first step in establishing baseline patterns of early social attention development.
Collapse
Affiliation(s)
- Sarah E Maylott
- Department of Psychology, University of Miami, Coral Gables, FL, USA
| | - Annika Paukner
- Department of Psychology, Nottingham Trent University, Nottingham, UK
| | - Yeojin A Ahn
- Department of Psychology, University of Miami, Coral Gables, FL, USA
| | | |
Collapse
|
8
|
Dimigen O. Optimizing the ICA-based removal of ocular EEG artifacts from free viewing experiments. Neuroimage 2019; 207:116117. [PMID: 31689537 DOI: 10.1016/j.neuroimage.2019.116117] [Citation(s) in RCA: 99] [Impact Index Per Article: 16.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2019] [Revised: 07/01/2019] [Accepted: 08/20/2019] [Indexed: 11/30/2022] Open
Abstract
Combining EEG with eye-tracking is a promising approach to study neural correlates of natural vision, but the resulting recordings are also heavily contaminated by activity of the eye balls, eye lids, and extraocular muscles. While Independent Component Analysis (ICA) is commonly used to suppress these ocular artifacts, its performance under free viewing conditions has not been systematically evaluated and many published reports contain residual artifacts. Here I evaluated and optimized ICA-based correction for two tasks with unconstrained eye movements: visual search in images and sentence reading. In a first step, four parameters of the ICA pipeline were varied orthogonally: the (1) high-pass and (2) low-pass filter applied to the training data, (3) the proportion of training data containing myogenic saccadic spike potentials (SP), and (4) the threshold for eye tracker-based component rejection. In a second step, the eye-tracker was used to objectively quantify the correction quality of each ICA solution, both in terms of undercorrection (residual artifacts) and overcorrection (removal of neurogenic activity). As a benchmark, results were compared to those obtained with an alternative spatial filter, Multiple Source Eye Correction (MSEC). With commonly used settings, Infomax ICA not only left artifacts in the data, but also distorted neurogenic activity during eye movement-free intervals. However, correction results could be strongly improved by training the ICA on optimally filtered data in which SPs were massively overweighted. With optimized procedures, ICA removed virtually all artifacts, including the SP and its associated spectral broadband artifact from both viewing paradigms, with little distortion of neural activity. It also outperformed MSEC in terms of SP correction. Matlab code is provided.
Collapse
Affiliation(s)
- Olaf Dimigen
- Humboldt-Universität zu Berlin, Unter den Linden 6, 10099, Berlin, Germany.
| |
Collapse
|
9
|
The comparative anatomy of frontal eye fields in primates. Cortex 2019; 118:51-64. [DOI: 10.1016/j.cortex.2019.02.023] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2018] [Revised: 01/24/2019] [Accepted: 02/22/2019] [Indexed: 12/25/2022]
|
10
|
Yuval-Greenberg S, Keren A, Hilo R, Paz A, Ratzon N. Gaze Control During Simulator Driving in Adolescents With and Without Attention Deficit Hyperactivity Disorder. Am J Occup Ther 2019; 73:7303345030p1-7303345030p8. [DOI: 10.5014/ajot.2019.031500] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
Abstract
Importance: Attention deficit hyperactivity disorder (ADHD) is associated with driving deficits. Visual standards for driving define minimum qualifications for safe driving, including acuity and field of vision, but they do not consider the ability to explore the environment efficiently by shifting the gaze, which is a critical element of safe driving.
Objective: To examine visual exploration during simulated driving in adolescents with and without ADHD.
Design: Adolescents with and without ADHD drove a driving simulator for approximately 10 min while their gaze was monitored. They then completed a battery of questionnaires.
Setting: University lab.
Participants: Participants with (n = 16) and without (n = 15) ADHD were included. Participants had a history of neurological disorders other than ADHD and normal or corrected-to-normal vision. Control participants reported not having a diagnosis of ADHD. Participants with ADHD had been previously diagnosed by a qualified professional.
Outcomes and Measures: We compared the following measures between ADHD and non-ADHD groups: dashboard dwell times, fixation variance, entropy, and fixation duration.
Results: Findings showed that participants with ADHD were more restricted in their patterns of exploration than control group participants. They spent considerably more time gazing at the dashboard and had longer periods of fixation with lower variability and randomness.
Conclusions and Relevance: The results support the hypothesis that adolescents with ADHD engage in less active exploration during simulated driving.
What This Article Adds: This study raises concerns regarding the driving competence of people with ADHD and opens up new directions for potential training programs that focus on exploratory gaze control.
Collapse
Affiliation(s)
- Shlomit Yuval-Greenberg
- Shlomit Yuval-Greenberg, PhD, is Senior Lecturer, School of Psychological Sciences and Sagol School of Neuroscience, Tel Aviv University, Tel Aviv, Israel;
| | - Anat Keren
- Anat Keren, MSc, OT, is Account Manager, Futuring Up, and Occupational Therapist, Private Practice, Tel Aviv, Israel. At the time of the study, she was Master’s Student, Department of Occupational Therapy, Stanley Steyer School of Health Professions, Sackler Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel
| | - Rinat Hilo
- Rinat Hilo, MA, is PhD Student, School of Psychological Sciences, Tel Aviv University, Tel Aviv, Israel
| | - Adar Paz
- Adar Paz, BA, is PhD Student, Department of Psychology, Bar Ilan University. At the time of the study, he was BA Student, School of Psychological Sciences, Tel Aviv University, Tel Aviv, Israel
| | - Navah Ratzon
- Navah Ratzon, PhD, MPH, OT, is Professor, Department of Occupational Therapy, Stanley Steyer School of Health Professions, Sackler Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel
| |
Collapse
|
11
|
Hegdé J. Neural Mechanisms of High-Level Vision. Compr Physiol 2018; 8:903-953. [PMID: 29978891 DOI: 10.1002/cphy.c160035] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
The last three decades have seen major strides in our understanding of neural mechanisms of high-level vision, or visual cognition of the world around us. Vision has also served as a model system for the study of brain function. Several broad insights, as yet incomplete, have recently emerged. First, visual perception is best understood not as an end unto itself, but as a sensory process that subserves the animal's behavioral goal at hand. Visual perception is likely to be simply a side effect that reflects the readout of visual information processing that leads to behavior. Second, the brain is essentially a probabilistic computational system that produces behaviors by collectively evaluating, not necessarily consciously or always optimally, the available information about the outside world received from the senses, the behavioral goals, prior knowledge about the world, and possible risks and benefits of a given behavior. Vision plays a prominent role in the overall functioning of the brain providing the lion's share of information about the outside world. Third, the visual system does not function in isolation, but rather interacts actively and reciprocally with other brain systems, including other sensory faculties. Finally, various regions of the visual system process information not in a strict hierarchical manner, but as parts of various dynamic brain-wide networks, collectively referred to as the "connectome." Thus, a full understanding of vision will ultimately entail understanding, in granular, quantitative detail, various aspects of dynamic brain networks that use visual sensory information to produce behavior under real-world conditions. © 2017 American Physiological Society. Compr Physiol 8:903-953, 2018.
Collapse
Affiliation(s)
- Jay Hegdé
- Brain and Behavior Discovery Institute, Augusta University, Augusta, Georgia, USA.,James and Jean Culver Vision Discovery Institute, Augusta University, Augusta, Georgia, USA.,Department of Ophthalmology, Medical College of Georgia, Augusta University, Augusta, Georgia, USA.,The Graduate School, Augusta University, Augusta, Georgia, USA
| |
Collapse
|
12
|
Kano F, Shepherd SV, Hirata S, Call J. Primate social attention: Species differences and effects of individual experience in humans, great apes, and macaques. PLoS One 2018; 13:e0193283. [PMID: 29474416 PMCID: PMC5825077 DOI: 10.1371/journal.pone.0193283] [Citation(s) in RCA: 21] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2017] [Accepted: 02/07/2018] [Indexed: 11/18/2022] Open
Abstract
When viewing social scenes, humans and nonhuman primates focus on particular features, such as the models' eyes, mouth, and action targets. Previous studies reported that such viewing patterns vary significantly across individuals in humans, and also across closely-related primate species. However, the nature of these individual and species differences remains unclear, particularly among nonhuman primates. In large samples of human and nonhuman primates, we examined species differences and the effects of experience on patterns of gaze toward social movies. Experiment 1 examined the species differences across rhesus macaques, nonhuman apes (bonobos, chimpanzees, and orangutans), and humans while they viewed movies of various animals' species-typical behaviors. We found that each species had distinct viewing patterns of the models' faces, eyes, mouths, and action targets. Experiment 2 tested the effect of individuals' experience on chimpanzee and human viewing patterns. We presented movies depicting natural behaviors of chimpanzees to three groups of chimpanzees (individuals from a zoo, a sanctuary, and a research institute) differing in their early social and physical experiences. We also presented the same movies to human adults and children differing in their expertise with chimpanzees (experts vs. novices) or movie-viewing generally (adults vs. preschoolers). Individuals varied within each species in their patterns of gaze toward models' faces, eyes, mouths, and action targets depending on their unique individual experiences. We thus found that the viewing patterns for social stimuli are both individual- and species-specific in these closely-related primates. Such individual/species-specificities are likely related to both individual experience and species-typical temperament, suggesting that primate individuals acquire their unique attentional biases through both ontogeny and evolution. Such unique attentional biases may help them learn efficiently about their particular social environments.
Collapse
Affiliation(s)
- Fumihiro Kano
- Kumamoto Sanctuary, Wildlife Research Center, Kyoto University, Kumamoto, Japan
| | | | - Satoshi Hirata
- Kumamoto Sanctuary, Wildlife Research Center, Kyoto University, Kumamoto, Japan
| | - Josep Call
- Department of Developmental and Comparative Psychology, Max-Planck Institute for Evolutionary Anthropology, Leipzig, Germany
- School of Psychology and Neuroscience, University of St Andrews, St Andrews, United Kingdom
| |
Collapse
|
13
|
Ito J, Yamane Y, Suzuki M, Maldonado P, Fujita I, Tamura H, Grün S. Switch from ambient to focal processing mode explains the dynamics of free viewing eye movements. Sci Rep 2017; 7:1082. [PMID: 28439075 PMCID: PMC5430715 DOI: 10.1038/s41598-017-01076-w] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2016] [Accepted: 03/22/2017] [Indexed: 11/21/2022] Open
Abstract
Previous studies have reported that humans employ ambient and focal modes of visual exploration while they freely view natural scenes. These two modes have been characterized based on eye movement parameters such as saccade amplitude and fixation duration, but not by any visual features of the viewed scenes. Here we propose a new characterization of eye movements during free viewing based on how eyes are moved from and to objects in a visual scene. We applied this characterization to data obtained from freely-viewing macaque monkeys. We show that the analysis based on this characterization gives a direct indication of a behavioral shift from ambient to focal processing mode along the course of free viewing exploration. We further propose a stochastic model of saccade sequence generation incorporating a switch between the two processing modes, which quantitatively reproduces the behavioral features observed in the data.
Collapse
Affiliation(s)
- Junji Ito
- Institute of Neuroscience and Medicine (INM-6) and Institute for Advanced Simulation (IAS-6) and JARA BRAIN Institute I, Jülich Research Centre, Jülich, Germany.
| | - Yukako Yamane
- Graduate School of Frontier Biosciences, Osaka University, Osaka, Japan
- Center for Information and Neural Networks, Osaka University and National Institute of Information and Communications Technology, Osaka, Japan
| | - Mika Suzuki
- Graduate School of Frontier Biosciences, Osaka University, Osaka, Japan
| | - Pedro Maldonado
- BNI, CENEM and Programa de Fisiología y Biofísica, ICBM, Facultad de Medicina, Universidad de Chile, Santiago, Chile
| | - Ichiro Fujita
- Graduate School of Frontier Biosciences, Osaka University, Osaka, Japan
- Center for Information and Neural Networks, Osaka University and National Institute of Information and Communications Technology, Osaka, Japan
| | - Hiroshi Tamura
- Graduate School of Frontier Biosciences, Osaka University, Osaka, Japan
- Center for Information and Neural Networks, Osaka University and National Institute of Information and Communications Technology, Osaka, Japan
| | - Sonja Grün
- Institute of Neuroscience and Medicine (INM-6) and Institute for Advanced Simulation (IAS-6) and JARA BRAIN Institute I, Jülich Research Centre, Jülich, Germany
- Graduate School of Frontier Biosciences, Osaka University, Osaka, Japan
- Theoretical Systems Neurobiology, RWTH Aachen University, Aachen, Germany
| |
Collapse
|
14
|
Segraves MA, Kuo E, Caddigan S, Berthiaume EA, Kording KP. Predicting rhesus monkey eye movements during natural-image search. J Vis 2017; 17:12. [PMID: 28355625 PMCID: PMC5373813 DOI: 10.1167/17.3.12] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
There are three prominent factors that can predict human visual-search behavior in natural scenes: the distinctiveness of a location (salience), similarity to the target (relevance), and features of the environment that predict where the object might be (context). We do not currently know how well these factors are able to predict macaque visual search, which matters because it is arguably the most popular model for asking how the brain controls eye movements. Here we trained monkeys to perform the pedestrian search task previously used for human subjects. Salience, relevance, and context models were all predictive of monkey eye fixations and jointly about as precise as for humans. We attempted to disrupt the influence of scene context on search by testing the monkeys with an inverted set of the same images. Surprisingly, the monkeys were able to locate the pedestrian at a rate similar to that for upright images. The best predictions of monkey fixations in searching inverted images were obtained by rotating the results of the model predictions for the original image. The fact that the same models can predict human and monkey search behavior suggests that the monkey can be used as a good model for understanding how the human brain enables natural-scene search.
Collapse
Affiliation(s)
- Mark A Segraves
- Department of Neurobiology, Weinberg College of Arts and Sciences, Northwestern University, Evanston, IL, USA
| | - Emory Kuo
- Department of Neurobiology, Weinberg College of Arts and Sciences, Northwestern University, Evanston, IL, USA
| | - Sara Caddigan
- Department of Neurobiology, Weinberg College of Arts and Sciences, Northwestern University, Evanston, IL, USA
| | - Emily A Berthiaume
- Department of Neurobiology, Weinberg College of Arts and Sciences, Northwestern University, Evanston, IL, USA
| | - Konrad P Kording
- Departments of Physical Medicine and Rehabilitation and Physiology, Feinberg School of Medicine, Northwestern University, Chicago, IL, USA
| |
Collapse
|
15
|
Wilming N, Onat S, Ossandón JP, Açık A, Kietzmann TC, Kaspar K, Gameiro RR, Vormberg A, König P. An extensive dataset of eye movements during viewing of complex images. Sci Data 2017; 4:160126. [PMID: 28140391 PMCID: PMC5283059 DOI: 10.1038/sdata.2016.126] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2016] [Accepted: 11/15/2016] [Indexed: 11/18/2022] Open
Abstract
We present a dataset of free-viewing eye-movement recordings that contains more than 2.7 million fixation locations from 949 observers on more than 1000 images from different categories. This dataset aggregates and harmonizes data from 23 different studies conducted at the Institute of Cognitive Science at Osnabrück University and the University Medical Center in Hamburg-Eppendorf. Trained personnel recorded all studies under standard conditions with homogeneous equipment and parameter settings. All studies allowed for free eye-movements, and differed in the age range of participants (~7-80 years), stimulus sizes, stimulus modifications (phase scrambled, spatial filtering, mirrored), and stimuli categories (natural and urban scenes, web sites, fractal, pink-noise, and ambiguous artistic figures). The size and variability of viewing behavior within this dataset presents a strong opportunity for evaluating and comparing computational models of overt attention, and furthermore, for thoroughly quantifying strategies of viewing behavior. This also makes the dataset a good starting point for investigating whether viewing strategies change in patient groups.
Collapse
Affiliation(s)
- Niklas Wilming
- Institute of Cognitive Science, University of Osnabrück, 49069 Osnabrück, Germany
- Department of Neurophysiology and Pathophysiology, University Medical Center Hamburg-Eppendorf, 20246 Hamburg, Germany
| | - Selim Onat
- Institute of Cognitive Science, University of Osnabrück, 49069 Osnabrück, Germany
- Department of Systems Neuroscience, University Medical Center Hamburg-Eppendorf, 20246 Hamburg, Germany
| | - José P. Ossandón
- Institute of Cognitive Science, University of Osnabrück, 49069 Osnabrück, Germany
- Biological Psychology & Neuropsychology, University of Hamburg, 20146 Hamburg, Germany
| | - Alper Açık
- Institute of Cognitive Science, University of Osnabrück, 49069 Osnabrück, Germany
- Department of Psychology, Ozyegin University, 34716 Istanbul, Turkey
| | - Tim C. Kietzmann
- Institute of Cognitive Science, University of Osnabrück, 49069 Osnabrück, Germany
- Medical Research Council, Cognition and Brain Sciences Unit, Cambridge CB2 7EF, UK
| | - Kai Kaspar
- Institute of Cognitive Science, University of Osnabrück, 49069 Osnabrück, Germany
- Department of Psychology, University of Cologne, 50931 Cologne, Germany
| | - Ricardo R. Gameiro
- Institute of Cognitive Science, University of Osnabrück, 49069 Osnabrück, Germany
| | - Alexandra Vormberg
- Institute of Cognitive Science, University of Osnabrück, 49069 Osnabrück, Germany
- Frankfurt Institute for Advanced Studies (FIAS), 60438 Frankfurt/Main, Germany
- Ernst Strüngmann Institute (ESI) for Neuroscience in Cooperation with Max Planck Society, 60528 Frankfurt/Main, Germany
| | - Peter König
- Institute of Cognitive Science, University of Osnabrück, 49069 Osnabrück, Germany
- Department of Neurophysiology and Pathophysiology, University Medical Center Hamburg-Eppendorf, 20246 Hamburg, Germany
| |
Collapse
|
16
|
Wilming N, Kietzmann TC, Jutras M, Xue C, Treue S, Buffalo EA, König P. Differential Contribution of Low- and High-level Image Content to Eye Movements in Monkeys and Humans. Cereb Cortex 2017; 27:279-293. [PMID: 28077512 PMCID: PMC5942390 DOI: 10.1093/cercor/bhw399] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2016] [Accepted: 12/13/2016] [Indexed: 11/25/2022] Open
Abstract
Oculomotor selection exerts a fundamental impact on our experience of the environment. To better understand the underlying principles, researchers typically rely on behavioral data from humans, and electrophysiological recordings in macaque monkeys. This approach rests on the assumption that the same selection processes are at play in both species. To test this assumption, we compared the viewing behavior of 106 humans and 11 macaques in an unconstrained free-viewing task. Our data-driven clustering analyses revealed distinct human and macaque clusters, indicating species-specific selection strategies. Yet, cross-species predictions were found to be above chance, indicating some level of shared behavior. Analyses relying on computational models of visual saliency indicate that such cross-species commonalities in free viewing are largely due to similar low-level selection mechanisms, with only a small contribution by shared higher level selection mechanisms and with consistent viewing behavior of monkeys being a subset of the consistent viewing behavior of humans.
Collapse
Affiliation(s)
- Niklas Wilming
- Institute of Cognitive Science, University of Osnabrück, Osnabrück, Germany.,Department of Physiology and Biophysics, University of Washington, Seattle, WA 98195, USA.,Yerkes National Primate Research Center, Atlanta, GA 30329, USA.,Department of Neurophysiology and Pathophysiology, University Medical Center Hamburg-Eppendorf, Hamburg, Germany.,Washington National Primate Research Center, Seattle, WA 09195, USA
| | - Tim C Kietzmann
- Institute of Cognitive Science, University of Osnabrück, Osnabrück, Germany.,Medical Research Council, Cognition and Brain Sciences Unit, Cambridge CB2 7EF, UK
| | - Megan Jutras
- Department of Physiology and Biophysics, University of Washington, Seattle, WA 98195, USA.,Yerkes National Primate Research Center, Atlanta, GA 30329, USA.,Washington National Primate Research Center, Seattle, WA 09195, USA
| | - Cheng Xue
- Cognitive Neuroscience Laboratory, German Primate Center - Leibniz-Institute for Primate Research, Goettingen, Germany
| | - Stefan Treue
- Cognitive Neuroscience Laboratory, German Primate Center - Leibniz-Institute for Primate Research, Goettingen, Germany.,Faculty of Biology and Psychology, Goettingen University, Goettingen, Germany.,Leibniz-ScienceCampus Primate Cognition, Goettingen, Germany
| | - Elizabeth A Buffalo
- Department of Physiology and Biophysics, University of Washington, Seattle, WA 98195, USA.,Yerkes National Primate Research Center, Atlanta, GA 30329, USA.,Washington National Primate Research Center, Seattle, WA 09195, USA
| | - Peter König
- Institute of Cognitive Science, University of Osnabrück, Osnabrück, Germany.,Department of Neurophysiology and Pathophysiology, University Medical Center Hamburg-Eppendorf, Hamburg, Germany
| |
Collapse
|
17
|
Beran MJ, Menzel CR, Parrish AE, Perdue BM, Sayers K, Smith JD, Washburn DA. Primate cognition: attention, episodic memory, prospective memory, self-control, and metacognition as examples of cognitive control in nonhuman primates. WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2016; 7:294-316. [PMID: 27284790 PMCID: PMC5173379 DOI: 10.1002/wcs.1397] [Citation(s) in RCA: 30] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/29/2016] [Revised: 04/21/2016] [Accepted: 04/28/2016] [Indexed: 11/09/2022]
Abstract
Primate Cognition is the study of cognitive processes, which represent internal mental processes involved in discriminations, decisions, and behaviors of humans and other primate species. Cognitive control involves executive and regulatory processes that allocate attention, manipulate and evaluate available information (and, when necessary, seek additional information), remember past experiences to plan future behaviors, and deal with distraction and impulsivity when they are threats to goal achievement. Areas of research that relate to cognitive control as it is assessed across species include executive attention, episodic memory, prospective memory, metacognition, and self-control. Executive attention refers to the ability to control what sensory stimuli one attends to and how one regulates responses to those stimuli, especially in cases of conflict. Episodic memory refers to memory for personally experienced, autobiographical events. Prospective memory refers to the formation and implementation of future-intended actions, such as remembering what needs to be done later. Metacognition consists of control and monitoring processes that allow individuals to assess what information they have and what information they still need, and then if necessary to seek information. Self-control is a regulatory process whereby individuals forego more immediate or easier to obtain rewards for more delayed or harder to obtain rewards that are objectively more valuable. The behavioral complexity shown by nonhuman primates when given tests to assess these capacities indicates psychological continuities with human cognitive control capacities. However, more research is needed to clarify the proper interpretation of these behaviors with regard to possible cognitive constructs that may underlie such behaviors. WIREs Cogn Sci 2016, 7:294-316. doi: 10.1002/wcs.1397 For further resources related to this article, please visit the WIREs website.
Collapse
Affiliation(s)
- Michael J Beran
- Department of Psychology, Georgia State University, Atlanta, GA, USA
| | - Charles R Menzel
- Language Research Center, Georgia State University, Atlanta, GA, USA
| | - Audrey E Parrish
- Department of Psychology, Georgia State University, Atlanta, GA, USA
| | - Bonnie M Perdue
- Department of Psychology, Agnes Scott College, Decatur, GA, USA
| | - Ken Sayers
- Language Research Center, Georgia State University, Atlanta, GA, USA
| | - J David Smith
- Department of Psychology, Georgia State University, Atlanta, GA, USA
| | - David A Washburn
- Department of Psychology, Georgia State University, Atlanta, GA, USA
| |
Collapse
|
18
|
König SD, Buffalo EA. Modeling Visual Exploration in Rhesus Macaques with Bottom-Up Salience and Oculomotor Statistics. Front Integr Neurosci 2016; 10:23. [PMID: 27445721 PMCID: PMC4928494 DOI: 10.3389/fnint.2016.00023] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2016] [Accepted: 06/16/2016] [Indexed: 11/25/2022] Open
Abstract
There is a growing interest in studying biological systems in natural settings, in which experimental stimuli are less artificial and behavior is less controlled. In primate vision research, free viewing of complex images has elucidated novel neural responses, and free viewing in humans has helped discover attentional and behavioral impairments in patients with neurological disorders. In order to fully interpret data collected from free viewing of complex scenes, it is critical to better understand what aspects of the stimuli guide viewing behavior. To this end, we have developed a novel viewing behavior model called a Biased Correlated Random Walk (BCRW) to describe free viewing behavior during the exploration of complex scenes in monkeys. The BCRW can predict fixation locations better than bottom-up salience. Additionally, we show that the BCRW can be used to test hypotheses regarding specific attentional mechanisms. For example, we used the BCRW to examine the source of the central bias in fixation locations. Our analyses suggest that the central bias may be caused by a natural tendency to reorient the eyes toward the center of the stimulus, rather than a photographer's bias to center salient items in a scene. Taken together these data suggest that the BCRW can be used to further our understanding of viewing behavior and attention, and could be useful in optimizing stimulus and task design.
Collapse
Affiliation(s)
- Seth D König
- Wallace H. Coulter Department of Biomedical Engineering at the Georgia Institute of Technology and Emory UniversityAtlanta, GA, USA; Yerkes National Primate Research CenterAtlanta, GA, USA; Graduate Program in Neuroscience, University of WashingtonSeattle, WA, USA; Washington National Primate Research CenterSeattle, WA, USA
| | - Elizabeth A Buffalo
- Yerkes National Primate Research CenterAtlanta, GA, USA; Washington National Primate Research CenterSeattle, WA, USA; Department of Neurology, Emory University School of MedicineAtlanta, GA, USA; Department of Physiology and Biophysics, University of WashingtonSeattle, WA, USA
| |
Collapse
|
19
|
Mustafar F, De Luna P, Rainer G. Enhanced visual exploration for real objects compared to pictures during free viewing in the macaque monkey. Behav Processes 2015; 118:8-20. [DOI: 10.1016/j.beproc.2015.05.009] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2015] [Revised: 05/19/2015] [Accepted: 05/19/2015] [Indexed: 01/23/2023]
|
20
|
Weiner KF, Ghose GM. Population coding in area V4 during rapid shape detections. J Neurophysiol 2015; 113:3021-34. [PMID: 25787961 DOI: 10.1152/jn.01044.2014] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2014] [Accepted: 03/17/2015] [Indexed: 11/22/2022] Open
Abstract
While previous studies have suggested that neuronal correlations are common in visual cortex over a range of timescales, the effect of correlations on rapid visually based decisions has received little attention. We trained Macaca mulatta to saccade to a peripherally presented shape embedded in dynamic noise as soon as the shape appeared. While the monkeys performed the task, we recorded from neuronal populations (5-29 cells) using a microelectrode array implanted in area V4, a visual area thought to be involved in form perception. While modest correlations were present between cells during visual stimulation, their magnitude did not change significantly subsequent to the appearance of a shape. We quantified the reliability and temporal precision with which neuronal populations signaled the appearance of the shape and predicted the animals' choices using mutual information analyses. To study the impact of correlations, we shuffled the activity from each cell across observations while retaining stimulus-dependent modulations in firing rate. We found that removing correlations by shuffling across trials minimally affected the reliability or timing with which pairs, or larger groups of cells, signaled the presence of a shape. To assess the downstream impact of correlations, we also studied how shuffling affected the ability of V4 populations to predict behavioral choices. Surprisingly, shuffling created a modest increase in the accuracy of such predictions, suggesting that the reliability of downstream neurons is slightly compromised by activity correlations. Our findings are consistent with neuronal correlations having a minimal effect on the reliability and timing of rapid perceptual decisions.
Collapse
Affiliation(s)
- Katherine F Weiner
- Graduate Program in Neuroscience, University of Minnesota, Minneapolis, Minnesota; and
| | - Geoffrey M Ghose
- Department of Neuroscience and Center for Magnetic Resonance Research, University of Minnesota, Minneapolis, Minnesota
| |
Collapse
|
21
|
Fernandes HL, Stevenson IH, Phillips AN, Segraves MA, Kording KP. Saliency and saccade encoding in the frontal eye field during natural scene search. Cereb Cortex 2014; 24:3232-45. [PMID: 23863686 PMCID: PMC4240184 DOI: 10.1093/cercor/bht179] [Citation(s) in RCA: 48] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/14/2022] Open
Abstract
The frontal eye field (FEF) plays a central role in saccade selection and execution. Using artificial stimuli, many studies have shown that the activity of neurons in the FEF is affected by both visually salient stimuli in a neuron's receptive field and upcoming saccades in a certain direction. However, the extent to which visual and motor information is represented in the FEF in the context of the cluttered natural scenes we encounter during everyday life has not been explored. Here, we model the activities of neurons in the FEF, recorded while monkeys were searching natural scenes, using both visual and saccade information. We compare the contribution of bottom-up visual saliency (based on low-level features such as brightness, orientation, and color) and saccade direction. We find that, while saliency is correlated with the activities of some neurons, this relationship is ultimately driven by activities related to movement. Although bottom-up visual saliency contributes to the choice of saccade targets, it does not appear that FEF neurons actively encode the kind of saliency posited by popular saliency map theories. Instead, our results emphasize the FEF's role in the stages of saccade planning directly related to movement generation.
Collapse
Affiliation(s)
- Hugo L. Fernandes
- Department of Physical Medicine and Rehabilitation, Northwestern University and Rehabilitation Institute of Chicago, Chicago, IL 60611, USA
- PDBC, Instituto Gulbenkian de Ciência, 2780 Oeiras, Portugal
- Instituto de Tecnologia Química e Biológica, Universidade Nova de Lisboa, 2780 Oeiras, Portugal
| | - Ian H. Stevenson
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, CA 94720, USA
| | - Adam N. Phillips
- Tamagawa University, Brain Science Institute, Machida 194-8610, Japan
- Department of Neurobiology, Northwestern University, Evanston, IL 60208, USA
| | - Mark A. Segraves
- Department of Neurobiology, Northwestern University, Evanston, IL 60208, USA
| | - Konrad P. Kording
- Department of Physical Medicine and Rehabilitation, Northwestern University and Rehabilitation Institute of Chicago, Chicago, IL 60611, USA
- Department of Engineering Sciences and Applied Mathematics, Northwestern University, Evanston, IL 60208, USA
- Department of Physiology, Northwestern University, Chicago, IL 60611, USA
| |
Collapse
|
22
|
Weiner KF, Ghose GM. Rapid shape detection signals in area V4. Front Neurosci 2014; 8:294. [PMID: 25278828 PMCID: PMC4165234 DOI: 10.3389/fnins.2014.00294] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2014] [Accepted: 08/29/2014] [Indexed: 11/13/2022] Open
Abstract
Vision in foveate animals is an active process that requires rapid and constant decision-making. For example, when a new object appears in the visual field, we can quickly decide to inspect it by directing our eyes to the object's location. We studied the contribution of primate area V4 to these types of rapid foveation decisions. Animals performed a reaction time task that required them to report when any shape appeared within a peripherally-located noisy stimulus by making a saccade to the stimulus location. We found that about half of the randomly sampled V4 neurons not only rapidly and precisely represented the appearance of this shape, but they were also predictive of the animal's saccades. A neuron's ability to predict the animal's saccades was not related to the specificity with which the cell represented a single type of shape but rather to its ability to signal whether any shape was present. This relationship between sensory sensitivity and behavioral predictiveness was not due to global effects such as alertness, as it was equally likely to be observed for cells with increases and decreases in firing rate. Careful analysis of the timescales of reliability in these neurons implies that they reflect both feedforward and feedback shape detecting processes. In approximately 7% of our recorded sample, individual neurons were able to predict both the delay and precision of the animal's shape detection performance. This suggests that a subset of V4 neurons may have been directly and causally contributing to task performance and that area V4 likely plays a critical role in guiding rapid, form-based foveation decisions.
Collapse
Affiliation(s)
- Katherine F Weiner
- Graduate Program in Neuroscience, University of Minnesota Minneapolis, MN, USA
| | - Geoffrey M Ghose
- Graduate Program in Neuroscience, University of Minnesota Minneapolis, MN, USA ; Departments of Neuroscience, Psychology, and Radiology, Center for Magnetic Resonance Research, University of Minnesota Minneapolis, MN, USA
| |
Collapse
|
23
|
Krishna BS, Ipata AE, Bisley JW, Gottlieb J, Goldberg ME. Extrafoveal preview benefit during free-viewing visual search in the monkey. J Vis 2014; 14:6. [PMID: 24403392 PMCID: PMC5077276 DOI: 10.1167/14.1.6] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2013] [Accepted: 11/21/2013] [Indexed: 11/24/2022] Open
Abstract
Previous studies have shown that subjects require less time to process a stimulus at the fovea after a saccade if they have viewed the same stimulus in the periphery immediately prior to the saccade. This extrafoveal preview benefit indicates that information about the visual form of an extrafoveally viewed stimulus can be transferred across a saccade. Here, we extend these findings by demonstrating and characterizing a similar extrafoveal preview benefit in monkeys during a free-viewing visual search task. We trained two monkeys to report the orientation of a target among distractors by releasing one of two bars with their hand; monkeys were free to move their eyes during the task. Both monkeys took less time to indicate the orientation of the target after foveating it, when the target lay closer to the fovea during the previous fixation. An extrafoveal preview benefit emerged even if there was more than one intervening saccade between the preview and the target fixation, indicating that information about target identity could be transferred across more than one saccade and could be obtained even if the search target was not the goal of the next saccade. An extrafoveal preview benefit was also found for distractor stimuli. These results aid future physiological investigations of the extrafoveal preview benefit.
Collapse
Affiliation(s)
- B. Suresh Krishna
- Mahoney-Keck Center for Brain and Behavior Research, New York State Psychiatric Institute, New York, NY, USA
- Cognitive Neuroscience Laboratory, German Primate Center, Goettingen, Germany
| | - Anna E. Ipata
- Mahoney-Keck Center for Brain and Behavior Research, New York State Psychiatric Institute, New York, NY, USA
- Department of Neuroscience, Kavli Neuroscience Institute, Columbia University College of Physicians and Surgeons, New York, NY, USA
| | - James W. Bisley
- Mahoney-Keck Center for Brain and Behavior Research, New York State Psychiatric Institute, New York, NY, USA
- Department of Neurobiology, David Geffen School of Medicine at UCLA, Los Angeles, CA, USA
| | - Jacqueline Gottlieb
- Mahoney-Keck Center for Brain and Behavior Research, New York State Psychiatric Institute, New York, NY, USA
- Cognitive Neuroscience Laboratory, German Primate Center, Goettingen, Germany
| | - Michael E. Goldberg
- Mahoney-Keck Center for Brain and Behavior Research, New York State Psychiatric Institute, New York, NY, USA
- Department of Neuroscience, Kavli Neuroscience Institute, Columbia University College of Physicians and Surgeons, New York, NY, USA
- Departments of Neurology, Psychiatry, and Ophthalmology, Columbia University College of Physicians and Surgeons, New York, NY, USA
| |
Collapse
|
24
|
XU JIAWEI, YUE SHIGANG, TANG YUCHAO. A MOTION ATTENTION MODEL BASED ON RARITY WEIGHTING AND MOTION CUES IN DYNAMIC SCENES. INT J PATTERN RECOGN 2013. [DOI: 10.1142/s0218001413550094] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
Nowadays, motion attention model is a controversial topic in the biological computer vision area. The computational attention model can be decomposed into a set of features via predefined channels. Here we designed a bio-inspired vision attention model, and added the rarity measurement onto it. The priority of rarity is emphasized under the assumption of weighting effect upon the features logic fusion. At this stage, a final saliency map at each frame is adjusted by the spatiotemporal and rarity values. By doing this, the process of mimicking human vision attention becomes more realistic and logical to the real circumstance. The experiments are conducted on the benchmark dataset of static images and video sequences. We simulated the attention shift based on several dataset. Most importantly, our dynamic scenes are mostly selected from the objects moving on the highway and dynamic scenes. The former one can be developed on the detection of car collision and will be a useful tool for further application in robotics. We also conduct experiment on the other video clips to prove the rationality of rarity factor and feature cues fusion methods. Finally, the evaluation results indicate our visual attention model outperforms several state-of-the-art motion attention models.
Collapse
Affiliation(s)
- JIAWEI XU
- School of Computer Science, University of Lincoln, Lincoln LN6 7TS, UK
| | - SHIGANG YUE
- School of Computer Science, University of Lincoln, Lincoln LN6 7TS, UK
| | - YUCHAO TANG
- Institute for Information and System Science, Xi'an jiaotong University, Xi'an 710049, P. R. China
| |
Collapse
|
25
|
Bethell EJ, Holmes A, Maclarnon A, Semple S. Evidence that emotion mediates social attention in rhesus macaques. PLoS One 2012; 7:e44387. [PMID: 22952968 PMCID: PMC3431396 DOI: 10.1371/journal.pone.0044387] [Citation(s) in RCA: 83] [Impact Index Per Article: 6.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2012] [Accepted: 08/03/2012] [Indexed: 02/02/2023] Open
Abstract
BACKGROUND Recent work on non-human primates indicates that the allocation of social attention is mediated by characteristics of the attending animal, such as social status and genotype, as well as by the value of the target to which attention is directed. Studies of humans indicate that an individual's emotion state also plays a crucial role in mediating their social attention; for example, individuals look for longer towards aggressive faces when they are feeling more anxious, and this bias leads to increased negative arousal and distraction from other ongoing tasks. To our knowledge, no studies have tested for an effect of emotion state on allocation of social attention in any non-human species. METHODOLOGY We presented captive adult male rhesus macaques with pairs of adult male conspecific face images - one with an aggressive expression, one with a neutral expression - and recorded gaze towards these images. Each animal was tested twice, once during a putatively stressful condition (i.e. following a veterinary health check), and once during a neutral (or potentially positive) condition (i.e. a period of environmental enrichment). Initial analyses revealed that behavioural indicators of anxiety and stress were significantly higher after the health check than during enrichment, indicating that the former caused a negative shift in emotional state. PRINCIPLE FINDINGS The macaques showed initial vigilance for aggressive faces across both conditions, but subsequent responses differed between conditions. Following the health check, initial vigilance was followed by rapid and sustained avoidance of aggressive faces. By contrast, during the period of enrichment, the macaques showed sustained attention towards the same aggressive faces. CONCLUSIONS/SIGNIFICANCE These data provide, to our knowledge, the first evidence that shifts in emotion state mediate social attention towards and away from facial cues of emotion in a non-human animal. This work provides novel insights into the evolution of emotion-attention interactions in humans, and mechanisms of social behaviour in non-human primates, and may have important implications for understanding animal psychological wellbeing.
Collapse
Affiliation(s)
- Emily J Bethell
- Centre for Research in Evolutionary and Environmental Anthropology, University of Roehampton, London, United Kingdom.
| | | | | | | |
Collapse
|
26
|
Sakai H, Shin D, Kohama T, Uchiyama Y. Attentional effects on gaze preference for salient loci in traffic scenes. ERGONOMICS 2012; 55:743-751. [PMID: 22506809 DOI: 10.1080/00140139.2012.668563] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
UNLABELLED Alerting drivers for self-regulation of attention might decrease crash risks attributable to absent-minded driving. However, no reliable method exists for monitoring driver attention. Therefore, we examined attentional effects on gaze preference for salient loci (GPS) in traffic scenes. In an active viewing (AV) condition requiring endogenous attention for traffic scene comprehension, participants identified appropriate speeds for driving in presented traffic scene images. In a passive viewing (PV) condition requiring no endogenous attention, participants passively viewed traffic scene images. GPS was quantified by the mean saliency value averaged across fixation locations. Results show that GPS was less during AV than during PV. Additionally, gaze dwell time on signboards was shorter for AV than for PV. These results suggest that, in the absence of endogenous attention for traffic scene comprehension, gaze tends to concentrate on irrelevant salient loci in a traffic environment. Therefore, increased GPS can indicate absent-minded driving. PRACTITIONER SUMMARY The present study demonstrated that, without endogenous attention for traffic scene comprehension, gaze tends to concentrate on irrelevant salient loci in a traffic environment. This result suggests that increased gaze preference for salient loci indicates absent-minded driving, which is otherwise difficult to detect.
Collapse
Affiliation(s)
- Hiroyuki Sakai
- Frontier Research Center, Toyota Central Research and Development Laboratories, Inc., 41-1 Nagakute, Aichi, Japan.
| | | | | | | |
Collapse
|
27
|
Abstract
Visual saliency based on orientation contrast is a perceptual product attributed to the functional organization of the mammalian brain. We examined this visual phenomenon in barn owls by mounting a wireless video microcamera on the owls' heads and confronting them with visual scenes that contained one differently oriented target among similarly oriented distracters. Without being confined by any particular task, the owls looked significantly longer, more often, and earlier at the target, thus exhibiting visual search strategies so far demonstrated in similar conditions only in primates. Given the considerable differences in phylogeny and the structure of visual pathways between owls and humans, these findings suggest that orientation saliency has computational optimality in a wide variety of ecological contexts, and thus constitutes a universal building block for efficient visual information processing in general.
Collapse
|
28
|
Perceptual mechanism underlying gaze guidance in chimpanzees and humans. Anim Cogn 2011; 14:377-86. [PMID: 21305329 DOI: 10.1007/s10071-010-0372-3] [Citation(s) in RCA: 13] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2010] [Revised: 12/15/2010] [Accepted: 12/17/2010] [Indexed: 10/18/2022]
Abstract
Previous studies comparing eye movements between humans and their closest relatives, chimpanzees, have revealed similarities and differences between the species in terms of where individuals fixate their gaze during free viewing of a naturalistic scene, including social stimuli (e.g. body and face). However, those results were somewhat confounded by the fact that gaze behavior is influenced by low-level stimulus properties (e.g., color and form) and by high-level processes such as social sensitivity and knowledge about the scene. Given the known perceptual and cognitive similarities between chimpanzees and humans, it is expected that such low-level effects do not play a critical role in explaining the high-level similarities and differences between the species. However, there is no quantitative evidence to support this assumption. To estimate the effect of local stimulus saliency on such eye-movement patterns, this study used a well-established bottom-up saliency model. In addition, to elucidate the cues that the viewers use to guide their gaze, we presented scenes in which we had manipulated various stimulus properties. As expected, the saliency model did not fully predict the fixation patterns actually observed in chimpanzees and humans. In addition, both species used multiple cues to fixate socially significant areas such as the face. There was no evidence suggesting any differences between chimpanzees and humans in their responses to low-level saliency. Therefore, this study found a substantial amount of similarity in the perceptual mechanisms underlying gaze guidance in chimpanzees and humans and thereby offers a foundation for direct comparisons between them.
Collapse
|
29
|
Kollmorgen S, Nortmann N, Schröder S, König P. Influence of low-level stimulus features, task dependent factors, and spatial biases on overt visual attention. PLoS Comput Biol 2010; 6:e1000791. [PMID: 20502672 PMCID: PMC2873902 DOI: 10.1371/journal.pcbi.1000791] [Citation(s) in RCA: 41] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2010] [Accepted: 04/21/2010] [Indexed: 11/18/2022] Open
Abstract
Visual attention is thought to be driven by the interplay between low-level visual features and task dependent information content of local image regions, as well as by spatial viewing biases. Though dependent on experimental paradigms and model assumptions, this idea has given rise to varying claims that either bottom-up or top-down mechanisms dominate visual attention. To contribute toward a resolution of this discussion, here we quantify the influence of these factors and their relative importance in a set of classification tasks. Our stimuli consist of individual image patches (bubbles). For each bubble we derive three measures: a measure of salience based on low-level stimulus features, a measure of salience based on the task dependent information content derived from our subjects' classification responses and a measure of salience based on spatial viewing biases. Furthermore, we measure the empirical salience of each bubble based on our subjects' measured eye gazes thus characterizing the overt visual attention each bubble receives. A multivariate linear model relates the three salience measures to overt visual attention. It reveals that all three salience measures contribute significantly. The effect of spatial viewing biases is highest and rather constant in different tasks. The contribution of task dependent information is a close runner-up. Specifically, in a standardized task of judging facial expressions it scores highly. The contribution of low-level features is, on average, somewhat lower. However, in a prototypical search task, without an available template, it makes a strong contribution on par with the two other measures. Finally, the contributions of the three factors are only slightly redundant, and the semi-partial correlation coefficients are only slightly lower than the coefficients for full correlations. These data provide evidence that all three measures make significant and independent contributions and that none can be neglected in a model of human overt visual attention.
Collapse
Affiliation(s)
- Sepp Kollmorgen
- Institute of Neurobiopsychology, University of Osnabrück, Osnabrück, Germany
- Institute of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland
| | - Nora Nortmann
- Institute of Neurobiopsychology, University of Osnabrück, Osnabrück, Germany
| | - Sylvia Schröder
- Institute of Neurobiopsychology, University of Osnabrück, Osnabrück, Germany
- Institute of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland
- * E-mail:
| | - Peter König
- Institute of Neurobiopsychology, University of Osnabrück, Osnabrück, Germany
| |
Collapse
|
30
|
Shepherd SV, Steckenfinger SA, Hasson U, Ghazanfar AA. Human-monkey gaze correlations reveal convergent and divergent patterns of movie viewing. Curr Biol 2010; 20:649-56. [PMID: 20303267 PMCID: PMC2855404 DOI: 10.1016/j.cub.2010.02.032] [Citation(s) in RCA: 58] [Impact Index Per Article: 3.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2009] [Revised: 01/29/2010] [Accepted: 02/01/2010] [Indexed: 11/18/2022]
Abstract
The neuroanatomical organization of the visual system is largely similar across primate species, predicting similar visual behaviors and perceptions. Although responses to trial-by-trial presentation of static images suggest that primates share visual orienting strategies, these reduced stimuli fail to capture key elements of the naturalistic, dynamic visual world in which we evolved. Here, we compared the gaze behavior of humans and macaques when they viewed three different 3-minute movie clips. We found significant intersubject and interspecies gaze correlations, suggesting that both species attend a common set of events in each scene. Comparing human and monkey gaze behavior with a computational saliency model revealed that interspecies gaze correlations were driven by biologically relevant social stimuli overlooked by low-level saliency models. Additionally, humans, but not monkeys, tended to gaze toward the targets of viewed individual's actions or gaze. Together, these data suggest that human and monkey gaze behavior comprises converging and diverging informational strategies, driven by both scene content and context; they are not fully described by simple low-level visual models.
Collapse
|
31
|
Comparison of Redundancy and Relevance Measures for Feature Selection in Tissue Classification of CT Images. ACTA ACUST UNITED AC 2010. [DOI: 10.1007/978-3-642-14400-4_20] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register]
|
32
|
Nothdurft HC, Pigarev IN, Kastner S. Overt and covert visual search in primates: reaction times and gaze shift strategies. J Integr Neurosci 2009; 8:137-74. [PMID: 19618485 DOI: 10.1142/s0219635209002101] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2008] [Revised: 03/20/2009] [Indexed: 11/18/2022] Open
Abstract
In order to investigate the search performance and strategies of nonhuman primates, two macaque monkeys were trained to search for a target template among differently oriented distractors in both free-gaze and fixed-gaze viewing conditions (overt and covert search). In free-gaze search, reaction times (RT) and eye movements revealed the theoretically predicted characteristics of exhaustive and self-terminating serial search, with certain exceptions that are also observed in humans. RT was linearly related to the number of fixations but not necessarily to the number of items on display. Animals scanned the scenes in a nonrandom manner spending notably more time on targets and items inspected last (just before reaction). The characteristics of free-gaze search were then compared with search performance under fixed gaze (covert search) and with the performance of four human subjects tested in similar experiments. By and large the performance characteristics of both groups were similar; monkeys were slightly faster, and humans more accurate. Both species produced shorter RT in fixed-gaze than in free-gaze search. But while RT slopes of the human subjects still showed the theoretically predicted difference between hits and rejections, slopes of the two monkeys appeared to collapse. Despite considerable priming and short-term learning when similar tests were continuously repeated, no substantial long-term training effects were seen when test conditions and set sizes were frequently varied. Altogether, the data reveal many similarities between human and monkey search behavior but indicate that search is not necessarily restricted to exclusively serial processes.
Collapse
|
33
|
Saliency on a natural scene background: effects of color and luminance contrast add linearly. Atten Percept Psychophys 2009; 71:1337-52. [PMID: 19633349 DOI: 10.3758/app.71.6.1337] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
In natural vision, shifts in spatial attention are associated with shifts of gaze. Computational models of such overt attention typically use the concept of a saliency map: Normalized maps of center-surround differences are computed for individual stimulus features and added linearly to obtain the saliency map. Although the predictions of such models correlate with fixated locations better than chance, their mechanistic assumptions are less well investigated. Here, we tested one key assumption: Do the effects of different features add linearly or according to a max-type of interaction? We measured the eye position of observers viewing natural stimuli whose luminance contrast and/or color contrast (saturation) increased gradually toward one side. We found that these feature gradients biased fixations toward regions of high contrasts. When two contrast gradients (color and luminance) were superimposed, linear summation of their individual effects predicted their combined effect. This demonstrated that the interaction of color and luminance contrast with respect to human overt attention is--irrespective of the precise model--consistent with the assumption of linearity, but not with a max-type interaction of these features.
Collapse
|
34
|
Itti L, Baldi P. Bayesian surprise attracts human attention. Vision Res 2009; 49:1295-306. [PMID: 18834898 PMCID: PMC2782645 DOI: 10.1016/j.visres.2008.09.007] [Citation(s) in RCA: 391] [Impact Index Per Article: 24.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2007] [Revised: 09/02/2008] [Accepted: 09/02/2008] [Indexed: 11/18/2022]
Abstract
We propose a formal Bayesian definition of surprise to capture subjective aspects of sensory information. Surprise measures how data affects an observer, in terms of differences between posterior and prior beliefs about the world. Only data observations which substantially affect the observer's beliefs yield surprise, irrespectively of how rare or informative in Shannon's sense these observations are. We test the framework by quantifying the extent to which humans may orient attention and gaze towards surprising events or items while watching television. To this end, we implement a simple computational model where a low-level, sensory form of surprise is computed by simple simulated early visual neurons. Bayesian surprise is a strong attractor of human attention, with 72% of all gaze shifts directed towards locations more surprising than the average, a figure rising to 84% when focusing the analysis onto regions simultaneously selected by all observers. The proposed theory of surprise is applicable across different spatio-temporal scales, modalities, and levels of abstraction.
Collapse
Affiliation(s)
- Laurent Itti
- Computer Science Department, University of Southern California, Los Angeles, 90089, USA.
| | | |
Collapse
|
35
|
Açik A, Onat S, Schumann F, Einhäuser W, König P. Effects of luminance contrast and its modifications on fixation behavior during free viewing of images from different categories. Vision Res 2009; 49:1541-53. [DOI: 10.1016/j.visres.2009.03.011] [Citation(s) in RCA: 31] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2008] [Revised: 03/12/2009] [Accepted: 03/12/2009] [Indexed: 11/29/2022]
Affiliation(s)
- Alper Açik
- University of Osnabrück, Institute of Cognitive Science, Osnabrück, Germany.
| | | | | | | | | |
Collapse
|
36
|
Zhang L, Tong MH, Marks TK, Shan H, Cottrell GW. SUN: A Bayesian framework for saliency using natural statistics. J Vis 2008; 8:32.1-20. [PMID: 19146264 DOI: 10.1167/8.7.32] [Citation(s) in RCA: 286] [Impact Index Per Article: 16.8] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2007] [Accepted: 05/17/2008] [Indexed: 11/24/2022] Open
Abstract
We propose a definition of saliency by considering what the visual system is trying to optimize when directing attention. The resulting model is a Bayesian framework from which bottom-up saliency emerges naturally as the self-information of visual features, and overall saliency (incorporating top-down information with bottom-up saliency) emerges as the pointwise mutual information between the features and the target when searching for a target. An implementation of our framework demonstrates that our model's bottom-up saliency maps perform as well as or better than existing algorithms in predicting people's fixations in free viewing. Unlike existing saliency measures, which depend on the statistics of the particular image being viewed, our measure of saliency is derived from natural image statistics, obtained in advance from a collection of natural images. For this reason, we call our model SUN (Saliency Using Natural statistics). A measure of saliency based on natural image statistics, rather than based on a single test image, provides a straightforward explanation for many search asymmetries observed in humans; the statistics of a single test image lead to predictions that are not consistent with these asymmetries. In our model, saliency is computed locally, which is consistent with the neuroanatomy of the early visual system and results in an efficient algorithm with few free parameters.
Collapse
Affiliation(s)
- Lingyun Zhang
- Department of Computer Science and Engineering, UCSD, La Jolla, CA, USA.
| | | | | | | | | |
Collapse
|
37
|
Ohayon S, Harmening W, Wagner H, Rivlin E. Through a barn owl's eyes: interactions between scene content and visual attention. BIOLOGICAL CYBERNETICS 2008; 98:115-132. [PMID: 18066583 DOI: 10.1007/s00422-007-0199-4] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/31/2007] [Accepted: 10/22/2007] [Indexed: 05/25/2023]
Abstract
In this study we investigated visual attention properties of freely behaving barn owls, using a miniature wireless camera attached to their heads. The tubular eye structure of barn owls makes them ideal subjects for this research since it limits their eye movements. Video sequences recorded from the owl's point of view capture part of the visual scene as seen by the owl. Automated analysis of video sequences revealed that during an active search task, owls repeatedly and consistently direct their gaze in a way that brings objects of interest to a specific retinal location (retinal fixation area). Using a projective model that captures the geometry between the eye and the camera, we recovered the corresponding location in the recorded images (image fixation area). Recording in various types of environments (aviary, office, outdoors) revealed significant statistical differences of low level image properties at the image fixation area compared to values extracted at random image patches. These differences are in agreement with results obtained in primates in similar studies. To investigate the role of saliency and its contribution to drawing the owl's attention, we used a popular bottom-up computational model. Saliency values at the image fixation area were typically greater than at random patches, yet were only 20% out of the maximal saliency value, suggesting a top-down modulation of gaze control.
Collapse
Affiliation(s)
- Shay Ohayon
- Israel Institute of Technology (Technion), Haifa 32000, Israel.
| | | | | | | |
Collapse
|
38
|
Kayser C, Nielsen KJ, Logothetis NK. Fixations in natural scenes: interaction of image structure and image content. Vision Res 2006; 46:2535-45. [PMID: 16545420 DOI: 10.1016/j.visres.2006.02.003] [Citation(s) in RCA: 23] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2005] [Revised: 01/23/2006] [Accepted: 02/01/2006] [Indexed: 10/24/2022]
Abstract
Explorative eye movements specifically target some parts of a scene while ignoring others. Here, we investigate how local image structure--defined by spatial frequency contrast--and informative image content--defined by higher order image statistics-are weighted for the selection of fixation points. We measured eye movements of macaque monkeys freely viewing a set of natural and manipulated images outside a particular task. To probe the effect of scene content, we locally introduced patches of pink noise into natural images, and to probe the interaction with image structure, we altered the contrast of the noise. We found that fixations specifically targeted the natural image parts and spared the uninformative noise patches. However, both increasing and decreasing the contrast of the noise attracted more fixations, and, in the extreme cases, compensated the effect of missing content. Introducing patches from another natural image led to similar results. In all paradigms tested, the interaction between scene structure and informative scene content was the same in any of the first six fixations on an image, demonstrating that the weighting of these factors is constant during viewing of an image. These results question theories, which suggest that initial fixations are driven by stimulus structure whereas later fixations are determined by informative scene content.
Collapse
Affiliation(s)
- Christoph Kayser
- Max Planck Institute for Biological Cybernetics, Spemannstrasse 38, 72076 Tübingen, Germany.
| | | | | |
Collapse
|