1
|
Portugal AM, Taylor MJ, Tammimies K, Ronald A, Falck-Ytter T. Dissociable genetic influences on eye movements during abstract versus naturalistic social scene viewing in infancy. Sci Rep 2025; 15:4100. [PMID: 39900629 PMCID: PMC11791049 DOI: 10.1038/s41598-024-83557-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2024] [Accepted: 12/16/2024] [Indexed: 02/05/2025] Open
Abstract
Eye-movement metrics like fixation location and duration are increasingly being used in infancy research. We tested whether fixation durations during meaningful social stimulus viewing involve common or different familial influences than fixation durations during viewing of abstract stimulus. We analysed the duration of fixations, and the allocation of fixations to face and motion, from 536 dizygotic and monozygotic 5-month-old twins in: naturalistic scenes including low- and high-level social features, and abstract scenes only having low-level features. We observed significant genetic influences in both conditions (h2naturalistic = 0.30, 95% confidence interval (CI) 0.14 to 0.44; h2abstract = 0.25, 95% CI 0.09 to 0.39), while shared environmental influences were negligible. Although some genetic influences were shared between the two conditions, unique genetic factors were linked to naturalistic scene viewing, indicating that fixation durations index different phenomena dependent on the context. Heritability for face looking was moderate (h2 = 0.19, 95% CI 0.03 to 0.34), and no familial influences were found for motion looking. Exploratory polygenic score analyses revealed no significant associations with fixation measures. This study underscores the dissociable genetic influences on infants' visual exploration of abstract versus naturalistic stimuli and the importance of considering context when interpreting eye-tracking data.
Collapse
Affiliation(s)
- Ana Maria Portugal
- Development and Neurodiversity Lab (DIVE), Department of Psychology, Uppsala University, Uppsala, Sweden.
- Center of Neurodevelopmental Disorders (KIND), Centre for Psychiatry Research, Department of Women's and Children's Health, Karolinska Institutet and Stockholm Health Care Services, Region Stockholm, Stockholm, Sweden.
| | - Mark J Taylor
- Department of Medical Epidemiology and Biostatistics, Karolinska Institutet, Stockholm, Sweden
| | - Kristiina Tammimies
- Center of Neurodevelopmental Disorders (KIND), Centre for Psychiatry Research, Department of Women's and Children's Health, Karolinska Institutet and Stockholm Health Care Services, Region Stockholm, Stockholm, Sweden
- Astrid Lindgren Children's Hospital, Karolinska University Hospital, Region Stockholm, Stockholm, Sweden
| | - Angelica Ronald
- School of Psychology, Faculty of Health and Medical Sciences, University of Surrey, Surrey, UK
| | - Terje Falck-Ytter
- Development and Neurodiversity Lab (DIVE), Department of Psychology, Uppsala University, Uppsala, Sweden.
- Center of Neurodevelopmental Disorders (KIND), Centre for Psychiatry Research, Department of Women's and Children's Health, Karolinska Institutet and Stockholm Health Care Services, Region Stockholm, Stockholm, Sweden.
| |
Collapse
|
2
|
Roth N, McLaughlin J, Obermayer K, Rolfs M. Gaze Behavior Reveals Expectations of Potential Scene Changes. Psychol Sci 2024; 35:1350-1363. [PMID: 39570640 DOI: 10.1177/09567976241279198] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2024] Open
Abstract
Even if the scene before our eyes remains static for some time, we might explore it differently compared with how we examine static images, which are commonly used in studies on visual attention. Here we show experimentally that the top-down expectation of changes in natural scenes causes clearly distinguishable gaze behavior for visually identical scenes. We present free-viewing eye-tracking data of 20 healthy adults on a new video dataset of natural scenes, each mapped for its potential for change (PfC) in independent ratings. Observers looking at frozen videos looked significantly more often at the parts of the scene with a high PfC compared with static images, with substantially higher interobserver coherence. This viewing difference peaked right before a potential movement onset. Established concepts like object animacy or salience alone could not explain this finding. Images thus conceal experience-based expectations that affect gaze behavior in the potentially dynamic real world.
Collapse
Affiliation(s)
- Nicolas Roth
- Cluster of Excellence Science of Intelligence, Technische Universität Berlin
- Institute of Software Engineering and Theoretical Computer Science, Technische Universität Berlin
| | - Jasper McLaughlin
- Cluster of Excellence Science of Intelligence, Technische Universität Berlin
| | - Klaus Obermayer
- Cluster of Excellence Science of Intelligence, Technische Universität Berlin
- Institute of Software Engineering and Theoretical Computer Science, Technische Universität Berlin
- Bernstein Center for Computational Neuroscience Berlin, Berlin, Germany
| | - Martin Rolfs
- Cluster of Excellence Science of Intelligence, Technische Universität Berlin
- Bernstein Center for Computational Neuroscience Berlin, Berlin, Germany
- Department of Psychology, Humboldt-Universität zu Berlin
| |
Collapse
|
3
|
Jing M, Kadooka K, Franchak J, Kirkorian HL. The effect of narrative coherence and visual salience on children's and adults' gaze while watching video. J Exp Child Psychol 2023; 226:105562. [PMID: 36257254 DOI: 10.1016/j.jecp.2022.105562] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2021] [Revised: 09/12/2022] [Accepted: 09/14/2022] [Indexed: 11/05/2022]
Abstract
Low-level visual features (e.g., motion, contrast) predict eye gaze during video viewing. The current study investigated the effect of narrative coherence on the extent to which low-level visual salience predicts eye gaze. Eye movements were recorded as 4-year-olds (n = 20) and adults (n = 20) watched a cohesive versus random sequence of video shots from a 4.5-min full vignette from Sesame Street. Overall, visual salience was a stronger predictor of gaze in adults than in children, especially when viewing a random shot sequence. The impact of narrative coherence on children's gaze was limited to the short period of time surrounding cuts to new video shots. The discussion considers potential direct effects of visual salience as well as incidental effects due to overlap between salient features and semantic content. The findings are also discussed in the context of developing video comprehension.
Collapse
Affiliation(s)
- Mengguo Jing
- Department of Human Development and Family Studies, University of Wisconsin-Madison, Madison, WI 53705, USA.
| | - Kellan Kadooka
- Department of Psychology, University of California, Riverside, Riverside, CA 92521, USA
| | - John Franchak
- Department of Psychology, University of California, Riverside, Riverside, CA 92521, USA
| | - Heather L Kirkorian
- Department of Human Development and Family Studies, University of Wisconsin-Madison, Madison, WI 53705, USA
| |
Collapse
|
4
|
Franchak JM, Kadooka K. Age differences in orienting to faces in dynamic scenes depend on face centering, not visual saliency. INFANCY 2022; 27:1032-1051. [PMID: 35932474 DOI: 10.1111/infa.12492] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
Abstract
The current study investigated how infants (6-24 months), children (2-12 years), and adults differ in how visual cues-visual saliency and centering-guide their attention to faces in videos. We report a secondary analysis of Kadooka and Franchak (2020), in which observers' eye movements were recorded during viewing of television clips containing a variety of faces. For every face on every video frame, we calculated its visual saliency (based on both static and dynamic image features) and calculated how close the face was to the center of the image. Results revealed that participants of every age looked more often at each face when it was more salient compared to less salient. In contrast, centering did not increase the likelihood that infants looked at a given face, but in later childhood and adulthood, centering became a stronger cue for face looking. A control analysis determined that the age-related change in centering was specific to face looking; participants of all ages were more likely to look at the center of the image, and this center bias did not change with age. The implications for using videos in educational and diagnostic contexts are discussed.
Collapse
|
5
|
Effect of sequential video shot comprehensibility on attentional synchrony: A comparison of children and adults. Proc Natl Acad Sci U S A 2018; 115:9867-9874. [PMID: 30275303 DOI: 10.1073/pnas.1611606114] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
To comprehend edited video, viewers must infer the meaning conveyed by successive video shots (i.e., continuous video segments separated by edit points, such as camera cuts). The central question here was whether comprehension-related top-down cognitive processes drive eye movements during sequential processing of video montage. Eye movements were recorded as 4 year olds and adults (n = 62) watched a video with the same constituent shots in either normal or random sequence. The key analyses compared eye movements to constituent shots when presented in normal order with those to the same shots presented in random order. The dependent variable was attentional synchrony or the extent to which viewers looked at the same location at the same time, indicating commonality of processing the video. This was calculated as the bivariate contour ellipse area within which points of gaze fell during each video frame. Results indicated that children were more scattered in their gaze locations than adults. Viewers became more similar to each other as normal vignettes unfolded over time; this was especially true in adults and possibly reflects a growing and shared understanding of the content. Conversely, adult attentional synchrony was reduced when watching random shot sequences. Thus, attentional synchrony during normal video viewing is driven not only by salient visual features, such as movement and areas of high contrast, but also, by the unfolding sequential comprehension of video montage, especially in adults. Differences between children and adults indicate that this top-down control of eye movements while watching video changes systematically over development.
Collapse
|
6
|
Le Meur O, Coutrot A, Liu Z, Rama P, Le Roch A, Helo A. Visual Attention Saccadic Models Learn to Emulate Gaze Patterns From Childhood to Adulthood. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2017; 26:4777-4789. [PMID: 28682255 DOI: 10.1109/tip.2017.2722238] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
How people look at visual information reveals fundamental information about themselves, their interests and their state of mind. While previous visual attention models output static 2D saliency maps, saccadic models aim to predict not only where observers look at but also how they move their eyes to explore the scene. In this paper, we demonstrate that saccadic models are a flexible framework that can be tailored to emulate observer's viewing tendencies. More specifically, we use fixation data from 101 observers split into five age groups (adults, 8-10 y.o., 6-8 y.o., 4-6 y.o., and 2 y.o.) to train our saccadic model for different stages of the development of human visual system. We show that the joint distribution of saccade amplitude and orientation is a visual signature specific to each age group, and can be used to generate age-dependent scan paths. Our age-dependent saccadic model does not only output human-like, age-specific visual scan paths, but also significantly outperforms other state-of-the-art saliency models. We demonstrate that the computational modeling of visual attention, through the use of saccadic model, can be efficiently adapted to emulate the gaze behavior of a specific group of observers.
Collapse
|