1
|
Liu J, Chi J, Yang Z. A review on personal calibration issues for video-oculographic-based gaze tracking. Front Psychol 2024; 15:1309047. [PMID: 38572211 PMCID: PMC10987702 DOI: 10.3389/fpsyg.2024.1309047] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2023] [Accepted: 03/08/2024] [Indexed: 04/05/2024] Open
Abstract
Personal calibration is a process of obtaining personal gaze-related information by focusing on some calibration benchmarks when the user initially uses a gaze tracking system. It not only provides conditions for gaze estimation, but also improves gaze tracking performance. Existing eye-tracking products often require users to conduct explicit personal calibration first, thereby tracking and interacting based on their gaze. This calibration mode has certain limitations, and there is still a significant gap between theoretical personal calibration methods and their practicality. Therefore, this paper reviews the issues of personal calibration for video-oculographic-based gaze tracking. The personal calibration information in typical gaze tracking methods is first summarized, and then some main settings in existing personal calibration processes are analyzed. Several personal calibration modes are discussed and compared subsequently. The performance of typical personal calibration methods for 2D and 3D gaze tracking is quantitatively compared through simulation experiments, highlighting the characteristics of different personal calibration settings. On this basis, we discuss several key issues in designing personal calibration. To the best of our knowledge, this is the first review on personal calibration issues for video-oculographic-based gaze tracking. It aims to provide a comprehensive overview of the research status of personal calibration, explore its main directions for further studies, and provide guidance for seeking personal calibration modes that conform to natural human-computer interaction and promoting the widespread application of eye-movement interaction.
Collapse
Affiliation(s)
- Jiahui Liu
- School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing, China
- Beijing Engineering Research Center of Industrial Spectrum Imaging, University of Science and Technology Beijing, Beijing, China
- Shunde Innovation School, University of Science and Technology Beijing, Foshan, China
| | - Jiannan Chi
- School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing, China
- Beijing Engineering Research Center of Industrial Spectrum Imaging, University of Science and Technology Beijing, Beijing, China
- Shunde Innovation School, University of Science and Technology Beijing, Foshan, China
| | - Zuoyun Yang
- School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing, China
| |
Collapse
|
2
|
Capparini C, To MPS, Dardenne C, Reid VM. Offline Calibration for Infant Gaze and Head Tracking across a Wide Horizontal Visual Field. SENSORS (BASEL, SWITZERLAND) 2023; 23:972. [PMID: 36679775 PMCID: PMC9866781 DOI: 10.3390/s23020972] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/14/2022] [Revised: 01/10/2023] [Accepted: 01/11/2023] [Indexed: 06/17/2023]
Abstract
Most well-established eye-tracking research paradigms adopt remote systems, which typically feature regular flat screens of limited width. Limitations of current eye-tracking methods over a wide area include calibration, the significant loss of data due to head movements, and the reduction of data quality over the course of an experimental session. Here, we introduced a novel method of tracking gaze and head movements that combines the possibility of investigating a wide field of view and an offline calibration procedure to enhance the accuracy of measurements. A 4-camera Smart Eye Pro system was adapted for infant research to detect gaze movements across 126° of the horizontal meridian. To accurately track this visual area, an online system calibration was combined with a new offline gaze calibration procedure. Results revealed that the proposed system successfully tracked infants' head and gaze beyond the average screen size. The implementation of an offline calibration procedure improved the validity and spatial accuracy of measures by correcting a systematic top-right error (1.38° mean horizontal error and 1.46° mean vertical error). This approach could be critical for deriving accurate physiological measures from the eye and represents a substantial methodological advance for tracking looking behaviour across both central and peripheral regions. The offline calibration is particularly useful for work with developing populations, such as infants, and for people who may have difficulties in following instructions.
Collapse
Affiliation(s)
- Chiara Capparini
- Center for Research in Cognition & Neuroscience (CRCN), Université Libre de Bruxelles, 1050 Brussels, Belgium
- Department of Psychology, Lancaster University, Lancaster LA1 4YF, UK
| | - Michelle P. S. To
- Department of Psychology, Lancaster University, Lancaster LA1 4YF, UK
| | | | - Vincent M. Reid
- School of Psychology, University of Waikato, Hamilton 3240, New Zealand
| |
Collapse
|
3
|
Akinyelu AA, Blignaut P. Convolutional Neural Network-Based Technique for Gaze Estimation on Mobile Devices. Front Artif Intell 2022; 4:796825. [PMID: 35156012 PMCID: PMC8826079 DOI: 10.3389/frai.2021.796825] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2021] [Accepted: 12/20/2021] [Indexed: 11/13/2022] Open
Abstract
Eye tracking is becoming a very popular, useful, and important technology. Many eye tracking technologies are currently expensive and only available to large corporations. Some of them necessitate explicit personal calibration, which makes them unsuitable for use in real-world or uncontrolled environments. Explicit personal calibration can also be cumbersome and degrades the user experience. To address these issues, this study proposes a Convolutional Neural Network (CNN) based calibration-free technique for improved gaze estimation in unconstrained environments. The proposed technique consists of two components, namely a face component and a 39-point facial landmark component. The face component is used to extract the gaze estimation features from the eyes, while the 39-point facial landmark component is used to encode the shape and location of the eyes (within the face) into the network. Adding this information can make the network learn free-head and eye movements. Another CNN model was designed in this study primarily for the sake of comparison. The CNN model accepts only the face images as input. Different experiments were performed, and the experimental result reveals that the proposed technique outperforms the second model. Fine-tuning was also performed using the VGG16 pre-trained model. Experimental results show that the fine-tuned results of the proposed technique perform better than the fine-tuned results of the second model. Overall, the results show that 39-point facial landmarks can be used to improve the performance of CNN-based gaze estimation models.
Collapse
|
4
|
Hsu WY, Chung CJ. A Novel Eye Center Localization Method for Head Poses With Large Rotations. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2020; 30:1369-1381. [PMID: 33332268 DOI: 10.1109/tip.2020.3044209] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Eye localization is undoubtedly crucial to acquiring large amounts of information. It not only helps people improve their understanding of others but is also a technology that enables machines to better understand humans. Although studies have reported satisfactory accuracy for frontal faces or head poses at limited angles, large head rotations generate numerous defects (e.g., disappearance of the eye), and existing methods are not effective enough to accurately localize eye centers. Therefore, this study makes three contributions to address these limitations. First, we propose a novel complete representation (CR) pipeline that can flexibly learn and generate two complete representations, namely the CR-center and CR-region, of the same identity. We also propose two novel eye center localization methods. This first method employs geometric transformation to estimate the rotational difference between two faces and an unknown-localization strategy for accurate transformation of the CR-center. The second method is based on image translation learning and uses the CR-region to train the generative adversarial network, which can then accurately generate and localize eye centers. Five image databases are employed to verify the proposed methods, and tests reveal that compared with existing methods, the proposed method can more accurately and robustly localize eye centers in challenging images, such as those showing considerable head rotation (both yaw rotation of -67.5° to +67.5° and roll rotation of +120° to -120°), complete occlusion of both eyes, poor illumination in addition to head rotation, head pose changes in the dark, and various gaze interaction.
Collapse
|
5
|
Larrazabal A, García Cena C, Martínez C. Video-oculography eye tracking towards clinical applications: A review. Comput Biol Med 2019; 108:57-66. [DOI: 10.1016/j.compbiomed.2019.03.025] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2018] [Revised: 02/20/2019] [Accepted: 03/26/2019] [Indexed: 10/27/2022]
|
6
|
Kasprowski P, Harezlak K, Skurowski P. Implicit Calibration Using Probable Fixation Targets. SENSORS 2019; 19:s19010216. [PMID: 30626162 PMCID: PMC6339230 DOI: 10.3390/s19010216] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/08/2018] [Revised: 12/13/2018] [Accepted: 12/25/2018] [Indexed: 11/16/2022]
Abstract
Proper calibration of eye movement signal registered by an eye tracker seems to be one of the main challenges in popularizing eye trackers as yet another user-input device. Classic calibration methods taking time and imposing unnatural behavior on eyes must be replaced by intelligent methods that are able to calibrate the signal without conscious cooperation by the user. Such an implicit calibration requires some knowledge about the stimulus a user is looking at and takes into account this information to predict probable gaze targets. This paper describes a possible method to perform implicit calibration: it starts with finding probable fixation targets (PFTs), then it uses these targets to build a mapping-probable gaze path. Various algorithms that may be used for finding PFTs and mappings are presented in the paper and errors are calculated using two datasets registered with two different types of eye trackers. The results show that although for now the implicit calibration provides results worse than the classic one, it may be comparable with it and sufficient for some applications.
Collapse
Affiliation(s)
- Pawel Kasprowski
- Institute of Informatics, Silesian University of Technology, Akademicka 16, 44-100 Gliwice, Poland.
| | - Katarzyna Harezlak
- Institute of Informatics, Silesian University of Technology, Akademicka 16, 44-100 Gliwice, Poland.
| | - Przemysław Skurowski
- Institute of Informatics, Silesian University of Technology, Akademicka 16, 44-100 Gliwice, Poland.
| |
Collapse
|
7
|
Harrar V, Le Trung W, Malienko A, Khan AZ. A nonvisual eye tracker calibration method for video-based tracking. J Vis 2018; 18:13. [PMID: 30208432 DOI: 10.1167/18.9.13] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Video-based eye trackers have enabled major advancements in our understanding of eye movements through their ease of use and their non-invasiveness. One necessity to obtain accurate eye recordings using video-based trackers is calibration. The aim of the current study was to determine the feasibility and reliability of alternative calibration methods for scenarios in which the standard visual-calibration is not possible. Fourteen participants were tested using the EyeLink 1000 Plus video-based eye tracker, and each completed the following 5-point calibration methods: 1) standard visual-target calibration, 2) described calibration where participants were provided with verbal instructions about where to direct their eyes (without vision of the screen), 3) proprioceptive calibration where participants were asked to look at their hidden finger, 4) replacement calibration, where the visual calibration was performed by 3 different people; the calibrators were temporary substitutes for the participants. Following calibration, participants performed a simple visually-guided saccade task to 16 randomly presented targets on a grid. We found that precision errors were comparable across the alternative calibration methods. In terms of accuracy, compared to the standard calibration, non-visual calibration methods (described and proprioception) led to significantly larger errors, whilst the replacement calibration method had much smaller errors. In conditions where calibration is not possible, for example when testing blind or visually impaired people who are unable to foveate the calibration targets, we suggest that using a single stand-in to perform the calibration is a simple and easy alternative calibration method, which should only cause a minimal decrease in accuracy.
Collapse
Affiliation(s)
- Vanessa Harrar
- Vision, Attention, and Action Laboratory (VISATTAC), School of Optometry, University of Montreal, Montreal, Quebec, Canada
| | - William Le Trung
- Vision, Attention, and Action Laboratory (VISATTAC), School of Optometry, University of Montreal, Montreal, Quebec, Canada
| | - Anton Malienko
- Vision, Attention, and Action Laboratory (VISATTAC), School of Optometry, University of Montreal, Montreal, Quebec, Canada
| | - Aarlenne Zein Khan
- Vision, Attention, and Action Laboratory (VISATTAC), School of Optometry, University of Montreal, Montreal, Quebec, Canada
| |
Collapse
|
8
|
Fujii K, Gras G, Salerno A, Yang GZ. Gaze gesture based human robot interaction for laparoscopic surgery. Med Image Anal 2017; 44:196-214. [PMID: 29277075 DOI: 10.1016/j.media.2017.11.011] [Citation(s) in RCA: 24] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2016] [Revised: 11/22/2017] [Accepted: 11/23/2017] [Indexed: 02/07/2023]
Abstract
While minimally invasive surgery offers great benefits in terms of reduced patient trauma, bleeding, as well as faster recovery time, it still presents surgeons with major ergonomic challenges. Laparoscopic surgery requires the surgeon to bimanually control surgical instruments during the operation. A dedicated assistant is thus required to manoeuvre the camera, which is often difficult to synchronise with the surgeon's movements. This article introduces a robotic system in which a rigid endoscope held by a robotic arm is controlled via the surgeon's eye movement, thus forgoing the need for a camera assistant. Gaze gestures detected via a series of eye movements are used to convey the surgeon's intention to initiate gaze contingent camera control. Hidden Markov Models (HMMs) are used for real-time gaze gesture recognition, allowing the robotic camera to pan, tilt, and zoom, whilst immune to aberrant or unintentional eye movements. A novel online calibration method for the gaze tracker is proposed, which overcomes calibration drift and simplifies its clinical application. This robotic system has been validated by comprehensive user trials and a detailed analysis performed on usability metrics to assess the performance of the system. The results demonstrate that the surgeons can perform their tasks quicker and more efficiently when compared to the use of a camera assistant or foot switches.
Collapse
Affiliation(s)
- Kenko Fujii
- The Hamlyn Centre for Robotic Surgery, Imperial College London, UK
| | - Gauthier Gras
- The Hamlyn Centre for Robotic Surgery, Imperial College London, UK
| | - Antonino Salerno
- The Hamlyn Centre for Robotic Surgery, Imperial College London, UK
| | - Guang-Zhong Yang
- The Hamlyn Centre for Robotic Surgery, Imperial College London, UK.
| |
Collapse
|
9
|
Abstract
This paper presents a method for computing the gaze point using camera data captured with a wearable gaze tracking device. The method utilizes a physical model of the human eye, advanced Bayesian computer vision algorithms, and Kalman filtering, resulting in high accuracy and low noise. Our C++ implementation can process camera streams with 30 frames per second in realtime. The performance of the system is validated in an exhaustive experimental setup with 19 participants, using a self-made device. Due to the used eye model and binocular cameras, the system is accurate for all distances and invariant to device movement. We also test our system against a best-in-class commercial device which is outperformed for spatial accuracy and precision. The software and hardware instructions as well as the experimental data are published as open source.
Collapse
Affiliation(s)
- Miika Toivanen
- Finnish Institute of Occupational Health, Helsinki, Finland.,University of Helsinki, Helsinki, Finland
| | | | - Kai Puolamäki
- Finnish Institute of Occupational Health, Helsinki, Finland
| |
Collapse
|
10
|
|
11
|
Lu F, Chen X. Person-independent eye gaze prediction from eye images using patch-based features. Neurocomputing 2016. [DOI: 10.1016/j.neucom.2015.07.125] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|