1
|
Hakim A, Zubak I, Marx C, Rhomberg T, Maragkou T, Slotboom J, Murek M. Feasibility of using Gramian angular field for preprocessing MR spectroscopy data in AI classification tasks: Differentiating glioblastoma from lymphoma. Eur J Radiol 2025; 184:111957. [PMID: 39892374 DOI: 10.1016/j.ejrad.2025.111957] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2024] [Revised: 01/13/2025] [Accepted: 01/28/2025] [Indexed: 02/03/2025]
Abstract
OBJECTIVES To convert 1D spectra into 2D images using the Gramian angular field, to be used as input for convolutional neural network for classification tasks such as glioblastoma versus lymphoma. MATERIALS AND METHODS Retrospective study including patients with histologically confirmed glioblastoma and lymphoma between 2009-2020 who underwent preoperative MR spectroscopy, using single voxel spectroscopy acquired with a short echo time (TE 30). We compared: 1) the Fourier-transformed raw spectra, and 2) the fitted spectra generated during post-processing. Both spectra were independently converted into images using the Gramian angular field, and then served as inputs for a pretrained neural network. We compared the classification performance using data from the Fourier-transformed raw spectra and the post-processed fitted spectra. RESULTS This feasibility study included 98 patients, of whom 65 were diagnosed with glioblastomas and 33 with lymphomas. For algorithm testing, 20 % of the cases (19 in total) were randomly selected. By applying the Gramian angular field technique to the Fourier-transformed spectra, we achieved an accuracy of 73.7 % and a sensitivity of 92 % in classifying glioblastoma versus lymphoma, slightly higher than the fitted spectra pathway. CONCLUSION Spectroscopy data can be effectively transformed into distinct color graphical images using the Gramian angular field technique, enabling their use as input for deep learning algorithms. Accuracy tends to be higher when utilizing data derived from Fourier-transformed spectra compared to fitted spectra. This finding underscores the potential of using MR spectroscopy data in neural network-based classification purposes.
Collapse
Affiliation(s)
- Arsany Hakim
- University Institute of Diagnostic and Interventional Neuroradiology, Bern University Hospital, Inselspital, University of Bern, Bern, Switzerland.
| | - Irena Zubak
- Department of Neurosurgery, Bern University Hospital, Inselspital, University of Bern, Bern, Switzerland
| | - Christina Marx
- University Institute of Diagnostic and Interventional Neuroradiology, Bern University Hospital, Inselspital, University of Bern, Bern, Switzerland
| | - Thomas Rhomberg
- Department of Neurosurgery, Bern University Hospital, Inselspital, University of Bern, Bern, Switzerland
| | - Theoni Maragkou
- Institute of Tissue Medicine and Pathology, University of Bern, Bern, Switzerland
| | - Johannes Slotboom
- University Institute of Diagnostic and Interventional Neuroradiology, Bern University Hospital, Inselspital, University of Bern, Bern, Switzerland
| | - Michael Murek
- Department of Neurosurgery, Bern University Hospital, Inselspital, University of Bern, Bern, Switzerland
| |
Collapse
|
2
|
Wang C, Wang R, Leng Y, Iramina K, Yang Y, Ge S. An Eye Movement Classification Method Based on Cascade Forest. IEEE J Biomed Health Inform 2024; 28:7184-7194. [PMID: 39106144 DOI: 10.1109/jbhi.2024.3439568] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/09/2024]
Abstract
Eye tracking technology has become increasingly important in scientific research and practical applications. In the field of eye tracking research, analysis of eye movement data is crucial, particularly for classifying raw eye movement data into eye movement events. Current classification methods exhibit considerable variation in adaptability across different participants, and it is necessary to address the issues of class imbalance and data scarcity in eye movement classification. In the current study, we introduce a novel eye movement classification method based on cascade forest (EMCCF), which comprises two modules: 1) a feature extraction module that employs a multi-scale time window method to extract features from raw eye movement data; 2) a classification module that innovatively employs a layered ensemble architecture, integrating the cascade forest structure with ensemble learning principles, specifically for eye movement classification. Consequently, EMCCF not only enhanced the accuracy and efficiency of eye movement classification but also represents an advancement in applying ensemble learning techniques within this domain. Furthermore, experimental results indicated that EMCCF outperformed existing deep learning-based classification models in several metrics and demonstrated robust performance across different datasets and participants.
Collapse
|
3
|
Yang H, Wei X, Huang K, Wu Z, Zhang Q, Wen S, Wang Q, Feng L. Features of attention network impairment in patients with temporal lobe epilepsy: Evidence from eye-tracking and electroencephalogram. Epilepsy Behav 2024; 157:109887. [PMID: 38905916 DOI: 10.1016/j.yebeh.2024.109887] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/07/2024] [Revised: 06/03/2024] [Accepted: 06/08/2024] [Indexed: 06/23/2024]
Abstract
AIM To explore multiple features of attention impairments in patients with temporal lobe epilepsy (TLE). METHODS A total of 93 patients diagnosed with TLE at Xiangya Hospital during May 2022 and December 2022 and 85 healthy controls were included in this study. Participants were asked to complete neuropsychological scales and attention network test (ANT) with recording of eye-tracking and electroencephalogram. RESULTS All means of evaluation showed impaired attention functions in TLE patients. ANT results showed impaired orienting (p < 0.001) and executive control (p = 0.041) networks. Longer mean first saccade time (p = 0.046) and more total saccadic counts (p = 0.035) were found in eye-tracking results, indicating abnormal alerting and orienting networks. Both alerting, orienting and executive control networks were abnormal, manifesting as decreased amplitudes (N1 & P3, p < 0.001) and extended latency (P3, p = 0.002). The energy of theta, alpha and beta were all sensitive to the changes of alerting and executive control network with time, but only beta power was sensitive to the changes of orienting network. CONCLUSION Our findings are helpful for early identification of patients with TLE combined with attention impairments, which have strong clinical guiding significance for long-term monitoring and intervention.
Collapse
Affiliation(s)
- Haojun Yang
- Department of Anesthesiology, Xiangya Hospital of Central South University, Changsha, Hunan, China
| | - Xiaojie Wei
- Key Laboratory of Biomedical Spectroscopy of Xi'an, Xi'an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi'an 710119, China; Key Laboratory of Spectral Imaging Technology, Xi'an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi'an 710119, China; University of Chinese Academy of Sciences, Beijing 101400, China
| | - Kailing Huang
- Department of Neurology, Xiangya Hospital of Central South University, Changsha, Hunan, China; National Clinical Research Center for Geriatric Disorders, Xiangya Hospital, Central South University, Changsha, China
| | - Zhongling Wu
- Department of Neurology, Xiangya Hospital of Central South University, Changsha, Hunan, China; Department of Clinical Nursing Teaching and Research Section, Xiangya Hospital, Central South University, Changsha, China
| | - Qiong Zhang
- Department of Neurology, Xiangya Hospital of Central South University, Changsha, Hunan, China; National Clinical Research Center for Geriatric Disorders, Xiangya Hospital, Central South University, Changsha, China
| | - Shirui Wen
- Department of Neurology, Xiangya Hospital of Central South University, Changsha, Hunan, China; National Clinical Research Center for Geriatric Disorders, Xiangya Hospital, Central South University, Changsha, China
| | - Quan Wang
- Key Laboratory of Biomedical Spectroscopy of Xi'an, Xi'an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi'an 710119, China; Key Laboratory of Spectral Imaging Technology, Xi'an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi'an 710119, China.
| | - Li Feng
- Department of Neurology, Xiangya Hospital of Central South University, Changsha, Hunan, China; National Clinical Research Center for Geriatric Disorders, Xiangya Hospital, Central South University, Changsha, China.
| |
Collapse
|
4
|
Korda Ž, Walcher S, Körner C, Benedek M. Decoupling of the pupillary light response during internal attention: The modulating effect of luminance intensity. Acta Psychol (Amst) 2024; 242:104123. [PMID: 38181698 DOI: 10.1016/j.actpsy.2023.104123] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/11/2023] [Revised: 11/30/2023] [Accepted: 12/21/2023] [Indexed: 01/07/2024] Open
Abstract
In a world full of sensory stimuli, attention guides us between the external environment and our internal thoughts. While external attention involves processing sensory stimuli, internal attention is devoted to self-generated representations such as planning or spontaneous mind wandering. They both draw from common cognitive resources, thus simultaneous engagement in both often leads to interference between processes. In order to maintain internal focus, an attentional mechanism known as perceptual decoupling takes effect. This mechanism supports internal cognition by decoupling attention from the perception of sensory information. Two previous studies of our lab investigated to what extent perceptual decoupling is evident in voluntary eye movements. Findings showed that the effect is mediated by the internal task modality and workload (visuospatial > arithmetic and high > low, respectively). However, it remains unclear whether it extends to involuntary eye behavior, which may not share cognitive resources with internal activities. Therefore, the present experiment aimed to further elucidate attentional dynamics by examining whether internal attention affects the pupillary light response (PLR). Specifically, we consistently observed that workload and task modality of the internal task reduced the PLR to luminance changes of medium intensity. However, the PLR to strong luminance changes was less or not at all affected by the internal task. These results suggest that perceptual decoupling effects may be less consistent in involuntary eye behavior, particularly in the context of a salient visual stimulus.
Collapse
Affiliation(s)
- Živa Korda
- Department of Psychology, University of Graz, Graz, Austria.
| | - Sonja Walcher
- Department of Psychology, University of Graz, Graz, Austria
| | | | | |
Collapse
|
5
|
Vortmann LM, Weidenbach P, Putze F. AtAwAR Translate: Attention-Aware Language Translation Application in Augmented Reality for Mobile Phones. SENSORS (BASEL, SWITZERLAND) 2022; 22:6160. [PMID: 36015922 PMCID: PMC9412445 DOI: 10.3390/s22166160] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/31/2022] [Revised: 08/02/2022] [Accepted: 08/12/2022] [Indexed: 06/15/2023]
Abstract
As lightweight, low-cost EEG headsets emerge, the feasibility of consumer-oriented brain-computer interfaces (BCI) increases. The combination of portable smartphones and easy-to-use EEG dry electrode headbands offers intriguing new applications and methods of human-computer interaction. In previous research, augmented reality (AR) scenarios have been identified to profit from additional user state information-such as that provided by a BCI. In this work, we implemented a system that integrates user attentional state awareness into a smartphone application for an AR written language translator. The attentional state of the user is classified in terms of internally and externally directed attention by using the Muse 2 electroencephalography headband with four frontal electrodes. The classification results are used to adapt the behavior of the translation app, which uses the smartphone's camera to display translated text as augmented reality elements. We present the first mobile BCI system that uses a smartphone and a low-cost EEG device with few electrodes to provide attention awareness to an AR application. Our case study with 12 participants did not fully support the assumption that the BCI improves usability. However, we are able to show that the classification accuracy and ease of setup are promising paths toward mobile consumer-oriented BCI usage. For future studies, other use cases, applications, and adaptations will be tested for this setup to explore the usability.
Collapse
|
6
|
Kuo RJ, Chen HJ, Kuo YH. The development of an eye movement-based deep learning system for laparoscopic surgical skills assessment. Sci Rep 2022; 12:11036. [PMID: 35970911 PMCID: PMC9378740 DOI: 10.1038/s41598-022-15053-5] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2022] [Accepted: 06/17/2022] [Indexed: 11/23/2022] Open
Abstract
The development of valid, reliable, and objective methods of skills assessment is central to modern surgical training. Numerous rating scales have been developed and validated for quantifying surgical performance. However, many of these scoring systems are potentially flawed in their design in terms of reliability. Eye-tracking techniques, which provide a more objective investigation of the visual-cognitive aspects of the decision-making process, recently have been utilized in surgery domains for skill assessment and training, and their use has been focused on investigating differences between expert and novice surgeons to understand task performance, identify experienced surgeons, and establish training approaches. Ten graduate students at the National Taiwan University of Science and Technology with no prior laparoscopic surgical skills were recruited to perform the FLS peg transfer task. Then k-means clustering algorithm was used to split 500 trials into three dissimilar clusters, grouped as novice, intermediate, and expert levels, by an objective performance assessment parameter incorporating task duration with error score. Two types of data sets, namely, time series data extracted from coordinates of eye fixation and image data from videos, were used to implement and test our proposed skill level detection system with ensemble learning and a CNN algorithm. Results indicated that ensemble learning and the CNN were able to correctly classify skill levels with accuracies of 76.0% and 81.2%, respectively. Furthermore, the incorporation of coordinates of eye fixation and image data allowed the discrimination of skill levels with a classification accuracy of 82.5%. We examined more levels of training experience and further integrated an eye tracking technique and deep learning algorithms to develop a tool for objective assessment of laparoscopic surgical skill. With a relatively unbalanced sample, our results have demonstrated that the approach combining the features of visual fixation coordinates and images achieved a very promising level of performance for classifying skill levels of trainees.
Collapse
Affiliation(s)
- R J Kuo
- Department of Industrial Management, National Taiwan University of Science and Technology, Taipei, Taiwan
| | - Hung-Jen Chen
- Department of Data Science, Soochow University, No. 70, Linhsi Road, Shihlin District, Taipei City, 111, Taiwan.
- Department of Marketing and Distribution Management, National Kaohsiung University of Science and Technology, No.1, University Road, Yanchao District, Kaohsiung City, 82445, Taiwan.
| | - Yi-Hung Kuo
- Department of New Product Introduction, Solid State Storage Technology Corporation, Hsinchu City, Taiwan
| |
Collapse
|
7
|
Vortmann LM, Ceh S, Putze F. Multimodal EEG and Eye Tracking Feature Fusion Approaches for Attention Classification in Hybrid BCIs. FRONTIERS IN COMPUTER SCIENCE 2022. [DOI: 10.3389/fcomp.2022.780580] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022] Open
Abstract
Often, various modalities capture distinct aspects of particular mental states or activities. While machine learning algorithms can reliably predict numerous aspects of human cognition and behavior using a single modality, they can benefit from the combination of multiple modalities. This is why hybrid BCIs are gaining popularity. However, it is not always straightforward to combine features from a multimodal dataset. Along with the method for generating the features, one must decide when the modalities should be combined during the classification process. We compare unimodal EEG and eye tracking classification of internally and externally directed attention to multimodal approaches for early, middle, and late fusion in this study. On a binary dataset with a chance level of 0.5, late fusion of the data achieves the highest classification accuracy of 0.609–0.675 (95%-confidence interval). In general, the results indicate that for these modalities, middle or late fusion approaches are better suited than early fusion approaches. Additional validation of the observed trend will require the use of additional datasets, alternative feature generation mechanisms, decision rules, and neural network designs. We conclude with a set of premises that need to be considered when deciding on a multimodal attentional state classification approach.
Collapse
|
8
|
Vortmann LM, Putze F. Combining Implicit and Explicit Feature Extraction for Eye Tracking: Attention Classification Using a Heterogeneous Input. SENSORS 2021; 21:s21248205. [PMID: 34960295 PMCID: PMC8707750 DOI: 10.3390/s21248205] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/29/2021] [Revised: 11/29/2021] [Accepted: 12/02/2021] [Indexed: 01/24/2023]
Abstract
Statistical measurements of eye movement-specific properties, such as fixations, saccades, blinks, or pupil dilation, are frequently utilized as input features for machine learning algorithms applied to eye tracking recordings. These characteristics are intended to be interpretable aspects of eye gazing behavior. However, prior research has demonstrated that when trained on implicit representations of raw eye tracking data, neural networks outperform these traditional techniques. To leverage the strengths and information of both feature sets, we integrated implicit and explicit eye tracking features in one classification approach in this work. A neural network was adapted to process the heterogeneous input and predict the internally and externally directed attention of 154 participants. We compared the accuracies reached by the implicit and combined features for different window lengths and evaluated the approaches in terms of person- and task-independence. The results indicate that combining implicit and explicit feature extraction techniques for eye tracking data improves classification results for attentional state detection significantly. The attentional state was correctly classified during new tasks with an accuracy better than chance, and person-independent classification even outperformed person-dependently trained classifiers for some settings. For future experiments and applications that require eye tracking data classification, we suggest to consider implicit data representation in addition to interpretable explicit features.
Collapse
|