Magnetic resonance-based eye tracking can detect subject arousal state

Poster No:

1431 

Submission Type:

Abstract Submission 

Authors:

Rongjie Hu1, Jie Liang1, Yiwen Ding2, Shuang Jian2, Xiuwen Wu1, Yanming Wang1, Zhen Liang2, Bensheng Qiu1, Xiaoxiao Wang1

Institutions:

1University of Science and Technology of China, Hefei, China, 2Anhui Medical University, Hefei, China

First Author:

Rongjie Hu  
University of Science and Technology of China
Hefei, China

Co-Author(s):

Jie Liang  
University of Science and Technology of China
Hefei, China
Yiwen Ding  
Anhui Medical University
Hefei, China
Shuang Jian  
Anhui Medical University
Hefei, China
Xiuwen Wu  
University of Science and Technology of China
Hefei, China
Yanming Wang  
University of Science and Technology of China
Hefei, China
Zhen Liang  
Anhui Medical University
Hefei, China
Bensheng Qiu  
University of Science and Technology of China
Hefei, China
Xiaoxiao Wang  
University of Science and Technology of China
Hefei, China

Introduction:

Eye movements directly reflect human thoughts and play key roles in cognitive researches (Klein & Ettinger, 2019). fMRI combined with eye tracking provide a window into brain cognition and disease diagnosis (LaConte, Glielmi, Heberlein, & Hu, 2007). However, conventional eye tracking methods reliant on MRI-compatible equipment are expensive and intricate. Frey et al. designed an exciting CNN model 'DeepMReye' which performed camera-free eye tracking (Frey, Nau, & Doeller, 2021). But DeepMReye and our previous work (Wu et al., 2023) operate at interval of time of repetition (TR, in seconds), unable to fast eye movements such as blink and saccade. Therefore, we propose an eye-movement deep learning prediction pipeline, MRGazerII, which reads in fMRI slices and reports the eye movements at the interval of tens of milliseconds.

Methods:

Our study employed data from the Human Connectome Project (HCP) 7T Movie release (Van Essen et al., 2013), in which subjects watched four movies with eye tracking (416 fMRI runs from 158 subjects). The eye balls were segmented by a binary morphology method, and the intermediate 6-layer slices of the eye balls were fed into the deep model. The model comprised a ResNet-CBAM backbone, a Transformer encoder and a fully connected layer for outputting slice-level eye movements (Fig 1a). Cross-individual 5-fold cross-validation was executed.
Preprocessing procedures, including artifact removal, registration, and regression of 12 motion parameters and signals in ventricles (Jo, Saad, Simmons, Milbury, & Cox, 2010), were applied to HCP 7T Rest1 dataset. Poudel, Innes, Bones, Watts, and Jones (2014) has found the arousal state was correlated with long blink, thus, the subjects with eye tracking data (N=131) were separated into the high arousal group (long blink proportion < 5%, N=73) and low arousal group (long blink proportion > 5%, N=58, Fig 2a). Schaefer et al. (2017) 300 parcellations were then used to generate functional connectivity (FC).
FCs-based behavioral prediction was also applied for predicting subjects list sorting score. In the feature selection step, Pearson correlation was performed between each edge and list sorting score across subjects in the training set, with top 3000 FCs being selected. Principal component analysis reduced the selected 3000 FCs in training set to 30 dimensions. Linear support vector machine (SVM) regression was used for prediction, with 10-fold cross validation. Beyond training and validating in all the subjects, we trained and validated the behavior prediction in high arousal group, then tested it in the low arousal group, and vice versal.
Supporting Image: Fig1.JPG
   ·Fig 1. Model for eye movement time series classification (a), a sample illustration (b) and the model’s performance (c).
 

Results:

The proposed model achieves an f1-score of 51.2% in classifying eye movements with f1-scores of 63%, 45%, 28% and 43% for the fixation, blink, saccade and long blink respectively, surpassing the chance level (Fig 1c). Cross-dataset testing (training on HCP 7T Movie and Wu et al. (2023)'s dataset, testing on Zhang and Naya (2022)'s) reports an averaged accuracy of 38.0% (chance level at 26.0%). FC analysis reveals significantly reduced FCs in the high arousal group compared to the low arousal group (FDR corrected p < 0.01, Fig 2d). The FC-based behavior prediction achieves a Pearson's r=0.26 using data from all subjects. While training and validating on the high arousal group, the r increases to 0.38, but fails in the low arousal group (r<0.1). Training and validating on the low arousal group resulting in r=0.42, but fails in the high arousal group (r<0.1).
Supporting Image: Fig2.JPG
   ·Fig 2. The histogram of proportion of long blinks (a); the FC of high arousal group (b) and low arousal group (c); group difference (d).
 

Conclusions:

The proposed method, MRGazerII, can predict fixation, blink, long blink, saccade and fixation point in an acceptable accuracy at a fine interval of tens of milliseconds from fMRI data, both across individual and datasets. Eye movement based arousal grouping shows that the brain–phenotype model trained in one arousal state fails on another arousal state. This study underscores the indispensable role of eye tracking in understanding fMRI data and introduces MRGazerII as a solution.

Modeling and Analysis Methods:

Classification and Predictive Modeling 1
Methods Development
Task-Independent and Resting-State Analysis 2

Keywords:

Computing
Data analysis
FUNCTIONAL MRI
Machine Learning
NORMAL HUMAN
Vision

1|2Indicates the priority used for review

Provide references using author date format

Frey, M., Nau, M., & Doeller, C. F. (2021). Magnetic resonance-based eye tracking using deep neural networks. Nat Neurosci, 24(12), 1772-1779. doi:10.1038/s41593-021-00947-w
Jo, H. J., Saad, Z. S., Simmons, W. K., Milbury, L. A., & Cox, R. W. (2010). Mapping sources of correlation in resting state FMRI, with artifact detection and removal. Neuroimage, 52(2), 571-582. doi:https://doi.org/10.1016/j.neuroimage.2010.04.246
Klein, C., & Ettinger, U. (2019). Eye movement research: An introduction to its scientific foundations and applications: Springer Nature.
LaConte, S., Glielmi, C., Heberlein, K., & Hu, X. (2007). Verifying visual fixation to improve fMRI with predictive eye estimation regression (PEER). Paper presented at the Proc. Intl. Soc. Magn. Reson. Med.
Poudel, G. R., Innes, C. R. H., Bones, P. J., Watts, R., & Jones, R. D. (2014). Losing the struggle to stay awake: Divergent thalamic and cortical activity during microsleeps. Human Brain Mapping, 35(1), 257-269. doi:10.1002/hbm.22178
Schaefer, A., Kong, R., Gordon, E. M., Laumann, T. O., Zuo, X.-N., Holmes, A. J., . . . Yeo, B. T. T. (2017). Local-Global Parcellation of the Human Cerebral Cortex from Intrinsic Functional Connectivity MRI. Cereb Cortex, 28(9), 3095-3114. doi:10.1093/cercor/bhx179
Van Essen, D. C., Smith, S. M., Barch, D. M., Behrens, T. E. J., Yacoub, E., Ugurbil, K., & Consortium, W. U.-M. H. (2013). The WU-Minn Human Connectome Project: An overview. Neuroimage, 80, 62-79. doi:10.1016/j.neuroimage.2013.05.041
Wu, X., Hu, R., Liang, J., Wang, Y., Qiu, B., & Wang, X. (2023). MRGazer: Decoding Eye Gaze Points from Functional Magnetic Resonance Imaging in Individual Space. arXiv. doi:10.48550/arXiv.2311.13372
Zhang, B., & Naya, Y. (2022). A dataset of human fMRI/MEG experiments with eye tracking for spatial memory research using virtual reality. Data in Brief, 43, 108380. doi:https://doi.org/10.1016/j.dib.2022.108380