1. Machine learning-based classification of viewing behavior using a wide range of statistical oculomotor features
- Author
-
Kootstra, Timo, Teuwen, Jonas, Goudsmit, Jeroen, Nijboer, Tanja, Dodd, Michael, Van der Stigchel, Stefan, Leerstoel Stigchel, LS Logica en grondslagen v.d. wiskunde, Helmholtz Institute, Experimental Psychology (onderzoeksprogramma PF), Leerstoel Postma, Afd Psychologische functieleer, VU SBE Executive Education, Leerstoel Stigchel, LS Logica en grondslagen v.d. wiskunde, Helmholtz Institute, Experimental Psychology (onderzoeksprogramma PF), Leerstoel Postma, and Afd Psychologische functieleer
- Subjects
Eye Movements ,Computer science ,Machine learning ,computer.software_genre ,Article ,050105 experimental psychology ,Machine Learning ,03 medical and health sciences ,Cognition ,0302 clinical medicine ,fixations ,Relative magnitude ,Humans ,features ,0501 psychology and cognitive sciences ,Invariant (mathematics) ,eye movement ,Feature ranking ,business.industry ,logistic regression ,05 social sciences ,Pupil size ,saccades ,Sensory Systems ,Women's cancers Radboud Institute for Health Sciences [Radboudumc 17] ,Ophthalmology ,Logistic Models ,classification ,Eye tracking ,Artificial intelligence ,business ,computer ,Classifier (UML) ,random forest ,030217 neurology & neurosurgery ,Decoding methods - Abstract
Contains fulltext : 225885.pdf (Publisher’s version ) (Open Access) Since the seminal work of Yarbus, multiple studies have demonstrated the influence of task-set on oculomotor behavior and the current cognitive state. In more recent years, this field of research has expanded by evaluating the costs of abruptly switching between such different tasks. At the same time, the field of classifying oculomotor behavior has been moving toward more advanced, data-driven methods of decoding data. For the current study, we used a large dataset compiled over multiple experiments and implemented separate state-of-the-art machine learning methods for decoding both cognitive state and task-switching. We found that, by extracting a wide range of oculomotor features, we were able to implement robust classifier models for decoding both cognitive state and task-switching. Our decoding performance highlights the feasibility of this approach, even invariant of image statistics. Additionally, we present a feature ranking for both models, indicating the relative magnitude of different oculomotor features for both classifiers. These rankings indicate a separate set of important predictors for decoding each task, respectively. Finally, we discuss the implications of the current approach related to interpreting the decoding results.
- Published
- 2020