Eye fixation versus pupil diameter as eye- tracking features for virtual reality emotion classification

The usage of eye-tracking technology is becoming increasingly popular in machine learning applications, particularly in the area of affective computing and emotion recognition. Typically, emotion recognition studies utilize popular physiological signals such as electroencephalography (EEG), while th...

Full description

Saved in:
Bibliographic Details
Main Authors: Lim Jia Zheng, James Mountstephens, Jason Teo
Format: Proceedings
Language:English
English
Published: Institute of Electrical and Electronics Engineers 2022
Subjects:
Online Access:https://eprints.ums.edu.my/id/eprint/33604/2/Eye%20fixation%20versus%20pupil%20diameter%20as%20eye-%20tracking%20features%20for%20virtual%20reality%20emotion%20classification.ABSTRACT.pdf
https://eprints.ums.edu.my/id/eprint/33604/1/Eye%20fixation%20versus%20pupil%20diameter%20as%20eye-tracking%20features%20for%20virtual%20reality%20emotion%20classification.pdf
https://eprints.ums.edu.my/id/eprint/33604/
https://ieeexplore.ieee.org/document/9673503
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Universiti Malaysia Sabah
Language: English
English
Description
Summary:The usage of eye-tracking technology is becoming increasingly popular in machine learning applications, particularly in the area of affective computing and emotion recognition. Typically, emotion recognition studies utilize popular physiological signals such as electroencephalography (EEG), while the research on emotion detection that relies solely on eye-tracking data is limited. In this study, an empirical comparison of the accuracy of eye-tracking-based emotion recognition in a virtual reality (VR) environment using eye fixation versus pupil diameter as the classification feature is performed. We classified emotions into four distinct classes according to Russell’s four-quadrant Circumplex Model of Affect. 3600 videos are presented as emotional stimuli to participants in a VR environment to evoke the user’s emotions. Three separate experiments were conducted using Support Vector Machines (SVMs) as the classification algorithm for the two chosen eye features. The results showed that emotion classification using fixation position obtained an accuracy of 75% while pupil diameter obtained an accuracy of 57%. For four-quadrant emotion recognition, eye fixation as a learning feature produces better classification accuracy compared to pupil diameter. Therefore, this empirical study has shown that eyetracking- based emotion recognition systems would benefit from using features based on eye fixation data rather than pupil size.