TY - GEN
T1 - Eye-based Continuous Affect Prediction
AU - O'Dwyer, Jonny
AU - Murray, Niall
AU - Flynn, Ronan
N1 - Publisher Copyright:
© 2019 IEEE.
PY - 2019/9
Y1 - 2019/9
N2 - Eye-based information channels include the pupils, gaze, saccades, fixational movements, and numerous forms of eye opening and closure. Pupil size variation indicates cognitive load and emotion, while a person's gaze direction is said to be congruent with the motivation to approach or avoid stimuli. The eyelids are involved in facial expressions that can encode basic emotions. Additionally, eye-based cues can have implications for human annotators of affect. Despite these facts, the use of eye-based cues in affective computing is in its infancy and this work is intended to start to address this. Eye-based feature sets, incorporating data from all of the aforementioned information channels, that can be estimated from video are proposed. Feature set refinement is provided by way of continuous arousal and valence learning and prediction experiments on the RECOLA validation set. The eye-based features are then combined with a speech feature set to provide confirmation of their usefulness and assess affect prediction performance compared with group-of-humans-level performance on the RECOLA test set. The core contribution of this paper, a refined eye-based feature set, is shown to provide benefits for affect prediction. It is hoped that this work stimulates further research into eye-based affective computing.
AB - Eye-based information channels include the pupils, gaze, saccades, fixational movements, and numerous forms of eye opening and closure. Pupil size variation indicates cognitive load and emotion, while a person's gaze direction is said to be congruent with the motivation to approach or avoid stimuli. The eyelids are involved in facial expressions that can encode basic emotions. Additionally, eye-based cues can have implications for human annotators of affect. Despite these facts, the use of eye-based cues in affective computing is in its infancy and this work is intended to start to address this. Eye-based feature sets, incorporating data from all of the aforementioned information channels, that can be estimated from video are proposed. Feature set refinement is provided by way of continuous arousal and valence learning and prediction experiments on the RECOLA validation set. The eye-based features are then combined with a speech feature set to provide confirmation of their usefulness and assess affect prediction performance compared with group-of-humans-level performance on the RECOLA test set. The core contribution of this paper, a refined eye-based feature set, is shown to provide benefits for affect prediction. It is hoped that this work stimulates further research into eye-based affective computing.
KW - Affective computing
KW - Eye closure
KW - Eye gaze
KW - Feature engineering
KW - Pupillometry
UR - http://www.scopus.com/inward/record.url?scp=85077789090&partnerID=8YFLogxK
U2 - 10.1109/ACII.2019.8925470
DO - 10.1109/ACII.2019.8925470
M3 - Conference contribution
AN - SCOPUS:85077789090
T3 - 2019 8th International Conference on Affective Computing and Intelligent Interaction, ACII 2019
BT - 2019 8th International Conference on Affective Computing and Intelligent Interaction, ACII 2019
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 8th International Conference on Affective Computing and Intelligent Interaction, ACII 2019
Y2 - 3 September 2019 through 6 September 2019
ER -