Abstract
Emotion recognition in affective brain-computer interfaces (aBCI) has emerged as a prominent research area. However, existing experimental paradigms for collecting emotional data often rely on stimuli-based elicitation, which may not accurately reflect emotions experienced in everyday life. Moreover, these paradigms are limited in terms of stimulus types and lack investigation into decoding naturalistic emotional states. To address these limitations, we propose a novel experimental paradigm that enables the recording of physiological signals in a more natural way. In our approach, emotions are allowed to arise spontaneously, unrestricted by specific experimental activities. Participants have the autonomy to determine the start and end of each recording session and provide corresponding emotion label. Over a period of three months, we recruited six subjects and collected data through multiple recording sessions per subject. We utilized electroencephalogram (EEG) and eye movement signals in both subject-dependent and cross-subject settings. In the subject-dependent unimodal condition, our attentive simple graph convolutional network (ASGC) achieved the highest accuracy of 76.32% for emotion recognition based on EEG data. For the cross-subject unimodal condition, our domain adversarial neural network (DANN) outperformed other models, achieving an average accuracy of 71.90% based on EEG data. These experimental results demonstrate the feasibility of recognizing emotions in naturalistic settings. The proposed experimental paradigm holds significant potential for advancing emotion recognition in various practical applications. By allowing emotions to unfold naturally, our approach enables the future emergence of more robust and applicable emotion recognition models in the field of aBCI.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Faust, O., Hagiwara, Y., Hong, T.J., Lih, O.S., Acharya, U.R.: Deep learning for healthcare applications based on physiological signals: a review. Comput. Methods Programs Biomed. 161, 1–13 (2018)
Brunner, C., et al.: BNCI horizon 2020: towards a roadmap for the BCI community. Brain-Comput. Interfaces 2(1), 1–10 (2015)
Bijanzadeh, M., et al.: Decoding naturalistic affective behaviour from spectro-spatial features in multiday human iEEG. Nat. Hum. Behav. 6(6), 823–836 (2022)
Lu, Y., Zheng, W., Li, B., Lu, B.: Combining eye movements and EEG to enhance emotion recognition. In: Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence (IJCAI), pp. 1170–1176 (2015)
Petrantonakis, P.C., Hadjileontiadis, L.J.: Emotion recognition from EEG using higher order crossings. IEEE Trans. Inf. Technol. Biomed. 14(2), 186–197 (2009)
Zheng, W.L., Lu, B.L.: Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Trans. Auton. Ment. Dev. 7(3), 162–175 (2015)
Kensinger, E.A.: Remembering emotional experiences: the contribution of valence and arousal. Rev. Neurosci. 15(4), 241–252 (2004)
Plutchik, R.: Emotion. A Psychoevolutionary Synthesis. Harper & Row, New York (1980)
Kollias, D., et al.: Deep affect prediction in-the-wild: aff-wild database and challenge, deep architectures, and beyond. Int. J. Comput. Vision 127(6), 907–929 (2019)
Duan, R.N., Zhu, J.Y., Lu, B.L.: Differential entropy feature for EEG-based emotion classification. In: 2013 6th International IEEE/EMBS Conference on Neural Engineering (NER), pp. 81–84. IEEE (2013)
Jiang, W.B., Zhao, L.M., Guo, P., Lu, B.L.: Discriminating surprise and anger from EEG and eye movements with a graph network. In: 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 1353–1357. IEEE (2021)
Liu, L.-D., Li, R., Liu, Y.-Z., Li, H.-L., Lu, B.-L.: EEG-based human decision confidence measurement using graph neural networks. In: Mantoro, T., Lee, M., Ayu, M.A., Wong, K.W., Hidayanto, A.N. (eds.) ICONIP 2021. CCIS, vol. 1517, pp. 291–298. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-92310-5_34
Wang, Y., Jiang, W.B., Li, R., Lu, B.L.: Emotion transformer fusion: complementary representation properties of EEG and eye movements on recognizing anger and surprise. In: 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 1575–1578. IEEE (2021)
Liu, W., Zheng, W.L., Li, Z., Wu, S.Y., Gan, L., Lu, B.L.: Identifying similarities and differences in emotion recognition with EEG and eye movements among Chinese, German, and French people. J. Neural Eng. 19(2), 026012 (2022)
Ganin, Y., et al.: Domain-adversarial training of neural networks. J. Mach. Learn. Res. 17(1), 2030–2096 (2016)
Zhao, L.M., Li, R., Zheng, W.L., Lu, B.L.: Classification of five emotions from EEG and eye movement signals: complementary representation properties. In: 2019 9th International IEEE/EMBS Conference on Neural Engineering (NER), pp. 611–614. IEEE (2019)
Zhao, L.M., Yan, X., Lu, B.L.: Plug-and-play domain adaptation for cross-subject EEG-based emotion recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 863–870 (2021)
Zhang, J.-M., et al.: A cross-subject and cross-modal model for multimodal emotion recognition. In: Mantoro, T., Lee, M., Ayu, M.A., Wong, K.W., Hidayanto, A.N. (eds.) ICONIP 2021. CCIS, vol. 1517, pp. 203–211. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-92310-5_24
Acknowledgments
This work was supported in part by grants from National Natural Science Foundation of China (Grant No. 61976135), STI 2030-Major Projects+2022ZD0208500, Shanghai Municipal Science and Technology Major Project (Grant No. 2021SHZDZX), Shanghai Pujiang Program (Grant No. 22PJ1408600), Medical-Engineering Interdisciplinary Research Foundation of Shanghai Jiao Tong University “Jiao Tong Star” Program (YG2023ZD25), and GuangCi Professorship Program of RuiJin Hospital Shanghai Jiao Tong University School of Medicine.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Zhang, JM. et al. (2024). Naturalistic Emotion Recognition Using EEG and Eye Movements. In: Luo, B., Cheng, L., Wu, ZG., Li, H., Li, C. (eds) Neural Information Processing. ICONIP 2023. Lecture Notes in Computer Science, vol 14449. Springer, Singapore. https://doi.org/10.1007/978-981-99-8067-3_20
Download citation
DOI: https://doi.org/10.1007/978-981-99-8067-3_20
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-8066-6
Online ISBN: 978-981-99-8067-3
eBook Packages: Computer ScienceComputer Science (R0)