IEEE/CAA Journal of Automatica Sinica
Citation: | Y. X. Wang, S. Qiu, D. Li, C. D. Du, B.-L. Lu, and H. G. He, “Multi-modal domain adaptation variational autoencoder for EEG-based emotion recognition,” IEEE/CAA J. Autom. Sinica, vol. 9, no. 9, pp. 1612–1626, Sept. 2022. doi: 10.1109/JAS.2022.105515 |
[1] |
S. Spence, “Descartes’ error: Emotion, reason and the human brain,” BMJ, vol. 310, no. 6988, p. 1213, 1995.
|
[2] |
B. Shneiderman, C. Plaisant, M. Cohen, S. Jacobs, N. Elmqvist, and N. Diakopoulos, Designing the User Interface: Strategies for Effective Human-Computer Interaction. Pearson, 2016.
|
[3] |
P. Ekman, “Expression and the nature of emotion,” Approaches to emotion, vol. 3, pp. 19–344, 1984.
|
[4] |
Z. Zeng, M. Pantic, G. I. Roisman, and T. S. Huang, “A survey of affect recognition methods: Audio, visual, and spontaneous expressions,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 31, no. 1, pp. 39–58, 2009. doi: 10.1109/TPAMI.2008.52
|
[5] |
M. Coulson, “Attributing emotion to static body postures: Recognition accuracy, confusions, and viewpoint dependence,” Journal of Nonverbal Behavior, vol. 28, no. 2, pp. 117–139, 2004. doi: 10.1023/B:JONB.0000023655.25550.be
|
[6] |
H. Gunes, B. Schuller, M. Pantic, and R. Cowie, “Emotion representation, analysis and synthesis in continuous space: A survey,” in Proc. IEEE Face and Gesture, pp. 827–834, 2011.
|
[7] |
R. A. Calvo and S. D’Mello, “Affect detection: An interdisciplinary review of models, methods, and their applications,” IEEE Trans. Affective Computing, vol. 1, no. 1, pp. 18–37, 2010. doi: 10.1109/T-AFFC.2010.1
|
[8] |
F. Lotte, M. Congedo, A. Lécuyer, F. Lamarche, and B. Arnaldi, “A review of classification algorithms for EEG-based brain–computer interfaces,” Journal of Neural Engineering, vol. 4, no. 2, p. R1, 2007.
|
[9] |
J. Zhang, Z. Yin, P. Cheng, and S. Nichele, “Emotion recognition using multi-modal data and machine learning techniques: A tutorial and review,” Information Fusion, 2020.
|
[10] |
M. Soleymani, S. Asghari-Esfeden, Y. Fu, and M. Pantic, “Analysis of EEG signals and facial expressions for continuous emotion detection,” IEEE Trans. Affective Computing, vol. 7, no. 1, pp. 17–28, 2015.
|
[11] |
Y. Lu, W.-L. Zheng, B. Li, and B.-L. Lu, “Combining eye movements and EEG to enhance emotion recognition,” in Proc. 24th Int. Joint Conf. Artificial Intelligence, 2015.
|
[12] |
H. Ranganathan, S. Chakraborty, and S. Panchanathan, “Multimodal emotion recognition using deep learning architectures,” in Proc. IEEE Winter Conf. Applications of Computer Vision, pp. 1–9, 2016.
|
[13] |
C. Mühl, B. Allison, A. Nijholt, and G. Chanel, “A survey of affective brain computer interfaces: Principles, state-of-the-art, and challenges,” Brain-Computer Interfaces, vol. 1, no. 2, pp. 66–84, 2014. doi: 10.1080/2326263X.2014.912881
|
[14] |
D. H. Lee and A. K. Anderson, “Reading what the mind thinks from how the eye sees,” Psychological Science, vol. 28, no. 4, pp. 494–503, 2017. doi: 10.1177/0956797616687364
|
[15] |
M. Soleymani, M. Pantic, and T. Pun, “Multi-modal emotion recognition in response to videos,” IEEE Trans. Affective Computing, vol. 3, no. 2, pp. 211–223, 2011.
|
[16] |
K. Weiss, T. M. Khoshgoftaar, and D. Wang, “A survey of transfer learning,” Journal of Big Data, vol. 3, no. 1, p. 9, 2016.
|
[17] |
T. Baltrušaitis, C. Ahuja, and L.-P. Morency, “Multimodal machine learning: A survey and taxonomy,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 41, no. 2, pp. 423–443, 2018.
|
[18] |
E. Principi, D. Rossetti, S. Squartini, and F. Piazza, “Unsupervised electric motor fault detection by using deep autoencoders,” IEEE/CAA J. Autom. Sinica, vol. 6, no. 2, pp. 441–451, 2019. doi: 10.1109/JAS.2019.1911393
|
[19] |
L. Pang, S. Zhu, and C.-W. Ngo, “Deep multi-modal learning for affective analysis and retrieval,” IEEE Trans. Multimedia, vol. 17, no. 11, pp. 2008–2020, 2015. doi: 10.1109/TMM.2015.2482228
|
[20] |
W. Liu, W.-L. Zheng, and B.-L. Lu, “Emotion recognition using multi-modal deep learning,” in Proc. Int. Conf. Neural Information Processing, pp. 521–529, Springer, 2016.
|
[21] |
P. Tzirakis, G. Trigeorgis, M. A. Nicolaou, B. W. Schuller, and S. Zafeiriou, “End-to-end multi-modal emotion recognition using deep neural networks,” IEEE Journal of Selected Topics in Signal Processing, vol. 11, no. 8, pp. 1301–1309, 2017. doi: 10.1109/JSTSP.2017.2764438
|
[22] |
X. Kang, F. Ren, and Y. Wu, “Exploring latent semantic information for textual emotion recognition in blog articles,” IEEE/CAA J. Autom. Sinica, vol. 5, no. 1, pp. 204–216, 2017.
|
[23] |
J. Chen, X. Luo, and M. Zhou, “Hierarchical particle swarm optimization-incorporated latent factor analysis for large-scale incomplete matrices,” IEEE Trans. Big Data, 2021. DOI: 10.1109/TBDATA.2021.3090905
|
[24] |
W.-L. Zheng and B.-L. Lu, “Personalizing EEG-based affective models with transfer learning,” in Proc. 25th Int. Joint Conf. Artificial Intelligence, pp. 2732–2738, AAAI Press, 2016.
|
[25] |
P. Sinno Jialin, I. W. Tsang, J. T. Kwok, and Y. Qiang, “Domain adaptation via transfer component analysis,” IEEE Trans. Neural Networks, vol. 22, no. 2, pp. 199–210, 2011. doi: 10.1109/TNN.2010.2091281
|
[26] |
B. Schölkopf, A. Smola, and K.-R. Müller, “Nonlinear component analysis as a kernel eigenvalue problem,” Neural Computation, vol. 10, no. 5, pp. 1299–1319, 1998. doi: 10.1162/089976698300017467
|
[27] |
R. Collobert, F. Sinz, J. Weston, L. Bottou, and T. Joachims, “Large scale transductive SVMs,” Journal of Machine Learning Research, vol. 7, no. 1, pp. 1687–1712, 2006.
|
[28] |
E. Sangineto, G. Zen, E. Ricci, and N. Sebe, “We are not all equal: Personalizing models for facial expression analysis with transductive parameter transfer,” in Proc. ACM Int. Conf. Multimedia, pp. 357–366, 2014.
|
[29] |
X. Chai, Q. Wang, Y. Zhao, Y. Li, D. Liu, X. Liu, and O. Bai, “A fast, efficient domain adaptation technique for cross-domain electroencephalography (EEG)-based emotion recognition,” Sensors, vol. 17, no. 5, p. 1014, 2017.
|
[30] |
Y. Li, W. Zheng, Y. Zong, Z. Cui, T. Zhang, and X. Zhou, “A bihemisphere domain adversarial neural network model for EEG emotion recognition,” IEEE Trans. Affective Computing, vol. 12, no. 2, pp. 494–504, 2021. doi: 10.1109/TAFFC.2018.2885474
|
[31] |
J. Li, S. Qiu, C. Du, Y. Wang, and H. He, “Domain adaptation for EEG emotion recognition based on latent representation similarity,” IEEE Trans. Cognitive and Developmental Systems, vol. 12, no. 2, pp. 344–353, 2020. doi: 10.1109/TCDS.2019.2949306
|
[32] |
D. P. Kingma and M. Welling, “Auto-encoding variational bayes,” arXiv preprint arXiv: 1312.6114, 2013.
|
[33] |
T. Shen, T. Lei, R. Barzilay, and T. Jaakkola, “Style transfer from nonparallel text by cross-alignment,” in Advances in Neural Information Processing Systems, pp. 6830–6841, 2017.
|
[34] |
M.-Y. Liu, T. Breuel, and J. Kautz, “Unsupervised image-to-image translation networks,” in Advances in Neural Information Processing Systems, pp. 700–708, 2017.
|
[35] |
M. Shang, X. Luo, Z. Liu, J. Chen, Y. Yuan, and M. Zhou, “Randomized latent factor model for high-dimensional and sparse matrices from industrial applications,” IEEE/CAA J. Autom. Sinica, vol. 6, no. 1, pp. 131–141, 2018.
|
[36] |
X. Luo, Z. Liu, S. Li, M. Shang, and Z. Wang, “A fast non-negative latent factor model based on generalized momentum method,” IEEE Trans. Systems,Man,and Cybernetics: Systems, vol. 51, no. 1, pp. 610–620, 2021. doi: 10.1109/TSMC.2018.2875452
|
[37] |
D. Wu, Y. He, X. Luo, and M. Zhou, “A latent factor analysis-based approach to online sparse streaming feature selection,” IEEE Trans. Systems,Man,and Cybernetics: Systems, 2021. DOI: 10.1109/TSMC.2021.3096065
|
[38] |
M. Wu and N. Goodman, “Multi-modal generative models for scalable weakly-supervised learning,” in Advances in Neural Information Processing Systems, pp. 5575–5585, 2018.
|
[39] |
G. E. Hinton, “Training products of experts by minimizing contrastive divergence,” Neural Computation, vol. 14, no. 8, pp. 1771–1800, 2002. doi: 10.1162/089976602760128018
|
[40] |
W.-L. Zheng, W. Liu, Y. Lu, B.-L. Lu, and A. Cichocki, “Emotionmeter: A multi-modal framework for recognizing human emotions,” IEEE Trans. Cybernetics, vol. 49, no. 3, pp. 1110–1122, 2018.
|
[41] |
W.-L. Zheng, J.-Y. Zhu, and B.-L. Lu, “Identifying stable patterns over time for emotion recognition from EEG,” IEEE Trans. Affective Computing, vol. 10, no. 3, pp. 417–429, 2019. doi: 10.1109/TAFFC.2017.2712143
|
[42] |
M. S. AL-Quraishi, I. Elamvazuthi, T. B. Tang, A.-Q. Muhammad, S. Parasuraman, and A. Borboni, “Multi-modal fusion approach based on EEG and EMG signals for lower limb movement recognition,” IEEE Sensors Journal, vol. 21, no. 24, pp. 27640–37650, 2021. doi: 10.1109/JSEN.2021.3119074
|
[43] |
E. Tzeng, J. Hoffman, T. Darrell, and K. Saenko, “Simultaneous deep transfer across domains and tasks,” in Proc. IEEE Int. Conf. Computer Vision, pp. 4068–4076, 2015.
|
[44] |
M. Long, Y. Cao, J. Wang, and M. Jordan, “Learning transferable features with deep adaptation networks,” in Proc. Int. Conf. Machine Learning, pp. 97–105, 2015.
|
[45] |
Y. Ganin, E. Ustinova, H. Ajakan, P. Germain, H. Larochelle, F. Laviolette, M. Marchand, and V. Lempitsky, “Domain-adversarial training of neural networks,” The Journal of Machine Learning Research, vol. 17, no. 1, pp. 2096–2030, 2016.
|
[46] |
M. Long, H. Zhu, J. Wang, and M. I. Jordan, “Deep transfer learning with joint adaptation networks,” in Proc. 34th Int. Conf. Machine Learning-Volume 70, pp. 2208–2217, JMLR. org, 2017.
|
[47] |
P. Haeusser, T. Frerix, A. Mordvintsev, and D. Cremers, “Associative domain adaptation,” in Proc. IEEE Int. Conf. Computer Vision, pp. 2765–2773, 2017.
|
[48] |
M. Long, Z. Cao, J. Wang, and M. I. Jordan, “Conditional adversarial domain adaptation,” in Advances in Neural Information Processing Systems, pp. 1640–1650, 2018.
|
[49] |
M.-Y. Liu and O. Tuzel, “Coupled generative adversarial networks,” in Advances in Neural Information Processing Systems, pp. 469–477, 2016.
|
[50] |
C. Yu, J. Wang, Y. Chen, and M. Huang, “Transfer learning with dynamic adversarial adaptation network,” in Proc. IEEE Int. Conf. Data Mining, pp. 778–786, 2019.
|
[51] |
R. Horlings, D. Datcu, and L. J. Rothkrantz, “Emotion recognition using brain activity,” in Proc. 9th Int. Conf. Computer Systems and Technologies and Workshop for Ph.D. Students in Computing, 2008.
|
[52] |
S. Katsigiannis and N. Ramzan, “Dreamer: A database for emotion recognition through EEG and ECG signals from wireless low-cost off-the-shelf devices,” IEEE Journal of Biomedical and Health Informatics, vol. 22, no. 1, pp. 98–107, 2017.
|
[53] |
M. Liu, D. Fan, X. Zhang, and X. Gong, “Human emotion recognition based on galvanic skin response signal feature selection and SVM,” in Proc. Int. Conf. Smart City and Systems Engineering, pp. 157–160, 2016.
|