首页 | 本学科首页   官方微博 | 高级检索  
     


Multi-Modal Domain Adaptation Variational Autoencoder for EEG-Based Emotion Recognition
Authors:Yixin Wang  Shuang Qiu  Dan Li  Changde Du  Bao-Liang Lu  Huiguang He
Abstract:Traditional electroencephalograph (EEG)-based emotion recognition requires a large number of calibration samples to build a model for a specific subject, which restricts the application of the affective brain computer interface (BCI) in practice. We attempt to use the multi-modal data from the past session to realize emotion recognition in the case of a small amount of calibration samples. To solve this problem, we propose a multi-modal domain adaptive variational autoencoder (MMDA-VAE) method, which learns shared cross-domain latent representations of the multi-modal data. Our method builds a multi-modal variational autoencoder (MVAE) to project the data of multiple modalities into a common space. Through adversarial learning and cycle-consistency regularization, our method can reduce the distribution difference of each domain on the shared latent representation layer and realize the transfer of knowledge. Extensive experiments are conducted on two public datasets, SEED and SEED-IV, and the results show the superiority of our proposed method. Our work can effectively improve the performance of emotion recognition with a small amount of labelled multi-modal data. 
Keywords:Cycle-consistency   domain adaptation   electroencephalograph (EEG)   multi modality   variational autoencoder
点击此处可从《IEEE/CAA Journal of Automatica Sinica》浏览原始摘要信息
点击此处可从《IEEE/CAA Journal of Automatica Sinica》下载免费的PDF全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号