UM  > Faculty of Science and Technology
Residential Collegefalse
Status已發表Published
Comprehensive Multisource Learning Network for Cross-Subject Multimodal Emotion Recognition
Chen, Chuangquan1; Li, Zhencheng1; Kou, Kit Ian2; Du, Jie3; Li, Chen4; Wang, Hongtao1; Vong, Chi Man5
2024-06-27
Source PublicationIEEE Transactions on Emerging Topics in Computational Intelligence
ISSN2471-285X
Abstract

Electroencephalography (EEG) signals and eye movement signals, which represent internal physiological responses and external subconscious behaviors, respectively, have been shown to be reliable indicators for recognizing emotions. However, integrating these two modalities across multiple subjects presents several challenges: 1) designing a robust consistency metric that balances the consistency and divergences between heterogeneous modalities across multiple subjects; 2) simultaneously considering intra-modality and inter-modality information across multiple subjects; and 3) overcoming individual differences among multiple subjects and generating subject-invariant representations of the multimodal fused features. To address these challenges associated with multisource data (i.e., multiple modalities and subjects), we propose a novel comprehensive multisource learning network (CMSLNet) for cross-subject multimodal emotion recognition. Specifically, an instance-level adaptive robust consistency metric is first designed to better align the information between EEG signals and eye movement signals, identifying their consistency and divergences across various emotions. Subsequently, an attentive low-rank multimodal fusion (Att-LMF) method is developed to account for individual differences and dynamically learn intra-modality and inter-modality information, resulting in highly discriminative fused features. Finally, domain generalization is utilized to extract subject-invariant representations of the fused features, thus adapting to new subjects and enhancing the model's generalization. Through these elaborate designs, CMSLNet effectively incorporates the information from multisource data, thus significantly improving the accuracy and reliability of emotion recognition. Extensive experiments on two public datasets demonstrate the superior performance of CMSLNet. CMSLNet achieves high accuracies of 83.15% on the SEED-IV dataset and 87.32% on the SEED-V dataset, surpassing the state-of-the-art methods by 3.62% and 4.60%, respectively.

KeywordMultisource Learning Multimodal Emotion Recognition Cross-subject Low-rank Multimodal Fusion Domain Generalization
DOI10.1109/TETCI.2024.3406422
URLView the original
Indexed BySCIE
Language英語English
WOS Research AreaComputer Science
WOS SubjectComputer Science, Artificial Intelligence
WOS IDWOS:001258759300001
PublisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC, 445 HOES LANE, PISCATAWAY, NJ 08855-4141
Scopus ID2-s2.0-85197038316
Fulltext Access
Citation statistics
Document TypeJournal article
CollectionFaculty of Science and Technology
DEPARTMENT OF MATHEMATICS
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorKou, Kit Ian; Vong, Chi Man
Affiliation1.School of Electronics and Information Engineering, Wuyi University, Jiangmen, China
2.Department of Mathematics, University of Macau, Macao, China
3.School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen University, Shenzhen, China
4.School of Mathematics and Big Data, Foshan University, Foshan, China
5.Department of Computer and Information Science, University of Macau, Macao, China
Corresponding Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Chen, Chuangquan,Li, Zhencheng,Kou, Kit Ian,et al. Comprehensive Multisource Learning Network for Cross-Subject Multimodal Emotion Recognition[J]. IEEE Transactions on Emerging Topics in Computational Intelligence, 2024.
APA Chen, Chuangquan., Li, Zhencheng., Kou, Kit Ian., Du, Jie., Li, Chen., Wang, Hongtao., & Vong, Chi Man (2024). Comprehensive Multisource Learning Network for Cross-Subject Multimodal Emotion Recognition. IEEE Transactions on Emerging Topics in Computational Intelligence.
MLA Chen, Chuangquan,et al."Comprehensive Multisource Learning Network for Cross-Subject Multimodal Emotion Recognition".IEEE Transactions on Emerging Topics in Computational Intelligence (2024).
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Chen, Chuangquan]'s Articles
[Li, Zhencheng]'s Articles
[Kou, Kit Ian]'s Articles
Baidu academic
Similar articles in Baidu academic
[Chen, Chuangquan]'s Articles
[Li, Zhencheng]'s Articles
[Kou, Kit Ian]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Chen, Chuangquan]'s Articles
[Li, Zhencheng]'s Articles
[Kou, Kit Ian]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.