[1]PAN Jiahui,HE Zhipeng,LI Zina,et al.A review of multimodal emotion recognition[J].CAAI Transactions on Intelligent Systems,2020,15(4):633-645.[doi:10.11992/tis.202001032]
Copy
CAAI Transactions on Intelligent Systems[ISSN 1673-4785/CN 23-1538/TP] Volume:
15
Number of periods:
2020 4
Page number:
633-645
Column:
综述
Public date:
2020-07-05
- Title:
-
A review of multimodal emotion recognition
- Author(s):
-
PAN Jiahui1; HE Zhipeng1; LI Zina2; LIANG Yan1; QIU Lina1
-
1. School of Software, South China Normal University, Foshan 528225, China;
2. School of Computer, South China Normal University, Guangzhou 510641, China
-
- Keywords:
-
emotion recognition; emotion description model; emotion inducing mode; information fusion; fusion strategy; emotion representation; modality blend
- CLC:
-
TP391.4
- DOI:
-
10.11992/tis.202001032
- Abstract:
-
This paper reviews the emerging field of multimodal emotion recognition. Firstly, the research foundation of emotion recognition is summarized from two aspects: emotion description model and emotion-inducing mode. Then, aiming at the key and difficult problem of information fusion in multi-modal emotion recognition, some mainstream and high-efficiency information fusion strategies are introduced from four fusion levels: data-level fusion, feature-level fusion, decision-level fusion, and model-level fusion. By exemplifying representative multi-modal mixing examples from three perspectives: the mixing of multiple external presentation modalities, the mixing of multiple neurophysiological modalities, and the mixing of neurophysiology and external presentation modalities, it fully demonstrates that multi-modality is more capable of emotional discrimination and emotional representation than single-modality. At the same time, some thoughts on the conversion of multi-modal recognition methods to engineering technology applications are put forward. Finally, based on the analysis and grasp of the current situation of emotion recognition research, the ways and strategies for improving and enhancing the performance of the emotion recognition models are discussed and prospected.