Electronic Science and Technology ›› 2024, Vol. 37 ›› Issue (10): 81-87.doi: 10.16180/j.cnki.issn1007-7820.2024.10.011

Previous Articles     Next Articles

Emotion Recognition Algorithm Based on Multimodal Cross-Interaction

ZHANG Hui, LI Feifei   

  1. School of Optical-Electrical and Computer Engineering,University of Shanghai for Science and Technology, Shanghai 200093,China
  • Received:2023-03-10 Online:2024-10-15 Published:2024-11-04
  • Supported by:
    The Program for Professor of Special Appointment(Eastern Scholar) at Shanghai Institutions of Higher Learning(ES2015XX)

Abstract:

Due to the limitations of single modality emotion recognition, many researchers have shifted their focus to the field of multimodal emotion recognition. Multi-modal emotion recognition focuses on two problems: The optimal extraction of the features of each mode and the effective fusion of the extracted features. This study proposes an emotion recognition method based on multimodal cross-interaction to capture the diversity of modality expressions. The editors of various modalities separately extract features with emotional information, and the stacked interaction modules based on the attention mechanism between modalities model the potential relationship among vision, text and audio. Experiments are conducted on CMU-MOSI and CMU-MOSEI datasets for emotion recognition based on text, audio and visual. The results show that the method achieved the scores of 86.5%, 47.7%, 86.4%, 0.718, 0.776, and 83.4%, 51.5%, 83.4%, 0.566, 0.737 on five indicators, Acc2(Accuracy2)、Acc7(Accuracy7)、F1、MAE(Mean Absolute Error) and Corr(Correlation). This demonstrates that the proposed algorithm significantly improves performance, and also validates that the cross-mapping mutual representation mechanism perform better than single-modal representation methods.

Key words: multimodal, feature fusion, emotion recognition, emotion analysis, attention mechanism, transformer, bidirectional encoder representation from transformers, interactive mapping

CLC Number: 

  • TP391.41