Explainable Multimodal Emotion Recognition

Multimodal emotion recognition is an important research topic in artificial intelligence, whose main goal is to integrate multimodal clues to identify human emotional states. Current works generally assume accurate labels for benchmark datasets and focus on developing more effective architectures. H...

Full description

Saved in:
Bibliographic Details
Main Authors Lian, Zheng, Sun, Haiyang, Sun, Licai, Gu, Hao, Wen, Zhuofan, Zhang, Siyuan, Chen, Shun, Xu, Mingyu, Xu, Ke, Chen, Kang, Chen, Lan, Liang, Shan, Li, Ya, Yi, Jiangyan, Liu, Bin, Tao, Jianhua
Format Journal Article
LanguageEnglish
Published 27.06.2023
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Multimodal emotion recognition is an important research topic in artificial intelligence, whose main goal is to integrate multimodal clues to identify human emotional states. Current works generally assume accurate labels for benchmark datasets and focus on developing more effective architectures. However, emotion annotation relies on subjective judgment. To obtain more reliable labels, existing datasets usually restrict the label space to some basic categories, then hire plenty of annotators and use majority voting to select the most likely label. However, this process may result in some correct but non-candidate or non-majority labels being ignored. To ensure reliability without ignoring subtle emotions, we propose a new task called ``Explainable Multimodal Emotion Recognition (EMER)''. Unlike traditional emotion recognition, EMER takes a step further by providing explanations for these predictions. Through this task, we can extract relatively reliable labels since each label has a certain basis. Meanwhile, we borrow large language models (LLMs) to disambiguate unimodal clues and generate more complete multimodal explanations. From them, we can extract richer emotions in an open-vocabulary manner. This paper presents our initial attempt at this task, including introducing a new dataset, establishing baselines, and defining evaluation metrics. In addition, EMER can serve as a benchmark task to evaluate the audio-video-text understanding performance of multimodal LLMs.
DOI:10.48550/arxiv.2306.15401