Attention-based Multi-modal Sentiment Analysis and Emotion Detection in Conversation using RNN
Author | |
Keywords | |
Abstract |
The availability of an enormous quantity of multimodal data and its widespread applications, automatic sentiment analysis and emotion classification in the conversation has become an interesting research topic among the research community. The interlocutor state, context state between the neighboring utterances and multimodal fusion play an important role in multimodal sentiment analysis and emotion detection in conversation. In this article, the recurrent neural network (RNN) based method is developed to capture the interlocutor state and contextual state between the utterances. The pair-wise attention mechanism is used to understand the relationship between the modalities and their importance before fusion. First, two-two combinations of modalities are fused at a time and finally, all the modalities are fused to form the trimodal representation feature vector. The experiments are conducted on three standard datasets such as IEMOCAP, CMU-MOSEI, and CMU-MOSI. The proposed model is evaluated using two metrics such as accuracy and F1-Score and the results demonstrate that the proposed model performs better than the standard baselines.
|
Year of Publication |
2021
|
Journal |
International Journal of Interactive Multimedia and Artificial Intelligence
|
Volume |
6
|
Issue |
Regular Issue
|
Number |
6
|
Number of Pages |
112-121
|
Date Published |
06/2021
|
ISSN Number |
1989-1660
|
URL | |
DOI | |
Attachment |
ijimai_6_6_12.pdf1.45 MB
|