Open Access iconOpen Access

ARTICLE

crossmark

Multimodal Sentiment Analysis Based on a Cross-Modal Multihead Attention Mechanism

by Lujuan Deng, Boyi Liu*, Zuhe Li

School of Computer and Communication Engineering, Zhengzhou University of Light Industry, Zhengzhou, 450002, China

* Corresponding Author: Boyi Liu. Email: email

Computers, Materials & Continua 2024, 78(1), 1157-1170. https://doi.org/10.32604/cmc.2023.042150

Abstract

Multimodal sentiment analysis aims to understand people’s emotions and opinions from diverse data. Concatenating or multiplying various modalities is a traditional multi-modal sentiment analysis fusion method. This fusion method does not utilize the correlation information between modalities. To solve this problem, this paper proposes a model based on a multi-head attention mechanism. First, after preprocessing the original data. Then, the feature representation is converted into a sequence of word vectors and positional encoding is introduced to better understand the semantic and sequential information in the input sequence. Next, the input coding sequence is fed into the transformer model for further processing and learning. At the transformer layer, a cross-modal attention consisting of a pair of multi-head attention modules is employed to reflect the correlation between modalities. Finally, the processed results are input into the feedforward neural network to obtain the emotional output through the classification layer. Through the above processing flow, the model can capture semantic information and contextual relationships and achieve good results in various natural language processing tasks. Our model was tested on the CMU Multimodal Opinion Sentiment and Emotion Intensity (CMU-MOSEI) and Multimodal EmotionLines Dataset (MELD), achieving an accuracy of 82.04% and F1 parameters reached 80.59% on the former dataset.

Keywords


Cite This Article

APA Style
Deng, L., Liu, B., Li, Z. (2024). Multimodal sentiment analysis based on a cross-modal multihead attention mechanism. Computers, Materials & Continua, 78(1), 1157-1170. https://doi.org/10.32604/cmc.2023.042150
Vancouver Style
Deng L, Liu B, Li Z. Multimodal sentiment analysis based on a cross-modal multihead attention mechanism. Comput Mater Contin. 2024;78(1):1157-1170 https://doi.org/10.32604/cmc.2023.042150
IEEE Style
L. Deng, B. Liu, and Z. Li, “Multimodal Sentiment Analysis Based on a Cross-Modal Multihead Attention Mechanism,” Comput. Mater. Contin., vol. 78, no. 1, pp. 1157-1170, 2024. https://doi.org/10.32604/cmc.2023.042150



cc Copyright © 2024 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 721

    View

  • 332

    Download

  • 0

    Like

Share Link