Trimodal Attention Module for Multimodal Sentiment Analysis
| dc.contributor.author | Harish, A.B. | |
| dc.contributor.author | Sadat, F. | |
| dc.date.accessioned | 2026-02-06T06:36:55Z | |
| dc.date.issued | 2020 | |
| dc.description.abstract | In our research, we propose a new multimodal fusion architecture for the task of sentiment analysis. The 3 modalities used in this paper are text, audio and video. Most of the current methods deal with either a feature level or a decision level fusion. In contrast, we propose an attention-based deep neural network and a training approach to facilitate both feature and decision level fusion. Our network effectively leverages information across all three modalities using a 2 stage fusion process. We test our network on the individual utterance based contextual information extracted from the CMUMOSI Dataset. A comparison is drawn between the state-ofthe- A rt and our network. © 2020 The Twenty-Fifth AAAI/SIGAI Doctoral Consortium (AAAI-20). All Rights Reserved. | |
| dc.identifier.citation | AAAI 2020 - 34th AAAI Conference on Artificial Intelligence, 2020, Vol., , p. 13803-13804 | |
| dc.identifier.uri | https://doi.org/ | |
| dc.identifier.uri | https://idr.nitk.ac.in/handle/123456789/30761 | |
| dc.publisher | AAAI press | |
| dc.title | Trimodal Attention Module for Multimodal Sentiment Analysis |
