UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

MAG+: An Extended Multimodal Adaptation Gate for Multimodal Sentiment Analysis

Zhao, X; Chen, Y; Li, W; Gao, L; Tang, B; (2022) MAG+: An Extended Multimodal Adaptation Gate for Multimodal Sentiment Analysis. In: ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings. (pp. pp. 4753-4757). Institute of Electrical and Electronics Engineers (IEEE) Green open access

[thumbnail of MAG___An_EXTENDED_Multimodal_Adaptation_Gate_for_Multimodal_Sentiment_Analysis.pdf]
Preview
Text
MAG___An_EXTENDED_Multimodal_Adaptation_Gate_for_Multimodal_Sentiment_Analysis.pdf - Accepted Version

Download (1MB) | Preview

Abstract

Human multimodal sentiment analysis is a challenging task that devotes to extract and integrate information from multiple resources, such as language, acoustic and visual information. Recently, multimodal adaptation gate (MAG), an attachment to transformer-based pre-trained language representation models, such as BERT and XLNet, has shown state-of-the-art performance on multimodal sentiment analysis. MAG only uses a 1-layer network to fuse multimodal information directly, and does not pay attention to relationships among different modalities. In this paper, we propose an extended MAG, called MAG+, to reinforce multimodal fusion. MAG+ contains two modules: multi-layer MAGs with modality reinforcement (M3R) and Adaptive Layer Aggregation (ALA). In the MAG with modality reinforcement of M3R, each modality is reinforced by all other modalities via crossmodal attention at first, and then all modalities are fused via MAG. The ALA module leverages the multimodal representations at low and high levels as the final multimodal representation. Similar to MAG, MAG+ is also attached to BERT and XLNet. Experimental results on two widely used datasets demonstrate the efficacy of our proposed MAG+.

Type: Proceedings paper
Title: MAG+: An Extended Multimodal Adaptation Gate for Multimodal Sentiment Analysis
Event: ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
Location: Singapore, Singapore
Dates: 23rd-27th May 2022
ISBN-13: 978-1-6654-0540-9
Open access status: An open access version is available from UCL Discovery
DOI: 10.1109/ICASSP43922.2022.9746536
Publisher version: http://dx.doi.org/10.1109/icassp43922.2022.9746536
Language: English
Additional information: This version is the author accepted manuscript. For information on re-use, please refer to the publisher's terms and conditions.
UCL classification: UCL
UCL > Provost and Vice Provost Offices > UCL BEAMS
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science > Dept of Computer Science
URI: https://discovery.ucl.ac.uk/id/eprint/10188927
Downloads since deposit
19Downloads
Download activity - last month
Download activity - last 12 months
Downloads by country - last 12 months

Archive Staff Only

View Item View Item