University of Leicester
Browse

Video based Cross-modal Auxiliary Network for Multimodal Sentiment Analysis

Download (1.48 MB)
journal contribution
posted on 2022-08-09, 10:27 authored by R Chen, W Zhou, Y Li, Huiyu Zhou

Multimodal sentiment analysis has a wide range of applications due to its information complementarity in multimodal interactions. Previous works focus more on investigating efficient joint representations, but they rarely consider the insufficient unimodal features extraction and data redundancy of multimodal fusion. In this paper, a Video-based Cross-modal Auxiliary Network (VCAN) is proposed, which is comprised of an audio features map module and a cross-modal selection module. The first module is designed to substantially increase feature diversity in audio feature extraction, aiming to improve classification accuracy by providing more comprehensive acoustic representations. To empower the model to handle redundant visual features, the second module is addressed to efficiently filter the redundant visual frames during integrating audiovisual data. Moreover, a classifier group consisting of several image classification networks is introduced to predict sentiment polarities and emotion categories. Extensive experimental results on RAVDESS, CMU-MOSI, and CMU-MOSEI benchmarks indicate that VCAN is significantly superior to the state-of-the-art methods for improving the classification accuracy of multimodal sentiment analysis.

Funding

National Natural Science Foundation of China (No. 61877065)

History

Author affiliation

School of Computing and Mathematical Sciences, University of Leicester

Version

  • AM (Accepted Manuscript)

Published in

IEEE Transactions on Circuits and Systems for Video Technology

Volume

32

Issue

12

Pagination

8703 - 8716

Publisher

Institute of Electrical and Electronics Engineers

issn

1051-8215

Acceptance date

2022-07-27

Copyright date

2022

Available date

2022-08-09

Language

en

Usage metrics

    University of Leicester Publications

    Categories

    No categories selected

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC