Adaptive Modality Distillation for Separable Multimodal Sentiment Analysis

Multimodal sentiment analysis has increasingly attracted attention since with the arrival of complementary data streams, it has great potential to improve and go beyond unimodal sentiment analysis. In this paper, we present an efficient separable multimodal learning method to deal with the tasks with modality missing issue. In this method, the multimodal tensor is utilized to guide the evolution of each separated modality representation. To save the computational expense, Tucker decomposition is introduced, which leads to a general extension of the low-rank tensor fusion method with more modality interactions. The method, in turn, enhances our modality distillation processing. Comprehensive experiments on three popular multimodal sentiment analysis datasets, CMU-MOSI, POM, and IEMOCAP, show a superior performance especially when only partial modalities are available.

Peng Wei, Hong Xiaopeng, Zhao Guoying

A1 Journal article – refereed

W. Peng, X. Hong and G. Zhao, "Adaptive Modality Distillation for Separable Multimodal Sentiment Analysis," in IEEE Intelligent Systems, vol. 36, no. 3, pp. 82-89, 1 May-June 2021, doi: 10.1109/MIS.2021.3057757

https://doi.org/10.1109/MIS.2021.3057757 http://urn.fi/urn:nbn:fi-fe202104099805