4.7 Article

M2Lens: Visualizing and Explaining Multimodal Models for Sentiment Analysis

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TVCG.2021.3114794

关键词

Analytical models; Sentiment analysis; Computational modeling; Predictive models; Data models; Lenses; Communication channels; Multimodal models; sentiment analysis; explainable machine learning

资金

  1. Foshan-HKUST Projects [FSNH20EGO1]

向作者/读者索取更多资源

Multimodal sentiment analysis is a vital research area for recognizing people's attitudes from multiple communication channels, but current models often lack explainability regarding how they utilize multimodal information. While recent advancements have been made in enhancing machine learning model explainability, there is limited research on explaining multimodal models.
Multimodal sentiment analysis aims to recognize people's attitudes from multiple communication channels such as verbal content (i.e., text), voice, and facial expressions. It has become a vibrant and important research topic in natural language processing. Much research focuses on modeling the complex intra- and inter-modal interactions between different communication channels. However, current multimodal models with strong performance are often deep-learning-based techniques and work like black boxes. It is not clear how models utilize multimodal information for sentiment predictions. Despite recent advances in techniques for enhancing the explainability of machine learning models, they often target unimodal scenarios (e.g., images, sentences), and little research has been done on explaining multimodal models. In this paper, we present an interactive visual analytics system, M2 Lens, to visualize and explain multimodal models for sentiment analysis. M2 Lens provides explanations on intra- and inter-modal interactions at the global, subset, and local levels. Specifically, it summarizes the influence of three typical interaction types (i.e., dominance, complement, and conflict) on the model predictions. Moreover, M2 Lens identifies frequent and influential multimodal features and supports the multi-faceted exploration of model behaviors from language, acoustic, and visual modalities. Through two case studies and expert interviews, we demonstrate our system can help users gain deep insights into the multimodal models for sentiment analysis.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据