菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-10
📄 Abstract - Progressive Representation Learning for Multimodal Sentiment Analysis with Incomplete Modalities

Multimodal Sentiment Analysis (MSA) seeks to infer human emotions by integrating textual, acoustic, and visual cues. However, existing approaches often rely on all modalities are completeness, whereas real-world applications frequently encounter noise, hardware failures, or privacy restrictions that result in missing modalities. There exists a significant feature misalignment between incomplete and complete modalities, and directly fusing them may even distort the well-learned representations of the intact modalities. To this end, we propose PRLF, a Progressive Representation Learning Framework designed for MSA under uncertain missing-modality conditions. PRLF introduces an Adaptive Modality Reliability Estimator (AMRE), which dynamically quantifies the reliability of each modality using recognition confidence and Fisher information to determine the dominant modality. In addition, the Progressive Interaction (ProgInteract) module iteratively aligns the other modalities with the dominant one, thereby enhancing cross-modal consistency while suppressing noise. Extensive experiments on CMU-MOSI, CMU-MOSEI, and SIMS verify that PRLF outperforms state-of-the-art methods across both inter- and intra-modality missing scenarios, demonstrating its robustness and generalization capability.

顶级标签: multi-modal natural language processing model training
详细标签: multimodal sentiment analysis missing modalities representation learning robustness cross-modal alignment 或 搜索:

面向模态缺失的多模态情感分析的渐进式表征学习 / Progressive Representation Learning for Multimodal Sentiment Analysis with Incomplete Modalities


1️⃣ 一句话总结

这篇论文提出了一种名为PRLF的新方法,它能在文本、声音或图像等部分信息缺失的情况下,通过动态评估各信息的可靠度并逐步对齐不同信息,来更准确地进行情感分析,并在多个数据集上验证了其优越性。

源自 arXiv: 2603.09111