Multimodal Learning 相关度: 9/10

MViR: Multi-View Visual-Semantic Representation for Fake News Detection

Haochen Liang, Xinqi Su, Jun Wang, Chaomeng Chen, Zitong Yu
arXiv: 2602.22944v1 发布: 2026-02-26 更新: 2026-02-26

AI 摘要

MViR通过多视角视觉语义表示提升假新闻检测性能,融合图像和文本信息。

主要贡献

  • 提出多视角视觉语义表示框架(MViR)
  • 使用金字塔空洞卷积捕获多视角视觉语义特征
  • 多视角特征融合与多视角语义线索提取

方法论

使用金字塔空洞卷积提取多视角视觉语义特征,然后融合文本信息,最后使用聚合器提取语义线索进行检测。

原文摘要

With the rise of online social networks, detecting fake news accurately is essential for a healthy online environment. While existing methods have advanced multimodal fake news detection, they often neglect the multi-view visual-semantic aspects of news, such as different text perspectives of the same image. To address this, we propose a Multi-View Visual-Semantic Representation (MViR) framework. Our approach includes a Multi-View Representation module using pyramid dilated convolution to capture multi-view visual-semantic features, a Multi-View Feature Fusion module to integrate these features with text, and multiple aggregators to extract multi-view semantic cues for detection. Experiments on benchmark datasets demonstrate the superiority of MViR. The source code of FedCoop is available at https://github.com/FlowerinZDF/FakeNews-MVIR.

标签

假新闻检测 多模态学习 视觉语义 多视角学习

arXiv 分类

cs.MM