MViR: Multi-View Visual-Semantic Representation for Fake News Detection
AI 摘要
MViR通过多视角视觉语义表示提升假新闻检测性能,融合图像和文本信息。
主要贡献
- 提出多视角视觉语义表示框架(MViR)
- 使用金字塔空洞卷积捕获多视角视觉语义特征
- 多视角特征融合与多视角语义线索提取
方法论
使用金字塔空洞卷积提取多视角视觉语义特征,然后融合文本信息,最后使用聚合器提取语义线索进行检测。
原文摘要
With the rise of online social networks, detecting fake news accurately is essential for a healthy online environment. While existing methods have advanced multimodal fake news detection, they often neglect the multi-view visual-semantic aspects of news, such as different text perspectives of the same image. To address this, we propose a Multi-View Visual-Semantic Representation (MViR) framework. Our approach includes a Multi-View Representation module using pyramid dilated convolution to capture multi-view visual-semantic features, a Multi-View Feature Fusion module to integrate these features with text, and multiple aggregators to extract multi-view semantic cues for detection. Experiments on benchmark datasets demonstrate the superiority of MViR. The source code of FedCoop is available at https://github.com/FlowerinZDF/FakeNews-MVIR.