FactGuard: Agentic Video Misinformation Detection via Reinforcement Learning
AI 摘要
FactGuard通过强化学习训练Agent进行视频虚假信息检测,提升了鲁棒性和泛化能力。
主要贡献
- 提出FactGuard Agent框架,迭代推理进行视频虚假信息检测
- 引入两阶段训练策略,优化工具使用和风险决策
- 实验证明FactGuard在多个数据集上的优越性能
方法论
构建基于MLLM的Agent框架,结合领域知识的监督微调和决策感知的强化学习,优化工具使用。
原文摘要
Multimodal large language models (MLLMs) have substantially advanced video misinformation detection through unified multimodal reasoning, but they often rely on fixed-depth inference and place excessive trust in internally generated assumptions, particularly in scenarios where critical evidence is sparse, fragmented, or requires external verification. To address these limitations, we propose FactGuard, an agentic framework for video misinformation detection that formulates verification as an iterative reasoning process built upon MLLMs. FactGuard explicitly assesses task ambiguity and selectively invokes external tools to acquire critical evidence, enabling progressive refinement of reasoning trajectories. To further strengthen this capability, we introduce a two-stage training strategy that combines domain-specific agentic supervised fine-tuning with decision-aware reinforcement learning to optimize tool usage and calibrate risk-sensitive decision making. Extensive experiments on FakeSV, FakeTT, and FakeVV demonstrate FactGuard's state-of-the-art performance and validate its excellent robustness and generalization capacity.