Seeing to Ground: Visual Attention for Hallucination-Resilient MDLLMs
AI 摘要
论文提出VISAGE框架,通过校准目标函数,减少多模态大语言模型中的幻觉问题。
主要贡献
- 提出VISAGE框架,用于减少多模态幻觉
- 分析了多模态幻觉的根本原因:目标不匹配
- 提供了VISAGE的理论稳定性保证
方法论
通过量化交叉注意力分布的空间熵来估计代理差异,并重新排序token承诺以支持视觉上可靠的结果。
原文摘要
Multimodal Diffusion Large Language Models (MDLLMs) achieve high-concurrency generation through parallel masked decoding, yet the architectures remain prone to multimodal hallucinations. This structural vulnerability stems from an algorithmic flaw: the decoder ranks candidate tokens based on textual likelihood without verifying localized visual support. We establish that this language-only ranking induces an objective mismatch, where language probability mass acts as a misspecified proxy for the intended multimodal task. Consequently, we reinterpret hallucination as a localized optimization error, a phenomenon where the decoder exploits language shortcuts to maximize a proxy score at the expense of visual grounding. To address this objective mismatch, we introduce VISAGE, a training-free decoding framework that calibrates the objective at inference time. VISAGE estimates the proxy discrepancy by quantifying the spatial entropy of cross-attention distributions. By enforcing a localization consensus across attention heads, the method penalizes spatially uniform distributions and re-ranks token commitments to favor visually grounded outcomes. We provide an analytical stability guarantee establishing that VISAGE maintains a bounded objective loss under estimation error. Evaluations across hallucination-sensitive and general-purpose benchmarks demonstrate the robustness of the framework, yielding relative gains of 8.59% on MMMU-val and 7.75% on HallusionBench.