Diagnosing Knowledge Conflict in Multimodal Long-Chain Reasoning
AI 摘要
该论文研究了多模态LLM在长链推理中因知识冲突导致的失败问题,并提出了诊断和控制方法。
主要贡献
- 形式化了知识冲突的概念,区分了输入层和过程层的冲突
- 通过探针实验揭示了冲突信号的线性可分性、深度定位、层次一致性和方向不对称性
- 提出了基于机制的知识冲突视角,用于诊断和控制长链推理失败
方法论
通过探针内部表征,分析不同类型的知识冲突在模型内部的表达和处理方式。
原文摘要
Multimodal large language models (MLLMs) in long chain-of-thought reasoning often fail when different knowledge sources provide conflicting signals. We formalize these failures under a unified notion of knowledge conflict, distinguishing input-level objective conflict from process-level effective conflict. Through probing internal representations, we reveal that: (I) Linear Separability: different conflict types are explicitly encoded as linearly separable features rather than entangled; (II) Depth Localization: conflict signals concentrate in mid-to-late layers, indicating a distinct processing stage for conflict encoding; (III) Hierarchical Consistency: aggregating noisy token-level signals along trajectories robustly recovers input-level conflict types; and (IV) Directional Asymmetry: reinforcing the model's implicit source preference under conflict is far easier than enforcing the opposite source. Our findings provide a mechanism-level view of multimodal reasoning under knowledge conflict and enable principled diagnosis and control of long-CoT failures.