Multimodal Learning 相关度: 9/10

What, Whether and How? Unveiling Process Reward Models for Thinking with Images Reasoning

Yujin Zhou, Pengcheng Wen, Jiale Chen, Boqin Yin, Han Zhu, Jiaming Ji, Juntao Dai, Chi-Min Chan, Sirui Han
arXiv: 2602.08346v1 发布: 2026-02-09 更新: 2026-02-09

AI 摘要

论文提出了一个用于评估LVLMs视觉推理过程奖励模型的综合基准测试。

主要贡献

  • 定义了7种细粒度的错误类型,揭示了专用PRM的必要性。
  • 构建了一个包含1206条人工标注推理轨迹的综合基准。
  • 分析表明现有LVLMs作为有效PRM存在不足。

方法论

通过分析推理轨迹和PRMs引导的搜索实验,构建基准测试并评估现有LVLMs的性能。

原文摘要

The rapid advancement of Large Vision Language Models (LVLMs) has demonstrated excellent abilities in various visual tasks. Building upon these developments, the thinking with images paradigm has emerged, enabling models to dynamically edit and re-encode visual information at each reasoning step, mirroring human visual processing. However, this paradigm introduces significant challenges as diverse errors may occur during reasoning processes. This necessitates Process Reward Models (PRMs) for distinguishing positive and negative reasoning steps, yet existing benchmarks for PRMs are predominantly text-centric and lack comprehensive assessment under this paradigm. To address these gaps, this work introduces the first comprehensive benchmark specifically designed for evaluating PRMs under the thinking with images paradigm. Our main contributions are: (1) Through extensive analysis of reasoning trajectories and guided search experiments with PRMs, we define 7 fine-grained error types and demonstrate both the necessity for specialized PRMs and the potential for improvement. (2) We construct a comprehensive benchmark comprising 1,206 manually annotated thinking with images reasoning trajectories spanning 4 categories and 16 subcategories for fine-grained evaluation of PRMs. (3) Our experimental analysis reveals that current LVLMs fall short as effective PRMs, exhibiting limited capabilities in visual reasoning process evaluation with significant performance disparities across error types, positive evaluation bias, and sensitivity to reasoning step positions. These findings demonstrate the effectiveness of our benchmark and establish crucial foundations for advancing PRMs in LVLMs.

标签

视觉推理 奖励模型 基准测试 LVLMs 图像理解

arXiv 分类

cs.CV