Multimodal Learning 相关度: 8/10

UniGRPO: Unified Policy Optimization for Reasoning-Driven Visual Generation

Jie Liu, Zilyu Ye, Linxiao Yuan, Shenhan Zhu, Yu Gao, Jie Wu, Kunchang Li, Xionghui Wang, Xiaonan Nie, Weilin Huang, Wanli Ouyang
arXiv: 2603.23500v1 发布: 2026-03-24 更新: 2026-03-24

AI 摘要

提出UniGRPO,用于联合优化推理和图像生成策略,提升图像生成质量,为多轮交互模型提供基线。

主要贡献

  • 提出UniGRPO框架,用于联合优化文本和图像生成策略。
  • 改进FlowGRPO,移除classifier-free guidance和替换KL惩罚。
  • 实验验证UniGRPO通过推理显著提升图像生成质量。

方法论

将多模态生成过程建模为MDP,使用GRPO优化文本策略,FlowGRPO优化图像策略,并进行改进。

原文摘要

Unified models capable of interleaved generation have emerged as a promising paradigm, with the community increasingly converging on autoregressive modeling for text and flow matching for image generation. To advance this direction, we propose a unified reinforcement learning framework tailored for interleaved generation. We validate our approach on its fundamental unit: a single round of reasoning-driven image generation, where the model first expands the user prompt through reasoning, followed by image synthesis. Formulating this multimodal generation process as a Markov Decision Process with sparse terminal rewards, we introduce UniGRPO to jointly optimize text and image generation policies using GRPO. Adopting a minimalist methodology to avoid over-design, we leverage established training recipes for both modalities by seamlessly integrating standard GRPO for reasoning and FlowGRPO for visual synthesis. To ensure scalability to multi-round interleaved generation, we introduce two critical modifications to the original FlowGRPO: (1) eliminating classifier-free guidance to maintain linear, unbranched rollouts, which is essential for scaling to complex scenarios involving multi-turn interactions and multi-condition generation (e.g., editing); and (2) replacing the standard latent KL penalty with an MSE penalty directly on the velocity fields, providing a more robust and direct regularization signal to mitigate reward hacking effectively. Our experiments demonstrate that this unified training recipe significantly enhances image generation quality through reasoning, providing a robust and scalable baseline for the future post-training of fully interleaved models.

标签

多模态学习 强化学习 图像生成 推理

arXiv 分类

cs.CV