Agent Tuning & Optimization 相关度: 9/10

GUI-Libra: Training Native GUI Agents to Reason and Act with Action-aware Supervision and Partially Verifiable RL

Rui Yang, Qianhui Wu, Zhaoyang Wang, Hanyang Chen, Ke Yang, Hao Cheng, Huaxiu Yao, Baoling Peng, Huan Zhang, Jianfeng Gao, Tong Zhang
arXiv: 2602.22190v1 发布: 2026-02-25 更新: 2026-02-25

AI 摘要

GUI-Libra提出了一种针对GUI智能体的训练方法,优化了数据、SFT和RL过程,显著提升了任务完成度。

主要贡献

  • 构建并发布了一个81K的GUI推理数据集,缓解了动作对齐推理数据稀缺的问题。
  • 提出了动作感知的SFT方法,平衡了推理和基础能力,提升了智能体的泛化性。
  • 针对GUI智能体的部分可验证性问题,改进了RL训练方法,增强了离线指标与在线性能的关联。

方法论

通过数据增强、动作感知SFT以及KL正则化的RLVR,优化GUI智能体的训练过程,提升其任务完成能力。

原文摘要

Open-source native GUI agents still lag behind closed-source systems on long-horizon navigation tasks. This gap stems from two limitations: a shortage of high-quality, action-aligned reasoning data, and the direct adoption of generic post-training pipelines that overlook the unique challenges of GUI agents. We identify two fundamental issues in these pipelines: (i) standard SFT with CoT reasoning often hurts grounding, and (ii) step-wise RLVR-tyle training faces partial verifiability, where multiple actions can be correct but only a single demonstrated action is used for verification. This makes offline step-wise metrics weak predictors of online task success. In this work, we present GUI-Libra, a tailored training recipe that addresses these challenges. First, to mitigate the scarcity of action-aligned reasoning data, we introduce a data construction and filtering pipeline and release a curated 81K GUI reasoning dataset. Second, to reconcile reasoning with grounding, we propose action-aware SFT that mixes reasoning-then-action and direct-action data and reweights tokens to emphasize action and grounding. Third, to stabilize RL under partial verifiability, we identify the overlooked importance of KL regularization in RLVR and show that a KL trust region is critical for improving offline-to-online predictability; we further introduce success-adaptive scaling to downweight unreliable negative gradients. Across diverse web and mobile benchmarks, GUI-Libra consistently improves both step-wise accuracy and end-to-end task completion. Our results suggest that carefully designed post-training and data curation can unlock significantly stronger task-solving capabilities without costly online data collection. We release our dataset, code, and models to facilitate further research on data-efficient post-training for reasoning-capable GUI agents.

标签

GUI Agent Reinforcement Learning Supervised Fine-tuning Reasoning

arXiv 分类

cs.LG cs.AI cs.CL