AI Agents 相关度: 9/10

SWE-CI: Evaluating Agent Capabilities in Maintaining Codebases via Continuous Integration

Jialong Chen, Xander Xu, Hu Wei, Chuan Chen, Bing Zhao
arXiv: 2603.03823v1 发布: 2026-03-04 更新: 2026-03-04

AI 摘要

提出了SWE-CI基准,评估LLM Agent在持续集成环境中维护代码库的能力。

主要贡献

  • 提出了SWE-CI基准
  • 关注代码长期可维护性而非短期功能正确性
  • 提供真实代码库演化历史的评估任务

方法论

构建包含100个任务的基准,每个任务对应真实代码库的长期演化历史,要求Agent通过多轮迭代解决。

原文摘要

Large language model (LLM)-powered agents have demonstrated strong capabilities in automating software engineering tasks such as static bug fixing, as evidenced by benchmarks like SWE-bench. However, in the real world, the development of mature software is typically predicated on complex requirement changes and long-term feature iterations -- a process that static, one-shot repair paradigms fail to capture. To bridge this gap, we propose \textbf{SWE-CI}, the first repository-level benchmark built upon the Continuous Integration loop, aiming to shift the evaluation paradigm for code generation from static, short-term \textit{functional correctness} toward dynamic, long-term \textit{maintainability}. The benchmark comprises 100 tasks, each corresponding on average to an evolution history spanning 233 days and 71 consecutive commits in a real-world code repository. SWE-CI requires agents to systematically resolve these tasks through dozens of rounds of analysis and coding iterations. SWE-CI provides valuable insights into how well agents can sustain code quality throughout long-term evolution.

标签

代码生成 LLM Agent 持续集成 软件维护 基准测试

arXiv 分类

cs.SE cs.AI cs.CL