AI Agents 相关度: 8/10

Multi-agent imitation learning with function approximation: Linear Markov games and beyond

Luca Viano, Till Freihaut, Emanuele Nevali, Volkan Cevher, Matthieu Geist, Giorgia Ramponi
arXiv: 2602.22810v1 发布: 2026-02-26 更新: 2026-02-26

AI 摘要

研究线性马尔可夫博弈中的多智能体模仿学习,提出理论分析和高效算法。

主要贡献

  • 提出线性马尔可夫博弈中多智能体模仿学习的理论分析
  • 提出特征层面的集中性系数,替代状态-动作层面的系数
  • 提出交互式多智能体模仿学习算法,样本复杂度只依赖于特征维度

方法论

理论分析与算法设计,利用线性马尔可夫博弈的结构,并构建深度模仿学习算法。

原文摘要

In this work, we present the first theoretical analysis of multi-agent imitation learning (MAIL) in linear Markov games where both the transition dynamics and each agent's reward function are linear in some given features. We demonstrate that by leveraging this structure, it is possible to replace the state-action level "all policy deviation concentrability coefficient" (Freihaut et al., arXiv:2510.09325) with a concentrability coefficient defined at the feature level which can be much smaller than the state-action analog when the features are informative about states' similarity. Furthermore, to circumvent the need for any concentrability coefficient, we turn to the interactive setting. We provide the first, computationally efficient, interactive MAIL algorithm for linear Markov games and show that its sample complexity depends only on the dimension of the feature map $d$. Building on these theoretical findings, we propose a deep MAIL interactive algorithm which clearly outperforms BC on games such as Tic-Tac-Toe and Connect4.

标签

多智能体模仿学习 线性马尔可夫博弈 强化学习 函数逼近

arXiv 分类

cs.LG