Team of Thoughts: Efficient Test-time Scaling of Agentic Systems through Orchestrated Tool Calling
AI 摘要
Team-of-Thoughts通过异构Agent协同,提升Agent系统在推理和代码生成任务上的性能。
主要贡献
- 提出Team-of-Thoughts架构,利用异构Agent互补能力
- 引入Orchestrator校准机制,选择最佳协调模型
- 设计自评估协议,使工具Agent评估自身领域专长
方法论
利用Orchestrator动态激活最合适的工具Agent,基于Agent的专长profile进行推理和代码生成。
原文摘要
Existing Multi-Agent Systems (MAS) typically rely on static, homogeneous model configurations, limiting their ability to exploit the distinct strengths of differently post-trained models. To address this, we introduce Team-of-Thoughts, a novel MAS architecture that leverages the complementary capabilities of heterogeneous agents via an orchestrator-tool paradigm. Our framework introduces two key mechanisms to optimize performance: (1) an orchestrator calibration scheme that identifies models with superior coordination capabilities, and (2) a self-assessment protocol where tool agents profile their own domain expertise to account for variations in post-training skills. During inference, the orchestrator dynamically activates the most suitable tool agents based on these proficiency profiles. Experiments on five reasoning and code generation benchmarks show that Team-of-Thoughts delivers consistently superior task performance. Notably, on AIME24 and LiveCodeBench, our approach achieves accuracies of 96.67% and 72.53%, respectively, substantially outperforming homogeneous role-play baselines, which score 80% and 65.93%.