Arbor: A Framework for Reliable Navigation of Critical Conversation Flows
AI 摘要
Arbor框架通过分解决策树导航任务,显著提升了LLM在复杂对话流程中的可靠性和效率。
主要贡献
- 提出Arbor框架,将决策树导航分解为节点级任务。
- 使用DAG进行流程编排,动态检索边缘信息,降低单次推理成本。
- 实验表明,Arbor提高了准确率,降低了延迟和成本。
方法论
将决策树转化为边缘列表,利用DAG进行流程编排,并通过专门的LLM调用评估有效转换,再进行响应生成。
原文摘要
Large language models struggle to maintain strict adherence to structured workflows in high-stakes domains such as healthcare triage. Monolithic approaches that encode entire decision structures within a single prompt are prone to instruction-following degradation as prompt length increases, including lost-in-the-middle effects and context window overflow. To address this gap, we present Arbor, a framework that decomposes decision tree navigation into specialized, node-level tasks. Decision trees are standardized into an edge-list representation and stored for dynamic retrieval. At runtime, a directed acyclic graph (DAG)-based orchestration mechanism iteratively retrieves only the outgoing edges of the current node, evaluates valid transitions via a dedicated LLM call, and delegates response generation to a separate inference step. The framework is agnostic to the underlying decision logic and model provider. Evaluated against single-prompt baselines across 10 foundation models using annotated turns from real clinical triage conversations. Arbor improves mean turn accuracy by 29.4 percentage points, reduces per-turn latency by 57.1%, and achieves an average 14.4x reduction in per-turn cost. These results indicate that architectural decomposition reduces dependence on intrinsic model capability, enabling smaller models to match or exceed larger models operating under single-prompt baselines.