LLM Memory & RAG 相关度: 9/10

Facts as First Class Objects: Knowledge Objects for Persistent LLM Memory

Oliver Zahn, Simran Chana
arXiv: 2603.17781v1 发布: 2026-03-18 更新: 2026-03-18

AI 摘要

论文提出Knowledge Objects (KOs)作为LLM持久记忆方案,解决传统in-context learning的不足。

主要贡献

  • 提出Knowledge Objects (KOs)的架构
  • 揭示in-context memory的三种失败模式:容量限制,压缩损失和目标漂移
  • 展示KOs在多跳推理和成本效益方面的优势

方法论

论文通过benchmark比较in-context memory和KOs的性能,并在多个前沿模型上验证结果。

原文摘要

Large language models increasingly serve as persistent knowledge workers, with in-context memory - facts stored in the prompt - as the default strategy. We benchmark in-context memory against Knowledge Objects (KOs), discrete hash-addressed tuples with O(1) retrieval. Within the context window, Claude Sonnet 4.5 achieves 100% exact-match accuracy from 10 to 7,000 facts (97.5% of its 200K window). However, production deployment reveals three failure modes: capacity limits (prompts overflow at 8,000 facts), compaction loss (summarization destroys 60% of facts), and goal drift (cascading compaction erodes 54% of project constraints while the model continues with full confidence). KOs achieve 100% accuracy across all conditions at 252x lower cost. On multi-hop reasoning, KOs reach 78.9% versus 31.6% for in-context. Cross-model replication across four frontier models confirms compaction loss is architectural, not model-specific. We additionally show that embedding retrieval fails on adversarial facts (20% precision at 1) and that neural memory (Titans) stores facts but fails to retrieve them on demand. We introduce density-adaptive retrieval as a switching mechanism and release the benchmark suite.

标签

LLM Memory Knowledge Objects Retrieval

arXiv 分类

cs.AI