Facts as First Class Objects: Knowledge Objects for Persistent LLM Memory
AI 摘要
论文提出Knowledge Objects (KOs)作为LLM持久记忆方案,解决传统in-context learning的不足。
主要贡献
- 提出Knowledge Objects (KOs)的架构
- 揭示in-context memory的三种失败模式:容量限制,压缩损失和目标漂移
- 展示KOs在多跳推理和成本效益方面的优势
方法论
论文通过benchmark比较in-context memory和KOs的性能,并在多个前沿模型上验证结果。
原文摘要
Large language models increasingly serve as persistent knowledge workers, with in-context memory - facts stored in the prompt - as the default strategy. We benchmark in-context memory against Knowledge Objects (KOs), discrete hash-addressed tuples with O(1) retrieval. Within the context window, Claude Sonnet 4.5 achieves 100% exact-match accuracy from 10 to 7,000 facts (97.5% of its 200K window). However, production deployment reveals three failure modes: capacity limits (prompts overflow at 8,000 facts), compaction loss (summarization destroys 60% of facts), and goal drift (cascading compaction erodes 54% of project constraints while the model continues with full confidence). KOs achieve 100% accuracy across all conditions at 252x lower cost. On multi-hop reasoning, KOs reach 78.9% versus 31.6% for in-context. Cross-model replication across four frontier models confirms compaction loss is architectural, not model-specific. We additionally show that embedding retrieval fails on adversarial facts (20% precision at 1) and that neural memory (Titans) stores facts but fails to retrieve them on demand. We introduce density-adaptive retrieval as a switching mechanism and release the benchmark suite.