LLM Memory & RAG 相关度: 9/10

SmartSearch: How Ranking Beats Structure for Conversational Memory Retrieval

Jesper Derehag, Carlos Calva, Timmy Ghiurau
arXiv: 2603.15599v1 发布: 2026-03-16 更新: 2026-03-16

AI 摘要

SmartSearch通过简单的排序方法,在对话记忆检索任务上超越了复杂的结构化方法。

主要贡献

  • 提出了一种基于排序的对话记忆检索方法SmartSearch
  • 证明了在对话记忆检索中排序比结构更重要
  • 在两个基准测试中,SmartSearch优于现有系统,且计算成本更低

方法论

使用NER加权子字符串匹配召回,规则实体发现进行多跳扩展,以及CrossEncoder+ColBERT排序融合。

原文摘要

Recent conversational memory systems invest heavily in LLM-based structuring at ingestion time and learned retrieval policies at query time. We show that neither is necessary. SmartSearch retrieves from raw, unstructured conversation history using a fully deterministic pipeline: NER-weighted substring matching for recall, rule-based entity discovery for multi-hop expansion, and a CrossEncoder+ColBERT rank fusion stage -- the only learned component -- running on CPU in ~650ms. Oracle analysis on two benchmarks identifies a compilation bottleneck: retrieval recall reaches 98.6%, but without intelligent ranking only 22.5% of gold evidence survives truncation to the token budget. With score-adaptive truncation and no per-dataset tuning, SmartSearch achieves 93.5% on LoCoMo and 88.4% on LongMemEval-S, exceeding all known memory systems under the same evaluation protocol on both benchmarks while using 8.5x fewer tokens than full-context baselines.

标签

对话记忆 信息检索 排序 检索增强生成 RAG

arXiv 分类

cs.LG