SmartSearch: How Ranking Beats Structure for Conversational Memory Retrieval
AI 摘要
SmartSearch通过简单的排序方法,在对话记忆检索任务上超越了复杂的结构化方法。
主要贡献
- 提出了一种基于排序的对话记忆检索方法SmartSearch
- 证明了在对话记忆检索中排序比结构更重要
- 在两个基准测试中,SmartSearch优于现有系统,且计算成本更低
方法论
使用NER加权子字符串匹配召回,规则实体发现进行多跳扩展,以及CrossEncoder+ColBERT排序融合。
原文摘要
Recent conversational memory systems invest heavily in LLM-based structuring at ingestion time and learned retrieval policies at query time. We show that neither is necessary. SmartSearch retrieves from raw, unstructured conversation history using a fully deterministic pipeline: NER-weighted substring matching for recall, rule-based entity discovery for multi-hop expansion, and a CrossEncoder+ColBERT rank fusion stage -- the only learned component -- running on CPU in ~650ms. Oracle analysis on two benchmarks identifies a compilation bottleneck: retrieval recall reaches 98.6%, but without intelligent ranking only 22.5% of gold evidence survives truncation to the token budget. With score-adaptive truncation and no per-dataset tuning, SmartSearch achieves 93.5% on LoCoMo and 88.4% on LongMemEval-S, exceeding all known memory systems under the same evaluation protocol on both benchmarks while using 8.5x fewer tokens than full-context baselines.