LLM Memory & RAG 相关度: 9/10

AttentionRetriever: Attention Layers are Secretly Long Document Retrievers

David Jiahao Fu, Lam Thanh Do, Jiayu Li, Kevin Chen-Chuan Chang
arXiv: 2602.12278v1 发布: 2026-02-12 更新: 2026-02-12

AI 摘要

AttentionRetriever利用注意力机制和实体检索,构建上下文感知嵌入,提升长文档检索性能和效率。

主要贡献

  • 提出AttentionRetriever模型,提升长文档检索性能
  • 利用注意力机制构建上下文感知嵌入
  • 通过实体检索确定检索范围

方法论

提出AttentionRetriever,结合注意力机制和实体检索,为长文档构建上下文感知嵌入,优化检索范围。

原文摘要

Retrieval augmented generation (RAG) has been widely adopted to help Large Language Models (LLMs) to process tasks involving long documents. However, existing retrieval models are not designed for long document retrieval and fail to address several key challenges of long document retrieval, including context-awareness, causal dependence, and scope of retrieval. In this paper, we proposed AttentionRetriever, a novel long document retrieval model that leverages attention mechanism and entity-based retrieval to build context-aware embeddings for long document and determine the scope of retrieval. With extensive experiments, we found AttentionRetriever is able to outperform existing retrieval models on long document retrieval datasets by a large margin while remaining as efficient as dense retrieval models.

标签

RAG 长文档检索 注意力机制 实体检索

arXiv 分类

cs.IR cs.AI