LLM Memory & RAG 相关度: 9/10

Cost-Efficient RAG for Entity Matching with LLMs: A Blocking-based Exploration

Chuangtao Ma, Zeyu Zhang, Arijit Khan, Sebastian Schelter, Paul Groth
arXiv: 2602.05708v1 发布: 2026-02-05 更新: 2026-02-05

AI 摘要

提出了基于分块的低成本RAG架构CE-RAG4EM,用于提升实体匹配效率。

主要贡献

  • 提出了一种基于分块的成本效益型RAG架构CE-RAG4EM
  • 提出了一个统一的实体匹配RAG系统分析与评估框架
  • 分析了性能和开销之间的权衡,为设计高效RAG系统提供指导

方法论

采用基于分块的批量检索和生成方法,减少计算开销,并对检索粒度进行优化。

原文摘要

Retrieval-augmented generation (RAG) enhances LLM reasoning in knowledge-intensive tasks, but existing RAG pipelines incur substantial retrieval and generation overhead when applied to large-scale entity matching. To address this limitation, we introduce CE-RAG4EM, a cost-efficient RAG architecture that reduces computation through blocking-based batch retrieval and generation. We also present a unified framework for analyzing and evaluating RAG systems for entity matching, focusing on blocking-aware optimizations and retrieval granularity. Extensive experiments suggest that CE-RAG4EM can achieve comparable or improved matching quality while substantially reducing end-to-end runtime relative to strong baselines. Our analysis further reveals that key configuration parameters introduce an inherent trade-off between performance and overhead, offering practical guidance for designing efficient and scalable RAG systems for entity matching and data integration.

标签

RAG Entity Matching Blocking Cost-Efficient LLM

arXiv 分类

cs.DB cs.CL