LLM Memory & RAG 相关度: 9/10

Supercharging Federated Intelligence Retrieval

Dimitris Stripelis, Patrick Foley, Mohammad Naseri, William Lindskog-Münzing, Chong Shen Ng, Daniel Janes Beutel, Nicholas D. Lane
arXiv: 2603.25374v1 发布: 2026-03-26 更新: 2026-03-26

AI 摘要

提出一种安全的联邦RAG系统,在保护隐私的同时实现分布式知识检索和远程LLM推理。

主要贡献

  • 提出安全联邦RAG系统
  • 使用Flower进行联邦学习
  • 引入可信执行环境
  • 提出级联推理方法

方法论

利用Flower进行本地检索聚合,服务端在可信环境中进行文本生成,并结合非机密模型进行辅助上下文推理。

原文摘要

RAG typically assumes centralized access to documents, which breaks down when knowledge is distributed across private data silos. We propose a secure Federated RAG system built using Flower that performs local silo retrieval, while server-side aggregation and text generation run inside an attested, confidential compute environment, enabling confidential remote LLM inference even in the presence of honest-but-curious or compromised servers. We also propose a cascading inference approach that incorporates a non-confidential third-party model (e.g., Amazon Nova) as auxiliary context without weakening confidentiality.

标签

联邦学习 RAG 隐私保护 LLM 可信计算

arXiv 分类

cs.IR cs.CL cs.CR cs.LG