LLM Memory & RAG 相关度: 7/10

Inference-time Unlearning Using Conformal Prediction

Somnath Basu Roy Chowdhury, Rahul Kidambi, Avinava Dubey, David Wang, Gokhan Mergen, Amr Ahmed, Aranyak Mehta
arXiv: 2602.03787v1 发布: 2026-02-03 更新: 2026-02-03

AI 摘要

提出一种基于Conformal Prediction的推理时免训练卸载框架,提升卸载性能并提供保证。

主要贡献

  • 提出了推理时卸载框架,无需模型参数更新
  • 利用Conformal Prediction减少计算开销
  • 提供分布无关的卸载保证

方法论

使用可验证器迭代优化生成响应,基于Conformal Prediction提供反馈,实现推理时卸载。

原文摘要

Machine unlearning is the process of efficiently removing specific information from a trained machine learning model without retraining from scratch. Existing unlearning methods, which often provide provable guarantees, typically involve retraining a subset of model parameters based on a forget set. While these approaches show promise in certain scenarios, their underlying assumptions are often challenged in real-world applications -- particularly when applied to generative models. Furthermore, updating parameters using these unlearning procedures often degrades the general-purpose capabilities the model acquired during pre-training. Motivated by these shortcomings, this paper considers the paradigm of inference time unlearning -- wherein, the generative model is equipped with an (approximately correct) verifier that judges whether the model's response satisfies appropriate unlearning guarantees. This paper introduces a framework that iteratively refines the quality of the generated responses using feedback from the verifier without updating the model parameters. The proposed framework leverages conformal prediction to reduce computational overhead and provide distribution-free unlearning guarantees. This paper's approach significantly outperforms existing state-of-the-art methods, reducing unlearning error by up to 93% across challenging unlearning benchmarks.

标签

Machine Unlearning Conformal Prediction Generative Models

arXiv 分类

cs.LG