LLM Memory & RAG 相关度: 8/10

Learn by Surprise, Commit by Proof

Kang-Sin Choi
arXiv: 2604.01951v1 发布: 2026-04-02 更新: 2026-04-02

AI 摘要

LSCP通过自验证方式,使模型学习新知识并修正已有知识,避免死记硬背。

主要贡献

  • 提出LSCP自门控后训练框架
  • 利用Q&A链迫使模型识别知识缺口
  • 通过β2门控防止邻近知识污染

方法论

LSCP利用模型自身的知识评估和Q&A过程,调整AdamW优化器的参数,实现自监督学习。

原文摘要

We propose LSCP, a self-gated post-training framework for autonomous knowledge acquisition: learning only what a model does not already know, verified against what it does know, at a strength proportional to conviction, with no external oracle. When a passage produces anomalously high per-token loss, LSCP flags it, generates a Q&A chain that forces the model to articulate its own knowledge and identify gaps, then adjusts AdamW's $β_2$ proportionally to conviction depth k (the number of self-verification steps the passage survives) via $β_2 = 0.999 \cdot r^k$. The entire learning intensity is governed by a single parameter $r$. Beyond new knowledge, this process sharpens weakly encoded existing knowledge, which is a primary source of hallucination. The framework is self-extinguishing: as the model learns, per-token loss on learned passages decreases toward the surprisal threshold and the system progressively converges to standard AdamW. This models biological memory consolidation: temporary information in the context window is selectively consolidated into parametric weights, the model's long-term memory. Experiments on the reference model (Qwen3-14B) and across six models (8B--32B, four families) show that standard fine-tuning produces rote memorization (perturbation gap (the ratio of paraphrase to original perplexity) of 11.6 +- 0.2 x baseline) while all LSCP conditions learn semantically (2.7--3.0x). The r=1.0 condition (identical optimizer, nearly identical data, only Q&A format differs) confirms that the training data format, not $β_2$ gating, is the primary mechanism preventing memorization; gating instead protects neighboring knowledge from contamination by corrupt content (93 +- 7% accuracy on adjacent questions at r=0.98 vs. 90% baseline).

标签

自监督学习 知识获取 模型修正 后训练

arXiv 分类

cs.LG