Rethinking Weight Tying: Pseudo-Inverse Tying for Stable LM Training and Updates
AI 摘要
提出伪逆权重绑定(PIT),通过共享的潜在token记忆同步embedding和unembedding,提升训练稳定性和语义一致性。
主要贡献
- 提出Pseudo-Inverse Tying (PIT)权重绑定方法
- 设计正交共享记忆和可学习的对称正定变换
- 提高训练稳定性和层间语义一致性
方法论
通过维持正交共享记忆和引入可学习的变换,实现embedding和unembedding的同步,避免显式的伪逆计算。
原文摘要
Weight tying is widely used in compact language models to reduce parameters by sharing the token table between the input embedding and the output projection. However, weight sharing does not guarantee a stable token interface: during training, the correspondence between encoding tokens into hidden states and decoding hidden states into logits can drift, worsening optimization sensitivity and making post-training interventions such as editing, patching, and lightweight adaptation less predictable. We propose Pseudo-Inverse Tying (PIT), which synchronizes embedding and unembedding as coupled projections of a shared latent token memory, guaranteeing a pseudo-inverse-consistent interface throughout training. PIT maintains an orthonormal shared memory, obtained by thin polar decomposition for teacher initialization or random orthonormal initialization from scratch, and introduces a fully learned symmetric positive definite hidden-space transform parameterized via a Cholesky factor. The output head applies this transform to hidden states before the vocabulary projection, while the embedding applies the inverse transform to token vectors using stable triangular solves, avoiding explicit pseudo-inverse recomputation and any vocabulary-sized auxiliary parameters. We evaluate PIT on on-device models spanning 256M-1.3B parameters across pretraining and adaptation, and consistently observe improved training stability, stronger layerwise semantic consistency, and substantially reduced side effects.