Revealing Behavioral Plasticity in Large Language Models: A Token-Conditional Perspective
AI 摘要
LLM具有行为可塑性,可通过token条件生成和强化学习进行调控,实现行为模式切换。
主要贡献
- 揭示LLM内在的行为可塑性
- 提出Token-Conditioned Reinforcement Learning (ToCoRL)框架
- ToCoRL能有效控制LLM行为,且不降低能力
方法论
通过Token条件生成引导探索,利用强化学习增强利用,学习并稳定LLM的token-conditional行为模式。
原文摘要
In this work, we reveal that Large Language Models (LLMs) possess intrinsic behavioral plasticity-akin to chameleons adapting their coloration to environmental cues-that can be exposed through token-conditional generation and stabilized via reinforcement learning. Specifically, by conditioning generation on carefully selected token prefixes sampled from responses exhibiting desired behaviors, LLMs seamlessly adapt their behavioral modes at inference time (e.g., switching from step-by-step reasoning to direct answering) without retraining. Based on this insight, we propose Token-Conditioned Reinforcement Learning (ToCoRL), a principled framework that leverages RL to internalize this chameleon-like plasticity, transforming transient inference-time adaptations into stable and learnable behavioral patterns. ToCoRL guides exploration with token-conditional generation and keep enhancing exploitation, enabling emergence of appropriate behaviors. Extensive experiments show that ToCoRL enables precise behavioral control without capability degradation. Notably, we show that large reasoning models, while performing strongly on complex mathematics, can be effectively adapted to excel at factual question answering, which was a capability previously hindered by their step-by-step reasoning patterns.