Explainable Token-level Noise Filtering for LLM Fine-tuning Datasets
AI 摘要
论文提出XTF框架,通过解释性的token级噪声过滤提升LLM微调性能。
主要贡献
- 提出XTF框架,分解token贡献为可解释的属性
- 利用token级噪声过滤改进LLM微调
- 实验证明XTF在多个任务上显著提升性能
方法论
XTF将token贡献分解为推理重要性、知识新颖性和任务相关性,并据此掩蔽噪声token的梯度以优化模型。
原文摘要
Large Language Models (LLMs) have seen remarkable advancements, achieving state-of-the-art results in diverse applications. Fine-tuning, an important step for adapting LLMs to specific downstream tasks, typically involves further training on corresponding datasets. However, a fundamental discrepancy exists between current fine-tuning datasets and the token-level optimization mechanism of LLMs: most datasets are designed at the sentence-level, which introduces token-level noise, causing negative influence to final performance. In this paper, we propose XTF, an explainable token-level noise filtering framework. XTF decomposes the complex and subtle contributions of token-level data to the fine-tuning process into three distinct and explicit attributes (reasoning importance, knowledge novelty, and task relevance), which can be assessed using scoring methods, and then masks the gradients of selected noisy tokens accordingly to optimize the performance of fine-tuned LLMs. We conduct extensive experiments on three representative downstream tasks (math, code and medicine) across 7 mainstream LLMs. The results demonstrate that XTF can significantly improve downstream performance by up to 13.7% compared to regular fine-tuning. Our work highlights the importance of token-level dataset optimization, and demonstrates the potential of strategies based on attribute decomposition for explaining complex training mechanisms.