LLM Reasoning 相关度: 7/10

Small LLMs for Medical NLP: a Systematic Analysis of Few-Shot, Constraint Decoding, Fine-Tuning and Continual Pre-Training in Italian

Pietro Ferrazzi, Mattia Franzin, Alberto Lavelli, Bernardo Magnini
arXiv: 2602.17475v1 发布: 2026-02-19 更新: 2026-02-19

AI 摘要

该论文研究了小型LLM在意大利语医疗NLP任务上的表现,并比较了多种优化策略。

主要贡献

  • 评估小型LLM在医疗NLP任务上的性能
  • 比较了不同的适应策略,包括微调和约束解码
  • 发布了意大利语医疗NLP数据集和预训练数据

方法论

在20个临床NLP任务上,比较了Llama-3、Gemma-3和Qwen3等小型LLM的微调、少样本学习和约束解码等方法。

原文摘要

Large Language Models (LLMs) consistently excel in diverse medical Natural Language Processing (NLP) tasks, yet their substantial computational requirements often limit deployment in real-world healthcare settings. In this work, we investigate whether "small" LLMs (around one billion parameters) can effectively perform medical tasks while maintaining competitive accuracy. We evaluate models from three major families-Llama-3, Gemma-3, and Qwen3-across 20 clinical NLP tasks among Named Entity Recognition, Relation Extraction, Case Report Form Filling, Question Answering, and Argument Mining. We systematically compare a range of adaptation strategies, both at inference time (few-shot prompting, constraint decoding) and at training time (supervised fine-tuning, continual pretraining). Fine-tuning emerges as the most effective approach, while the combination of few-shot prompting and constraint decoding offers strong lower-resource alternatives. Our results show that small LLMs can match or even surpass larger baselines, with our best configuration based on Qwen3-1.7B achieving an average score +9.2 points higher than Qwen3-32B. We release a comprehensive collection of all the publicly available Italian medical datasets for NLP tasks, together with our top-performing models. Furthermore, we release an Italian dataset of 126M words from the Emergency Department of an Italian Hospital, and 175M words from various sources that we used for continual pre-training.

标签

小型LLM 医疗NLP 意大利语 微调 约束解码

arXiv 分类

cs.CL