LLM Memory & RAG 相关度: 9/10

Do We Need Bigger Models for Science? Task-Aware Retrieval with Small Language Models

Florian Kelber, Matthias Jobst, Yuni Susanti, Michael Färber
arXiv: 2604.01965v1 发布: 2026-04-02 更新: 2026-04-02

AI 摘要

研究表明,检索增强能在一定程度上弥补小模型的能力不足,但模型容量对复杂推理仍然重要。

主要贡献

  • 设计轻量级的检索增强框架
  • 实现任务感知的检索策略选择
  • 结合全文和元数据进行信息检索
  • 使用小型指令调优语言模型生成带引用的答案

方法论

设计任务感知路由的检索增强框架,整合全文和元数据,使用小型指令调优语言模型进行实验。

原文摘要

Scientific knowledge discovery increasingly relies on large language models, yet many existing scholarly assistants depend on proprietary systems with tens or hundreds of billions of parameters. Such reliance limits reproducibility and accessibility for the research community. In this work, we ask a simple question: do we need bigger models for scientific applications? Specifically, we investigate to what extent carefully designed retrieval pipelines can compensate for reduced model scale in scientific applications. We design a lightweight retrieval-augmented framework that performs task-aware routing to select specialized retrieval strategies based on the input query. The system further integrates evidence from full-text scientific papers and structured scholarly metadata, and employs compact instruction-tuned language models to generate responses with citations. We evaluate the framework across several scholarly tasks, focusing on scholarly question answering (QA), including single- and multi-document scenarios, as well as biomedical QA under domain shift and scientific text compression. Our findings demonstrate that retrieval and model scale are complementary rather than interchangeable. While retrieval design can partially compensate for smaller models, model capacity remains important for complex reasoning tasks. This work highlights retrieval and task-aware design as key factors for building practical and reproducible scholarly assistants.

标签

检索增强 小型语言模型 科学问答 任务感知

arXiv 分类

cs.IR cs.AI cs.CL cs.DL