LLM Reasoning 相关度: 9/10

MoRI: Learning Motivation-Grounded Reasoning for Scientific Ideation in Large Language Models

Chenyang Gu, Jiahao Cheng, Meicong Zhang, Pujun Zheng, Jinquan Zheng, Guoxiu He
arXiv: 2603.19044v1 发布: 2026-03-19 更新: 2026-03-19

AI 摘要

MoRI框架通过动机驱动的推理,提升大语言模型在科学构思方面的技术深度和科学依据。

主要贡献

  • 提出MoRI框架,增强LLM的科学推理能力
  • 引入熵感知信息增益和对比语义增益的强化学习奖励
  • 实验证明MoRI在多个指标上优于现有方法

方法论

基于监督微调初始化LLM,再通过强化学习训练,鼓励模型挖掘技术细节并保持与科学的对齐。

原文摘要

Scientific ideation aims to propose novel solutions within a given scientific context. Existing LLM-based agentic approaches emulate human research workflows, yet inadequately model scientific reasoning, resulting in surface-level conceptual recombinations that lack technical depth and scientific grounding. To address this issue, we propose \textbf{MoRI} (\textbf{Mo}tivation-grounded \textbf{R}easoning for Scientific \textbf{I}deation), a framework that enables LLMs to explicitly learn the reasoning process from research motivations to methodologies. The base LLM is initialized via supervised fine-tuning to generate a research motivation from a given context, and is subsequently trained under a composite reinforcement learning reward that approximates scientific rigor: (1) entropy-aware information gain encourages the model to uncover and elaborate high-complexity technical details grounded in ground-truth methodologies, and (2) contrastive semantic gain constrains the reasoning trajectory to maintain conceptually aligned with scientifically valid solutions. Empirical results show that MoRI significantly outperforms strong commercial LLMs and complex agentic baselines across multiple dimensions, including novelty, technical rigor, and feasibility. The code will be made available on \href{https://github.com/ECNU-Text-Computing/IdeaGeneration}{GitHub}.

标签

科学构思 动机驱动推理 强化学习 大语言模型

arXiv 分类

cs.CL