AI Agents 相关度: 8/10

OMIND: Framework for Knowledge Grounded Finetuning and Multi-Turn Dialogue Benchmark for Mental Health LLMs

Suraj Racha, Prashant Harish Joshi, Utkarsh Maurya, Nitin Yadav, Mridul Sharma, Ananya Kunisetty, Saranya Darisipudi, Nirmal Punjabi, Ganesh Ramakrishnan
arXiv: 2603.25105v1 发布: 2026-03-26 更新: 2026-03-26

AI 摘要

oMind框架针对心理健康领域LLM,提出高质量数据集、训练方法和评估基准。

主要贡献

  • 构建了高质量的心理健康领域多任务SFT数据集 (~164k)。
  • 提出了oMind框架,用于训练和对齐LLM agents。
  • 创建了oMind-Chat,一个多轮对话benchmark数据集,带有专家标注。

方法论

利用结构化知识检索和LLM进行数据生成、剪枝和人工审核,最终得到高质量的训练数据集和评估基准。

原文摘要

Large Language Models (LLMs) have shown remarkable capabilities for complex tasks, yet adaptation in medical domain, specifically mental health, poses specific challenges. Mental health is a rising concern globally with LLMs having large potential to help address the same. We highlight three primary challenges for LLMs in mental health - lack of high quality interpretable and knowledge grounded training data; training paradigms restricted to core capabilities, and evaluation of multi turn dialogue settings. Addressing it, we present oMind framework which includes training and aligning LLM agents for diverse capabilities including conversations; high quality ~164k multi-task SFT dataset, as a result of our generation pipeline based on Structured Knowledge retrieval, LLM based pruning, and review actions. We also introduce oMind-Chat - a novel multi turn benchmark dataset with expert annotated turn level and conversation level rubrics. Our diverse experiments on both core capabilities and conversations shows oMind LLMs consistently outperform baselines. oMind-LLM also shows significantly better reasoning with up to 80% win rate.

标签

Mental Health Large Language Models Fine-tuning Multi-turn Dialogue

arXiv 分类

cs.CL