LLM Reasoning 相关度: 8/10

When Language Models Lose Their Mind: The Consequences of Brain Misalignment

Gabriele Merlin, Mariya Toneva
arXiv: 2603.23091v1 发布: 2026-03-24 更新: 2026-03-24

AI 摘要

研究表明,脑对齐对于LLM的语言能力至关重要,脑失调会导致下游任务性能显著下降。

主要贡献

  • 提出了脑失调LLM的概念
  • 评估了脑对齐对LLM语言能力的影响
  • 证明了脑对齐对于LLM鲁棒语言理解的重要性

方法论

训练脑失调LLM,通过200多个下游任务与脑对齐模型对比,评估其语言理解能力。

原文摘要

While brain-aligned large language models (LLMs) have garnered attention for their potential as cognitive models and for potential for enhanced safety and trustworthiness in AI, the role of this brain alignment for linguistic competence remains uncertain. In this work, we investigate the functional implications of brain alignment by introducing brain-misaligned models--LLMs intentionally trained to predict brain activity poorly while maintaining high language modeling performance. We evaluate these models on over 200 downstream tasks encompassing diverse linguistic domains, including semantics, syntax, discourse, reasoning, and morphology. By comparing brain-misaligned models with well-matched brain-aligned counterparts, we isolate the specific impact of brain alignment on language understanding. Our experiments reveal that brain misalignment substantially impairs downstream performance, highlighting the critical role of brain alignment in achieving robust linguistic competence. These findings underscore the importance of brain alignment in LLMs and offer novel insights into the relationship between neural representations and linguistic processing.

标签

LLM 脑对齐 语言理解 神经网络

arXiv 分类

cs.CL