LLM Reasoning 相关度: 8/10

Large Language Models and Impossible Language Acquisition: "False Promise" or an Overturn of our Current Perspective towards AI

Ziyan wang, Longlong Ma
arXiv: 2602.08437v1 发布: 2026-02-09 更新: 2026-02-09

AI 摘要

论文通过实验和理论分析,探讨LLM在学习不可能语言方面的能力,并对Chomsky的观点提出新的见解。

主要贡献

  • 通过实验验证GPT-2和小模型在学习不可能语言上的表现差异,揭示Transformer架构的重要性。
  • 提出在Chomsky框架内对LLM的新视角,以及从理性主义到功能主义/经验主义的理论范式转变。
  • 构建了一系列基于英语的句法上不可能的语言,为研究LLM的语言学习能力提供了实验基础。

方法论

构建不可能语言数据集,分别在GPT-2和LSTM模型上进行实验,采用Welch's t-test进行统计分析,并结合理论分析。

原文摘要

In Chomsky's provocative critique "The False Promise of CHATGPT," Large Language Models (LLMs) are characterized as mere pattern predictors that do not acquire languages via intrinsic causal and self-correction structures like humans, therefore are not able to distinguish impossible languages. It stands as a representative in a fundamental challenge to the intellectual foundations of AI, for it integrally synthesizes major issues in methodologies within LLMs and possesses an iconic a priori rationalist perspective. We examine this famous critic from both the perspective in pre-existing literature of linguistics and psychology as well as a research based on an experiment inquiring the capacity of learning both possible and impossible languages among LLMs. We constructed a set of syntactically impossible languages by applying certain transformations to English. These include reversing whole sentences, and adding negation based on word-count parity. Two rounds of controlled experiments were each conducted on GPT-2 small models and long short-term memory (LSTM) models. Statistical analysis (Welch's t-test) shows GPT2 small models underperform in learning all of the impossible languages compared to their performance on the possible language (p<.001). On the other hand, LSTM models' performance tallies with Chomsky's argument, suggesting the irreplaceable role of the evolution of transformer architecture. Based on theoretical analysis and empirical findings, we propose a new vision within Chomsky's theory towards LLMs, and a shift of theoretical paradigm outside Chomsky, from his "rationalist-romantics" paradigm to functionalism and empiricism in LLMs research.

标签

LLM Impossible Language Chomsky GPT-2 LSTM

arXiv 分类

cs.CL