LLM Reasoning 相关度: 9/10

Can Large Language Models Reason and Optimize Under Constraints?

Fabien Bernier, Salah Ghamizi, Pantelis Dogoulis, Maxime Cordy
arXiv: 2603.23004v1 发布: 2026-03-24 更新: 2026-03-24

AI 摘要

该论文评估了LLM在受约束优化问题(电力系统最优潮流问题)上的推理和优化能力,发现现有LLM表现不佳。

主要贡献

  • 提出了一个评估LLM在受约束优化问题上的能力的新框架。
  • 揭示了现有LLM在处理结构化推理和约束优化方面的不足。
  • 为开发能够解决实际电力系统优化问题的LLM助手提供了 rigorous 的测试环境。

方法论

通过构建Optimal Power Flow (OPF)问题,测试LLM在推理、结构化输入处理、算术和约束优化方面的能力,并评估SoTA LLM的性能。

原文摘要

Large Language Models (LLMs) have demonstrated great capabilities across diverse natural language tasks; yet their ability to solve abstraction and optimization problems with constraints remains scarcely explored. In this paper, we investigate whether LLMs can reason and optimize under the physical and operational constraints of Optimal Power Flow (OPF) problem. We introduce a challenging evaluation setup that requires a set of fundamental skills such as reasoning, structured input handling, arithmetic, and constrained optimization. Our evaluation reveals that SoTA LLMs fail in most of the tasks, and that reasoning LLMs still fail in the most complex settings. Our findings highlight critical gaps in LLMs' ability to handle structured reasoning under constraints, and this work provides a rigorous testing environment for developing more capable LLM assistants that can tackle real-world power grid optimization problems.

标签

LLM Reasoning Optimization Constraints Power Systems

arXiv 分类

cs.AI cs.LG