LLM Reasoning 相关度: 6/10

From Complex Dynamics to DynFormer: Rethinking Transformers for PDEs

Pengyu Lai, Yixiao Chen, Dewu Yang, Rui Wang, Feng Wang, Hui Xu
arXiv: 2603.03112v1 发布: 2026-03-03 更新: 2026-03-03

AI 摘要

DynFormer通过动力学信息指导Transformer,显著降低求解PDE的计算成本和误差。

主要贡献

  • 提出了DynFormer,一种动力学信息驱动的神经算子。
  • 引入Spectral Embedding和Kronecker结构注意力机制,高效捕捉大规模全局交互。
  • 设计Local-Global-Mixing模块,隐式重构小尺度快速变化的湍流级联。
  • 在PDE基准测试中,DynFormer显著降低误差并减少GPU内存消耗。

方法论

DynFormer利用谱嵌入分离尺度,使用Kronecker注意力处理全局,Local-Global-Mixing处理局部,构建混合进化架构。

原文摘要

Partial differential equations (PDEs) are fundamental for modeling complex physical systems, yet classical numerical solvers face prohibitive computational costs in high-dimensional and multi-scale regimes. While Transformer-based neural operators have emerged as powerful data-driven alternatives, they conventionally treat all discretized spatial points as uniform, independent tokens. This monolithic approach ignores the intrinsic scale separation of physical fields, applying computationally prohibitive global attention that redundantly mixes smooth large-scale dynamics with high-frequency fluctuations. Rethinking Transformers through the lens of complex dynamics, we propose DynFormer, a novel dynamics-informed neural operator. Rather than applying a uniform attention mechanism across all scales, DynFormer explicitly assigns specialized network modules to distinct physical scales. It leverages a Spectral Embedding to isolate low-frequency modes, enabling a Kronecker-structured attention mechanism to efficiently capture large-scale global interactions with reduced complexity. Concurrently, we introduce a Local-Global-Mixing transformation. This module utilizes nonlinear multiplicative frequency mixing to implicitly reconstruct the small-scale, fast-varying turbulent cascades that are slaved to the macroscopic state, without incurring the cost of global attention. Integrating these modules into a hybrid evolutionary architecture ensures robust long-term temporal stability. Extensive memory-aligned evaluations across four PDE benchmarks demonstrate that DynFormer achieves up to a 95% reduction in relative error compared to state-of-the-art baselines, while significantly reducing GPU memory consumption. Our results establish that embedding first-principles physical dynamics into Transformer architectures yields a highly scalable, theoretically grounded blueprint for PDE surrogate modeling.

标签

PDE Transformer Neural Operator Scientific Computing Deep Learning

arXiv 分类

cs.LG cs.AI nlin.CD