Separable neural architectures as a primitive for unified predictive and generative intelligence
AI 摘要
提出可分离神经网络架构(SNA),统一预测和生成智能,并在多个领域验证其有效性。
主要贡献
- 提出可分离神经网络架构 (SNA)
- 统一加性、二次和张量分解的神经模型
- 将连续物理状态视为平滑、可分离的嵌入
- 在多个领域证明SNA的通用性
方法论
通过约束交互阶数和张量秩,将高维映射分解为低元组件,实现领域无关的预测和生成建模。
原文摘要
Intelligent systems across physics, language and perception often exhibit factorisable structure, yet are typically modelled by monolithic neural architectures that do not explicitly exploit this structure. The separable neural architecture (SNA) addresses this by formalising a representational class that unifies additive, quadratic and tensor-decomposed neural models. By constraining interaction order and tensor rank, SNAs impose a structural inductive bias that factorises high-dimensional mappings into low-arity components. Separability need not be a property of the system itself: it often emerges in the coordinates or representations through which the system is expressed. Crucially, this coordinate-aware formulation reveals a structural analogy between chaotic spatiotemporal dynamics and linguistic autoregression. By treating continuous physical states as smooth, separable embeddings, SNAs enable distributional modelling of chaotic systems. This approach mitigates the nonphysical drift characteristics of deterministic operators whilst remaining applicable to discrete sequences. The compositional versatility of this approach is demonstrated across four domains: autonomous waypoint navigation via reinforcement learning, inverse generation of multifunctional microstructures, distributional modelling of turbulent flow and neural language modelling. These results establish the separable neural architecture as a domain-agnostic primitive for predictive and generative intelligence, capable of unifying both deterministic and distributional representations.