可分离神经架构:作为统一预测与生成智能的基础组件 / Separable neural architectures as a primitive for unified predictive and generative intelligence
1️⃣ 一句话总结
这篇论文提出了一种名为‘可分离神经架构’的新模型,它通过将复杂系统分解为低维度的简单组件,统一了从物理混沌系统到语言生成等多种预测和生成任务,为解决高维问题提供了一个通用且高效的方法。
Intelligent systems across physics, language and perception often exhibit factorisable structure, yet are typically modelled by monolithic neural architectures that do not explicitly exploit this structure. The separable neural architecture (SNA) addresses this by formalising a representational class that unifies additive, quadratic and tensor-decomposed neural models. By constraining interaction order and tensor rank, SNAs impose a structural inductive bias that factorises high-dimensional mappings into low-arity components. Separability need not be a property of the system itself: it often emerges in the coordinates or representations through which the system is expressed. Crucially, this coordinate-aware formulation reveals a structural analogy between chaotic spatiotemporal dynamics and linguistic autoregression. By treating continuous physical states as smooth, separable embeddings, SNAs enable distributional modelling of chaotic systems. This approach mitigates the nonphysical drift characteristics of deterministic operators whilst remaining applicable to discrete sequences. The compositional versatility of this approach is demonstrated across four domains: autonomous waypoint navigation via reinforcement learning, inverse generation of multifunctional microstructures, distributional modelling of turbulent flow and neural language modelling. These results establish the separable neural architecture as a domain-agnostic primitive for predictive and generative intelligence, capable of unifying both deterministic and distributional representations.
可分离神经架构:作为统一预测与生成智能的基础组件 / Separable neural architectures as a primitive for unified predictive and generative intelligence
这篇论文提出了一种名为‘可分离神经架构’的新模型,它通过将复杂系统分解为低维度的简单组件,统一了从物理混沌系统到语言生成等多种预测和生成任务,为解决高维问题提供了一个通用且高效的方法。
源自 arXiv: 2603.12244