菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-30
📄 Abstract - Physical Foundation Models: Fixed hardware implementations of large-scale neural networks

Foundation models are deep neural networks (such as GPT-5, Gemini~3, and Opus~4) trained on large datasets that can perform diverse downstream tasks -- text and code generation, question answering, summarization, image classification, and so on. The philosophy of foundation models is to put effort into a single, large (${\sim}10^{12}$-parameter) general-purpose model that can be adapted to many downstream tasks with no or minimal additional training. We argue that the rise of foundation models presents an opportunity for hardware engineers: in contrast to when different models were used for different tasks, it now makes sense to build special-purpose, fixed hardware implementations of neural networks, manufactured and released at the roughly 1-year cadence of major new foundation-model versions. Beyond conventional digital-electronic inference hardware with read-only weight memory, we advocate a more radical re-thinking: hardware in which the neural network is realized directly at the level of the physical design and operates via the hardware's natural physical dynamics -- \textit{Physical Foundation Models} (PFMs). PFMs could enable orders-of-magnitude advantages in energy efficiency, speed, and parameter density. For ${\sim}10^{12}$-parameter models, this would both reduce the high energy burden of AI in datacenters and enable AI in edge devices that today are power-constrained to far smaller models. PFMs could also enable inference hardware for models much larger than current ones: $10^{15}$- or even $10^{18}$-parameter PFMs seem plausible by some measures. We present back-of-the-envelope calculations illustrating PFM scaling using an optical example -- a 3D nanostructured glass medium -- and discuss prospects in nanoelectronics and other physical platforms. We conclude with the major research challenges that must be resolved for trillion-parameter PFMs and beyond to become reality.

顶级标签: systems machine learning
详细标签: foundation models hardware acceleration physical computing energy efficiency neural networks 或 搜索:

物理基础模型:大规模神经网络的固定硬件实现 / Physical Foundation Models: Fixed hardware implementations of large-scale neural networks


1️⃣ 一句话总结

本文提出一种全新思路:将万亿参数级别的基础模型(如GPT-5)直接设计成专用物理硬件,利用光、电子等物理现象进行运算,从而大幅提升能效、速度和参数密度,有望将AI从数据中心扩展到边缘设备,并支持更大规模的模型(如千万亿参数)。

源自 arXiv: 2604.27911