物理信息神经网络:完整训练周期的教学式推导 / Physics-Informed Neural Networks: A Didactic Derivation of the Complete Training Cycle
1️⃣ 一句话总结
这篇论文通过一个具体数值示例,手把手地推导了物理信息神经网络从正向传播、计算损失到反向传播梯度并更新参数的完整训练过程,旨在揭示通常被自动微分库隐藏的底层数学原理。
This paper is a step-by-step, self-contained guide to the complete training cycle of a Physics-Informed Neural Network (PINN) -- a topic that existing tutorials and guides typically delegate to automatic differentiation libraries without exposing the underlying algebra. Using a first-order initial value problem with a known analytical solution as a running example, we walk through every stage of the process: forward propagation of both the network output and its temporal derivative, evaluation of a composite loss function built from the ODE residual and the initial condition, backpropagation of gradients -- with particular attention to the product rule that arises in hidden layers -- and a gradient descent parameter update. Every calculation is presented with explicit, verifiable numerical values using a 1-3-3-1 multilayer perceptron with two hidden layers and 22 trainable parameters. From these concrete examples, we derive general recursive formulas -- expressed as sensitivity propagation relations -- that extend the gradient computation to networks of arbitrary depth, and we connect these formulas to the automatic differentiation engines used in practice. The trained network is then validated against the exact solution, achieving a relative $L^2$ error of $4.290 \times 10^{-4}$ using only the physics-informed loss, without any data from the true solution. A companion Jupyter/PyTorch notebook reproduces every manual calculation and the full training pipeline, providing mutual validation between hand-derived and machine-computed gradients.
物理信息神经网络:完整训练周期的教学式推导 / Physics-Informed Neural Networks: A Didactic Derivation of the Complete Training Cycle
这篇论文通过一个具体数值示例,手把手地推导了物理信息神经网络从正向传播、计算损失到反向传播梯度并更新参数的完整训练过程,旨在揭示通常被自动微分库隐藏的底层数学原理。
源自 arXiv: 2604.18481