ChatPaper.aiChatPaper

基于球面游走弱监督的算子学习方法

Operator Learning Using Weak Supervision from Walk-on-Spheres

March 1, 2026
作者: Hrishikesh Viswanath, Hong Chul Nam, Xi Deng, Julius Berner, Anima Anandkumar, Aniket Bera
cs.AI

摘要

训练神经偏微分方程求解器常受限于昂贵的数据生成成本,或由于高阶导数导致优化空间复杂而难以稳定的物理信息神经网络。为解决该问题,我们提出一种基于蒙特卡洛方法的替代方案,通过将偏微分方程解估计为随机过程,在训练过程中实施弱监督。利用球面行走算法,我们提出名为球面行走神经算子(WoS-NO)的学习框架,通过WoS提供的弱监督来训练任意给定的神经算子。该方法通过WoS算法的随机表示,将蒙特卡洛游走的成本分摊到偏微分方程实例分布中,在训练过程中生成廉价且含噪声的偏微分方程解估计。这一思路被形式化为无需数据的物理信息优化目标,通过训练神经算子回归这些弱监督信号,使其能够学习整个偏微分方程族的广义解映射。该策略无需昂贵的预计算数据集,避免了内存密集且不稳定的高阶导数损失计算,并展现出对新偏微分方程参数和领域的零样本泛化能力。实验表明,在相同训练步数下,相较于标准物理信息训练方案,本方法在L2误差上最高提升8.75倍,训练速度最高提升6.31倍,GPU内存消耗最高降低2.97倍。代码已发布于https://github.com/neuraloperator/WoS-NO。
English
Training neural PDE solvers is often bottlenecked by expensive data generation or unstable physics-informed neural network (PINN) involving challenging optimization landscapes due to higher-order derivatives. To tackle this issue, we propose an alternative approach using Monte Carlo approaches to estimate the solution to the PDE as a stochastic process for weak supervision during training. Leveraging the Walk-on-Spheres method, we introduce a learning scheme called Walk-on-Spheres Neural Operator (WoS-NO) which uses weak supervision from WoS to train any given neural operator. We propose to amortize the cost of Monte Carlo walks across the distribution of PDE instances using stochastic representations from the WoS algorithm to generate cheap, noisy, estimates of the PDE solution during training. This is formulated into a data-free physics-informed objective where a neural operator is trained to regress against these weak supervisions, allowing the operator to learn a generalized solution map for an entire family of PDEs. This strategy does not require expensive pre-computed datasets, avoids computing higher-order derivatives for loss functions that are memory-intensive and unstable, and demonstrates zero-shot generalization to novel PDE parameters and domains. Experiments show that for the same number of training steps, our method exhibits up to 8.75times improvement in L_2-error compared to standard physics-informed training schemes, up to 6.31times improvement in training speed, and reductions of up to 2.97times in GPU memory consumption. We present the code at https://github.com/neuraloperator/WoS-NO
PDF22May 8, 2026