论文标题
物理学从稀疏观测中限制了数据驱动的逆建模的学习
Physics Constrained Learning for Data-driven Inverse Modeling from Sparse Observations
论文作者
论文摘要
深度神经网络(DNN)已用于建模物理量之间的非线性关系。这些DNN嵌入了由部分微分方程(PDE)描述的物理系统中,并通过最大程度地减少损失函数来训练,该损失函数可以测量某些选定的规范中预测和观察结果之间的差异。当只有稀疏观测值可用时,此损耗函数通常包括PDE约束作为罚款项。结果,只有溶液大致满足PDE。但是,罚款术语通常会减慢优化器对严格问题的收敛性。我们提出了一种新的方法,该方法可以训练嵌入式DNN,同时数值满足PDE约束。我们开发了一种算法,该算法可以在反向模式自动分化中区分显式和隐式数值求解器。这允许在统一框架中计算DNN和PDE求解器的梯度。我们证明,与惩罚方法相比,我们的方法在相对严重的问题中具有更快的收敛性和更好的稳定性。我们的方法允许潜力求解和加速广泛的数据驱动的逆建模,其中物理约束由PDE描述,并且需要准确地满足。
Deep neural networks (DNN) have been used to model nonlinear relations between physical quantities. Those DNNs are embedded in physical systems described by partial differential equations (PDE) and trained by minimizing a loss function that measures the discrepancy between predictions and observations in some chosen norm. This loss function often includes the PDE constraints as a penalty term when only sparse observations are available. As a result, the PDE is only satisfied approximately by the solution. However, the penalty term typically slows down the convergence of the optimizer for stiff problems. We present a new approach that trains the embedded DNNs while numerically satisfying the PDE constraints. We develop an algorithm that enables differentiating both explicit and implicit numerical solvers in reverse-mode automatic differentiation. This allows the gradients of the DNNs and the PDE solvers to be computed in a unified framework. We demonstrate that our approach enjoys faster convergence and better stability in relatively stiff problems compared to the penalty method. Our approach allows for the potential to solve and accelerate a wide range of data-driven inverse modeling, where the physical constraints are described by PDEs and need to be satisfied accurately.