物理知情的神经网络(PINN)已获得极大的流行,作为用于求解PDE的替代方法。尽管取得了经验的成功,但我们仍在对培训对梯度下降的这种约束的融合特性建立了解。众所周知,在没有明确的归纳偏见的情况下,神经网络可能会以样本有效的方式学习或近似简单且知名的功能。因此,从少数搭配点诱导的数值近似可能无法概括整个域。同时,符号形式可以表现出良好的概括,并具有可解释性为有用的副产品。但是,符号近似可能会同时简洁明了。因此,在这项工作中,我们探索了一种神经肌符号方法,以近似PDE的溶液。我们观察到我们的方法在几个简单的情况下起作用。我们说明了我们方法对Navier Stokes的功效:Kovasznay流动,其中有多个物理量的兴趣,该物理数量由非线性耦合PDE系统控制。域分裂现在已成为帮助PINNS近似复杂功能的流行技巧。我们观察到神经肌符号方法也可以帮助这种复杂的功能。我们在暂时变化的二维汉堡方程上展示了域分裂的辅助神经符号方法。最后,我们考虑了必须解决参数化PDE的PINN的情况,以改变初始条件和PDE系数的变化。超级核武器已证明有望克服这些挑战。我们表明,可以设计超启动的网络,以结合速度的好处和提高准确性。我们观察到,神经词近似值始终是1-2个数量级,而不是神经或符号近似值。
translated by 谷歌翻译
机器学习方法最近在求解部分微分方程(PDE)中的承诺。它们可以分为两种广泛类别:近似解决方案功能并学习解决方案操作员。物理知识的神经网络(PINN)是前者的示例,而傅里叶神经操作员(FNO)是后者的示例。这两种方法都有缺点。 Pinn的优化是具有挑战性,易于发生故障,尤其是在多尺度动态系统上。 FNO不会遭受这种优化问题,因为它在给定的数据集上执行了监督学习,但获取此类数据可能太昂贵或无法使用。在这项工作中,我们提出了物理知识的神经运营商(Pino),在那里我们结合了操作学习和功能优化框架。这种综合方法可以提高PINN和FNO模型的收敛速度和准确性。在操作员学习阶段,Pino在参数PDE系列的多个实例上学习解决方案操作员。在测试时间优化阶段,Pino优化预先训练的操作员ANSATZ,用于PDE的查询实例。实验显示Pino优于许多流行的PDE家族的先前ML方法,同时保留与求解器相比FNO的非凡速度。特别是,Pino准确地解决了挑战的长时间瞬态流量,而其他基线ML方法无法收敛的Kolmogorov流程。
translated by 谷歌翻译
Solute transport in porous media is relevant to a wide range of applications in hydrogeology, geothermal energy, underground CO2 storage, and a variety of chemical engineering systems. Due to the complexity of solute transport in heterogeneous porous media, traditional solvers require high resolution meshing and are therefore expensive computationally. This study explores the application of a mesh-free method based on deep learning to accelerate the simulation of solute transport. We employ Physics-informed Neural Networks (PiNN) to solve solute transport problems in homogeneous and heterogeneous porous media governed by the advection-dispersion equation. Unlike traditional neural networks that learn from large training datasets, PiNNs only leverage the strong form mathematical models to simultaneously solve for multiple dependent or independent field variables (e.g., pressure and solute concentration fields). In this study, we construct PiNN using a periodic activation function to better represent the complex physical signals (i.e., pressure) and their derivatives (i.e., velocity). Several case studies are designed with the intention of investigating the proposed PiNN's capability to handle different degrees of complexity. A manual hyperparameter tuning method is used to find the best PiNN architecture for each test case. Point-wise error and mean square error (MSE) measures are employed to assess the performance of PiNNs' predictions against the ground truth solutions obtained analytically or numerically using the finite element method. Our findings show that the predictions of PiNN are in good agreement with the ground truth solutions while reducing computational complexity and cost by, at least, three orders of magnitude.
translated by 谷歌翻译
Deep learning has achieved remarkable success in diverse applications; however, its use in solving partial differential equations (PDEs) has emerged only recently. Here, we present an overview of physics-informed neural networks (PINNs), which embed a PDE into the loss of the neural network using automatic differentiation. The PINN algorithm is simple, and it can be applied to different types of PDEs, including integro-differential equations, fractional PDEs, and stochastic PDEs. Moreover, from the implementation point of view, PINNs solve inverse problems as easily as forward problems. We propose a new residual-based adaptive refinement (RAR) method to improve the training efficiency of PINNs. For pedagogical reasons, we compare the PINN algorithm to a standard finite element method. We also present a Python library for PINNs, DeepXDE, which is designed to serve both as an education tool to be used in the classroom as well as a research tool for solving problems in computational science and engineering. Specifically, DeepXDE can solve forward problems given initial and boundary conditions, as well as inverse problems given some extra measurements. DeepXDE supports complex-geometry domains based on the technique of constructive solid geometry, and enables the user code to be compact, resembling closely the mathematical formulation. We introduce the usage of DeepXDE and its customizability, and we also demonstrate the capability of PINNs and the user-friendliness of DeepXDE for five different examples. More broadly, DeepXDE contributes to the more rapid development of the emerging Scientific Machine Learning field.
translated by 谷歌翻译
随着计算能力的增加和机器学习的进步,基于数据驱动的学习方法在解决PDE方面引起了极大的关注。物理知识的神经网络(PINN)最近出现并成功地在各种前进和逆PDES问题中取得了成功,其优异的特性,例如灵活性,无网格解决方案和无监督的培训。但是,它们的收敛速度较慢和相对不准确的解决方案通常会限制其在许多科学和工程领域中的更广泛适用性。本文提出了一种新型的数据驱动的PDES求解器,物理知识的细胞表示(Pixel),优雅地结合了经典数值方法和基于学习的方法。我们采用来自数值方法的网格结构,以提高准确性和收敛速度并克服PINN中呈现的光谱偏差。此外,所提出的方法在PINN中具有相同的好处,例如,使用相同的优化框架来解决前进和逆PDE问题,并很容易通过现代自动分化技术强制执行PDE约束。我们为原始Pinn所努力的各种具有挑战性的PDE提供了实验结果,并表明像素达到了快速收敛速度和高精度。
translated by 谷歌翻译
在本文中,我们开发了一种物理知识的神经网络(PINN)模型,用于具有急剧干扰初始条件的抛物线问题。作为抛物线问题的一个示例,我们考虑具有点(高斯)源初始条件的对流 - 分散方程(ADE)。在$ d $维的ADE中,在初始条件衰减中的扰动随时间$ t $ as $ t^{ - d/2} $,这可能会在Pinn解决方案中造成较大的近似错误。 ADE溶液中的局部大梯度使该方程的残余效率低下的(PINN)拉丁高立方体采样(常见)。最后,抛物线方程的PINN解对损耗函数中的权重选择敏感。我们提出了一种归一化的ADE形式,其中溶液的初始扰动不会降低幅度,并证明该归一化显着降低了PINN近似误差。我们提出了与通过其他方法选择的权重相比,损耗函数中的权重标准更准确。最后,我们提出了一种自适应采样方案,该方案可显着减少相同数量的采样(残差)点的PINN溶液误差。我们证明了提出的PINN模型的前进,反向和向后ADE的准确性。
translated by 谷歌翻译
We present a unified hard-constraint framework for solving geometrically complex PDEs with neural networks, where the most commonly used Dirichlet, Neumann, and Robin boundary conditions (BCs) are considered. Specifically, we first introduce the "extra fields" from the mixed finite element method to reformulate the PDEs so as to equivalently transform the three types of BCs into linear forms. Based on the reformulation, we derive the general solutions of the BCs analytically, which are employed to construct an ansatz that automatically satisfies the BCs. With such a framework, we can train the neural networks without adding extra loss terms and thus efficiently handle geometrically complex PDEs, alleviating the unbalanced competition between the loss terms corresponding to the BCs and PDEs. We theoretically demonstrate that the "extra fields" can stabilize the training process. Experimental results on real-world geometrically complex PDEs showcase the effectiveness of our method compared with state-of-the-art baselines.
translated by 谷歌翻译
科学和工程学中的一个基本问题是设计最佳的控制政策,这些政策将给定的系统转向预期的结果。这项工作提出了同时求解给定系统状态和最佳控制信号的控制物理信息的神经网络(控制PINNS),在符合基础物理定律的一个阶段框架中。先前的方法使用两个阶段的框架,该框架首先建模然后按顺序控制系统。相比之下,控制PINN将所需的最佳条件纳入其体系结构和损耗函数中。通过解决以下开环的最佳控制问题来证明控制PINN的成功:(i)一个分析问题,(ii)一维热方程,以及(iii)二维捕食者捕食者问题。
translated by 谷歌翻译
物理信息的神经网络(PINN)是神经网络(NNS),它们作为神经网络本身的组成部分编码模型方程,例如部分微分方程(PDE)。如今,PINN是用于求解PDE,分数方程,积分分化方程和随机PDE的。这种新颖的方法已成为一个多任务学习框架,在该框架中,NN必须在减少PDE残差的同时拟合观察到的数据。本文对PINNS的文献进行了全面的综述:虽然该研究的主要目标是表征这些网络及其相关的优势和缺点。该综述还试图将出版物纳入更广泛的基于搭配的物理知识的神经网络,这些神经网络构成了香草·皮恩(Vanilla Pinn)以及许多其他变体,例如物理受限的神经网络(PCNN),各种HP-VPINN,变量HP-VPINN,VPINN,VPINN,变体。和保守的Pinn(CPINN)。该研究表明,大多数研究都集中在通过不同的激活功能,梯度优化技术,神经网络结构和损耗功能结构来定制PINN。尽管使用PINN的应用范围广泛,但通过证明其在某些情况下比有限元方法(FEM)等经典数值技术更可行的能力,但仍有可能的进步,最著名的是尚未解决的理论问题。
translated by 谷歌翻译
Physics-informed neural networks (PINNs) have lately received significant attention as a representative deep learning-based technique for solving partial differential equations (PDEs). Most fully connected network-based PINNs use automatic differentiation to construct loss functions that suffer from slow convergence and difficult boundary enforcement. In addition, although convolutional neural network (CNN)-based PINNs can significantly improve training efficiency, CNNs have difficulty in dealing with irregular geometries with unstructured meshes. Therefore, we propose a novel framework based on graph neural networks (GNNs) and radial basis function finite difference (RBF-FD). We introduce GNNs into physics-informed learning to better handle irregular domains with unstructured meshes. RBF-FD is used to construct a high-precision difference format of the differential equations to guide model training. Finally, we perform numerical experiments on Poisson and wave equations on irregular domains. We illustrate the generalizability, accuracy, and efficiency of the proposed algorithms on different PDE parameters, numbers of collection points, and several types of RBFs.
translated by 谷歌翻译
深入学习被证明是通过物理信息的神经网络(PINNS)求解部分微分方程(PDE)的有效工具。 Pinns将PDE残差嵌入到神经网络的损耗功能中,已成功用于解决各种前向和逆PDE问题。然而,第一代Pinns的一个缺点是它们通常具有许多训练点即使具有有限的准确性。在这里,我们提出了一种新的方法,梯度增强的物理信息的神经网络(GPInns),用于提高Pinns的准确性和培训效率。 GPInns利用PDE残差的梯度信息,并将梯度嵌入损耗功能。我们广泛地测试了GPinns,并证明了GPInns在前进和反向PDE问题中的有效性。我们的数值结果表明,GPInn比贴图更好地表现出较少的训练点。此外,我们将GPIn与基于残留的自适应细化(RAR)的方法组合,一种用于在训练期间自适应地改善训练点分布的方法,以进一步提高GPInn的性能,尤其是具有陡峭梯度的溶液的PDE。
translated by 谷歌翻译
部分微分方程(PDE)在研究大量科学和工程问题方面发挥着至关重要的作用。数值求解的非线性和/或高维PDE通常是一个具有挑战性的任务。灵感来自传统有限差分和有限元的方法和机器学习的新兴进步,我们提出了一个名为神经PDE的序列深度学习框架,这允许通过使用双向来自动学习从现有数据的任何时间依赖于现有数据的管理规则LSTM编码器,并预测下一个时间步长数据。我们所提出的框架的一个关键特征是,神经PDE能够同时学习和模拟多尺度变量。我们通过一维PDE的一系列示例测试神经PDE到高维和非线性复杂流体模型。结果表明,神经PDE能够学习初始条件,边界条件和差分运营商,而不知道PDE系统的特定形式。在我们的实验中,神经PDE可以有效地提取20个时期训练内的动态,并产生准确的预测。此外,与在学习PDE中的传统机器学习方法不同,例如CNN和MLP,这需要用于模型精度的巨大参数,神经PDE在所有时间步骤中共享参数,从而显着降低了计算复杂性并导致快速学习算法。
translated by 谷歌翻译
从经典动力学系统到量子力学的许多领域,在许多领域的进步核心,有效,准确地求解微分方程。人们对使用物理知识的神经网络(PINN)来解决此类问题,这引起了人们的兴趣,因为它们比传统的数值方法提供了许多好处。尽管它们在求解微分方程方面的潜在好处,但仍在探索转移学习。在这项研究中,我们提出了转移学习PINN的一般框架,该框架对普通和部分微分方程的线性系统进行了单次推断。这意味着,可以在不重新培训整个网络的情况下即时获得许多未知微分方程的方法。我们通过解决了几个现实世界中的问题,例如一阶线性普通方程,泊松方程以及时间依赖时间依赖的schrodinger复合物配合物部分差分方程来证明拟议的深度学习方法的功效。
translated by 谷歌翻译
Given ample experimental data from a system governed by differential equations, it is possible to use deep learning techniques to construct the underlying differential operators. In this work we perform symbolic discovery of differential operators in a situation where there is sparse experimental data. This small data regime in machine learning can be made tractable by providing our algorithms with prior information about the underlying dynamics. Physics Informed Neural Networks (PINNs) have been very successful in this regime (reconstructing entire ODE solutions using only a single point or entire PDE solutions with very few measurements of the initial condition). We modify the PINN approach by adding a neural network that learns a representation of unknown hidden terms in the differential equation. The algorithm yields both a surrogate solution to the differential equation and a black-box representation of the hidden terms. These hidden term neural networks can then be converted into symbolic equations using symbolic regression techniques like AI Feynman. In order to achieve convergence of these neural networks, we provide our algorithms with (noisy) measurements of both the initial condition as well as (synthetic) experimental data obtained at later times. We demonstrate strong performance of this approach even when provided with very few measurements of noisy data in both the ODE and PDE regime.
translated by 谷歌翻译
Recent years have witnessed a growth in mathematics for deep learning--which seeks a deeper understanding of the concepts of deep learning with mathematics, and explores how to make it more robust--and deep learning for mathematics, where deep learning algorithms are used to solve problems in mathematics. The latter has popularised the field of scientific machine learning where deep learning is applied to problems in scientific computing. Specifically, more and more neural network architectures have been developed to solve specific classes of partial differential equations (PDEs). Such methods exploit properties that are inherent to PDEs and thus solve the PDEs better than classical feed-forward neural networks, recurrent neural networks, and convolutional neural networks. This has had a great impact in the area of mathematical modeling where parametric PDEs are widely used to model most natural and physical processes arising in science and engineering, In this work, we review such methods and extend them for parametric studies as well as for solving the related inverse problems. We equally proceed to show their relevance in some industrial applications.
translated by 谷歌翻译
许多物理过程,例如天气现象或流体力学由部分微分方程(PDE)管辖。使用神经网络建模这种动态系统是一个新兴的研究领域。然而,目前的方法以各种方式限制:它们需要关于控制方程的先验知识,并限于线性或一阶方程。在这项工作中,我们提出了一种将卷积神经网络(CNNS)与可微分的颂歌求解器结合到模型动力系统的模型。我们表明,标准PDE求解器中使用的线路方法可以使用卷曲来表示,这使得CNN是对参数化任意PDE动态的自然选择。我们的模型可以应用于任何数据而不需要任何关于管理PDE的知识。我们评估通过求解各种PDE而产生的数据集的NeuralPDE,覆盖更高的订单,非线性方程和多个空间尺寸。
translated by 谷歌翻译
机器学习方法最近已用于求解微分方程和动态系统。这些方法已发展为一个新型的研究领域,称为科学机器学习,其中深层神经网络和统计学习等技术应用于应用数学的经典问题。由于神经网络提供了近似能力,因此在求解各种偏微分方程(PDE)时,通过机器学习和优化方法通过机器学习和优化方法实现了明显的性能。在本文中,我们开发了一种新颖的数值算法,该算法结合了机器学习和人工智能来解决PDE。特别是,我们基于Legendre-Galerkin神经网络提出了一种无监督的机器学习算法,以找到与不同类型PDE的解决方案的准确近似值。提出的神经网络应用于一般的1D和2D PDE,以及具有边界层行为的奇异扰动PDE。
translated by 谷歌翻译
We shed light on a pitfall and an opportunity in physics-informed neural networks (PINNs). We prove that a multilayer perceptron (MLP) only with ReLU (Rectified Linear Unit) or ReLU-like Lipschitz activation functions will always lead to a vanished Hessian. Such a network-imposed constraint contradicts any second- or higher-order partial differential equations (PDEs). Therefore, a ReLU-based MLP cannot form a permissible function space for the approximation of their solutions. Inspired by this pitfall, we prove that a linear PDE up to the $n$-th order can be strictly satisfied by an MLP with $C^n$ activation functions when the weights of its output layer lie on a certain hyperplane, as called the out-layer-hyperplane. An MLP equipped with the out-layer-hyperplane becomes "physics-enforced", no longer requiring a loss function for the PDE itself (but only those for the initial and boundary conditions). Such a hyperplane exists not only for MLPs but for any network architecture tailed by a fully-connected hidden layer. To our knowledge, this should be the first PINN architecture that enforces point-wise correctness of a PDE. We give the closed-form expression of the out-layer-hyperplane for second-order linear PDEs and provide an implementation.
translated by 谷歌翻译
深度学习方法的应用加快了挑战性电流问题的分辨率,最近显示出令人鼓舞的结果。但是,电力系统动力学不是快照,稳态操作。必须考虑这些动力学,以确保这些模型提供的最佳解决方案遵守实用的动力约束,避免频率波动和网格不稳定性。不幸的是,由于其高计算成本,基于普通或部分微分方程的动态系统模型通常不适合在控制或状态估计中直接应用。为了应对这些挑战,本文介绍了一种机器学习方法,以近乎实时近似电力系统动态的行为。该拟议的框架基于梯度增强的物理知识的神经网络(GPINNS),并编码有关电源系统的基本物理定律。拟议的GPINN的关键特征是它的训练能力而无需生成昂贵的培训数据。该论文说明了在单机无限总线系统中提出的方法在预测转子角度和频率的前进和反向问题中的潜力,以及不确定的参数,例如惯性和阻尼,以展示其在一系列电力系统应用中的潜力。
translated by 谷歌翻译
在本文中,我们提出了用于求解非线性微分方程(NDE)的神经网络的物理知情训练(PIAT)。众所周知,神经网络的标准培训会导致非平滑函数。对抗训练(AT)是针对对抗攻击的既定防御机制,这也可能有助于使解决方案平滑。 AT包括通过扰动增强训练迷你批量,使网络输出不匹配所需的输出对手。与正式AT仅依靠培训数据不同,在这里,我们使用对抗网络体系结构中的自动差异来以非线性微分方程的形式编码管理物理定律。我们将PIAT与PIAT进行了比较,以指示我们方法在求解多达10个维度方面的有效性。此外,我们提出了重量衰减和高斯平滑,以证明PIAT的优势。代码存储库可从https://github.com/rohban-lab/piat获得。
translated by 谷歌翻译