两个不混溶的流体的位移是多孔介质中流体流动的常见问题。这种问题可以作为局部微分方程(PDE)构成通常被称为Buckley-Leverett(B-L)问题。 B-L问题是一种非线性双曲守护法,众所周知,使用传统的数值方法难以解决。在这里,我们使用物理信息的神经网络(Pinns)使用非凸版通量函数来解决前向双曲线B-L问题。本文的贡献是双重的。首先,我们通过将Oleinik熵条件嵌入神经网络残差来提出一种Pinn方法来解决双曲线B-L问题。我们不使用扩散术语(人工粘度)在残留损失中,但我们依靠PDE的强形式。其次,我们使用ADAM优化器与基于残留的自适应细化(RAR)算法,实现不加权的超低损耗。我们的解决方案方法可以精确地捕获冲击前并产生精确的整体解决方案。我们报告了一个2 x 10-2的L2验证误差和1x 10-6的L2损耗。所提出的方法不需要任何额外的正则化或加权损失以获得这种准确的解决方案。
translated by 谷歌翻译
Deep learning has achieved remarkable success in diverse applications; however, its use in solving partial differential equations (PDEs) has emerged only recently. Here, we present an overview of physics-informed neural networks (PINNs), which embed a PDE into the loss of the neural network using automatic differentiation. The PINN algorithm is simple, and it can be applied to different types of PDEs, including integro-differential equations, fractional PDEs, and stochastic PDEs. Moreover, from the implementation point of view, PINNs solve inverse problems as easily as forward problems. We propose a new residual-based adaptive refinement (RAR) method to improve the training efficiency of PINNs. For pedagogical reasons, we compare the PINN algorithm to a standard finite element method. We also present a Python library for PINNs, DeepXDE, which is designed to serve both as an education tool to be used in the classroom as well as a research tool for solving problems in computational science and engineering. Specifically, DeepXDE can solve forward problems given initial and boundary conditions, as well as inverse problems given some extra measurements. DeepXDE supports complex-geometry domains based on the technique of constructive solid geometry, and enables the user code to be compact, resembling closely the mathematical formulation. We introduce the usage of DeepXDE and its customizability, and we also demonstrate the capability of PINNs and the user-friendliness of DeepXDE for five different examples. More broadly, DeepXDE contributes to the more rapid development of the emerging Scientific Machine Learning field.
translated by 谷歌翻译
Solute transport in porous media is relevant to a wide range of applications in hydrogeology, geothermal energy, underground CO2 storage, and a variety of chemical engineering systems. Due to the complexity of solute transport in heterogeneous porous media, traditional solvers require high resolution meshing and are therefore expensive computationally. This study explores the application of a mesh-free method based on deep learning to accelerate the simulation of solute transport. We employ Physics-informed Neural Networks (PiNN) to solve solute transport problems in homogeneous and heterogeneous porous media governed by the advection-dispersion equation. Unlike traditional neural networks that learn from large training datasets, PiNNs only leverage the strong form mathematical models to simultaneously solve for multiple dependent or independent field variables (e.g., pressure and solute concentration fields). In this study, we construct PiNN using a periodic activation function to better represent the complex physical signals (i.e., pressure) and their derivatives (i.e., velocity). Several case studies are designed with the intention of investigating the proposed PiNN's capability to handle different degrees of complexity. A manual hyperparameter tuning method is used to find the best PiNN architecture for each test case. Point-wise error and mean square error (MSE) measures are employed to assess the performance of PiNNs' predictions against the ground truth solutions obtained analytically or numerically using the finite element method. Our findings show that the predictions of PiNN are in good agreement with the ground truth solutions while reducing computational complexity and cost by, at least, three orders of magnitude.
translated by 谷歌翻译
物理信息的神经网络(PINN)是神经网络(NNS),它们作为神经网络本身的组成部分编码模型方程,例如部分微分方程(PDE)。如今,PINN是用于求解PDE,分数方程,积分分化方程和随机PDE的。这种新颖的方法已成为一个多任务学习框架,在该框架中,NN必须在减少PDE残差的同时拟合观察到的数据。本文对PINNS的文献进行了全面的综述:虽然该研究的主要目标是表征这些网络及其相关的优势和缺点。该综述还试图将出版物纳入更广泛的基于搭配的物理知识的神经网络,这些神经网络构成了香草·皮恩(Vanilla Pinn)以及许多其他变体,例如物理受限的神经网络(PCNN),各种HP-VPINN,变量HP-VPINN,VPINN,VPINN,变体。和保守的Pinn(CPINN)。该研究表明,大多数研究都集中在通过不同的激活功能,梯度优化技术,神经网络结构和损耗功能结构来定制PINN。尽管使用PINN的应用范围广泛,但通过证明其在某些情况下比有限元方法(FEM)等经典数值技术更可行的能力,但仍有可能的进步,最著名的是尚未解决的理论问题。
translated by 谷歌翻译
We propose characteristic-informed neural networks (CINN), a simple and efficient machine learning approach for solving forward and inverse problems involving hyperbolic PDEs. Like physics-informed neural networks (PINN), CINN is a meshless machine learning solver with universal approximation capabilities. Unlike PINN, which enforces a PDE softly via a multi-part loss function, CINN encodes the characteristics of the PDE in a general-purpose deep neural network trained with the usual MSE data-fitting regression loss and standard deep learning optimization methods. This leads to faster training and can avoid well-known pathologies of gradient descent optimization of multi-part PINN loss functions. If the characteristic ODEs can be solved exactly, which is true in important cases, the output of a CINN is an exact solution of the PDE, even at initialization, preventing the occurrence of non-physical outputs. Otherwise, the ODEs must be solved approximately, but the CINN is still trained only using a data-fitting loss function. The performance of CINN is assessed empirically in forward and inverse linear hyperbolic problems. These preliminary results indicate that CINN is able to improve on the accuracy of the baseline PINN, while being nearly twice as fast to train and avoiding non-physical solutions. Future extensions to hyperbolic PDE systems and nonlinear PDEs are also briefly discussed.
translated by 谷歌翻译
Deep learning techniques with neural networks have been used effectively in computational fluid dynamics (CFD) to obtain solutions to nonlinear differential equations. This paper presents a physics-informed neural network (PINN) approach to solve the Blasius function. This method eliminates the process of changing the non-linear differential equation to an initial value problem. Also, it tackles the convergence issue arising in the conventional series solution. It is seen that this method produces results that are at par with the numerical and conventional methods. The solution is extended to the negative axis to show that PINNs capture the singularity of the function at $\eta=-5.69$
translated by 谷歌翻译
数据驱动学习方法与经典仿真之间的接口造成了一个有趣的字段,提供了多种新应用。在这项工作中,我们建立了物理知识的神经网络(Pinns)的概念,并在浅水方程(SWE)模型中采用它们。这些模型在建模和模拟自由表面流程中起重要作用,例如洪波传播或海啸波。彼此比较Pinn残差的不同配方,并评估多种优化以加速收敛速率。我们用不同的1-D和2-D实验测试这些并最终证明关于具有不同沐浴浴的SWE场景,该方法能够与直接数值模拟相比,具有8.9美元的总相对$ L_2 $误差的直接数值模拟。e-3 $。
translated by 谷歌翻译
科学机器学习(Sciml)的出现在思路科学领域开辟了一个新的领域,通过在基于物理和数据建模的界面的界面中开发方法。为此,近年来介绍了物理知识的神经网络(Pinns),通过在所谓的焊点上纳入物理知识来应对培训数据的稀缺。在这项工作中,我们研究了Pinns关于用于强制基于物理惩罚术语的配偶数量的预测性能。我们表明Pinns可能会失败,学习通过定义来满足物理惩罚术语的琐碎解决方案。我们制定了一种替代的采样方法和新的惩罚术语,使我们能够在具有竞争性结果的数据稀缺设置中纠正Pinns中的核心问题,同时减少最多80 \%的基准问题所需的搭配数量。
translated by 谷歌翻译
地震波的频域模拟在地震反演中起着重要作用,但在大型模型中仍然具有挑战性。作为有效的深度学习方法,最近提出的物理知识的神经网络(PINN)在解决广泛的偏微分方程(PDES)方面取得了成功的应用,并且在这方面仍然有改进的余地。例如,当PDE系数不平滑并描述结构复合介质时,PINN可能导致溶液不准确。在本文中,我们通过使用PINN而不是波方程来求解频域中的声学和Visco声学散射的场波方程,以消除源奇异性。我们首先说明,当在损失函数中未实现边界条件时,非平滑速度模型导致波场不准确。然后,我们在PINN的损耗函数中添加了完美匹配的层(PML)条件,并设计了二次神经网络,以克服PINN中非平滑模型的有害影响。我们表明,PML和二次神经元改善了结果和衰减,并讨论了这种改进的原因。我们还说明,在波场模拟中训练的网络可用于预先训练PDE-Coeff及时改变后另一个波场模拟的神经网络,并相应地提高收敛速度。当两次连续迭代或两个连续的实验之间的模型扰动时,这种预训练策略应在迭代全波形反转(FWI)和时置目标成像中找到应用。
translated by 谷歌翻译
在本文中,我们演示并调查了一些挑战,这些挑战阻碍了使用物理知识的神经网络解决复杂问题的方式。特别是,我们可视化受过训练的模型的损失景观,并在存在物理学的情况下对反向传播梯度进行灵敏度分析。我们的发现表明,现有的方法产生了难以导航的高度非凸损失景观。此外,高阶PDE污染了可能阻碍或防止收敛的反向传播梯度。然后,我们提出了一种新的方法,该方法绕过了高阶PDE操作员的计算并减轻反向传播梯度的污染。为此,我们降低了解决方案搜索空间的维度,并通过非平滑解决方案促进学习问题。我们的配方还提供了一种反馈机制,可帮助我们的模型适应地专注于难以学习的领域的复杂区域。然后,我们通过调整Lagrange乘数方法来提出一个无约束的二重问题。我们运用我们的方法来解决由线性和非线性PDE控制的几个具有挑战性的基准问题。
translated by 谷歌翻译
物理信息神经网络(PINN)能够找到给定边界值问题的解决方案。我们使用有限元方法(FEM)的几个想法来增强工程问题中现有的PINN的性能。当前工作的主要贡献是促进使用主要变量的空间梯度作为分离神经网络的输出。后来,具有较高衍生物的强形式应用于主要变量的空间梯度作为物理约束。此外,该问题的所谓能量形式被应用于主要变量,作为训练的附加约束。所提出的方法仅需要一阶导数来构建物理损失函数。我们讨论了为什么通过不同模型之间的各种比较,这一点是有益的。基于配方混合的PINN和FE方法具有一些相似之处。前者利用神经网络的复杂非线性插值将PDE及其能量形式最小化及其能量形式,而后者则在元素节点借助Shape函数在元素节点上使用相同。我们专注于异质固体,以显示深学习在不同边界条件下在复杂环境中预测解决方案的能力。针对FEM的解决方案对两个原型问题的解决方案进行了检查:弹性和泊松方程(稳态扩散问题)。我们得出的结论是,通过正确设计PINN中的网络体系结构,深度学习模型有可能在没有其他来源的任何可用初始数据中解决异质域中的未知数。最后,关于Pinn和FEM的组合进行了讨论,以在未来的开发中快速准确地设计复合材料。
translated by 谷歌翻译
深入学习被证明是通过物理信息的神经网络(PINNS)求解部分微分方程(PDE)的有效工具。 Pinns将PDE残差嵌入到神经网络的损耗功能中,已成功用于解决各种前向和逆PDE问题。然而,第一代Pinns的一个缺点是它们通常具有许多训练点即使具有有限的准确性。在这里,我们提出了一种新的方法,梯度增强的物理信息的神经网络(GPInns),用于提高Pinns的准确性和培训效率。 GPInns利用PDE残差的梯度信息,并将梯度嵌入损耗功能。我们广泛地测试了GPinns,并证明了GPInns在前进和反向PDE问题中的有效性。我们的数值结果表明,GPInn比贴图更好地表现出较少的训练点。此外,我们将GPIn与基于残留的自适应细化(RAR)的方法组合,一种用于在训练期间自适应地改善训练点分布的方法,以进一步提高GPInn的性能,尤其是具有陡峭梯度的溶液的PDE。
translated by 谷歌翻译
在这项工作中,我们分析了不同程度的不同精度和分段多项式测试函数如何影响变异物理学知情神经网络(VPINN)的收敛速率,同时解决椭圆边界边界值问题,如何影响变异物理学知情神经网络(VPINN)的收敛速率。使用依靠INF-SUP条件的Petrov-Galerkin框架,我们在精确解决方案和合适的计算神经网络的合适的高阶分段插值之间得出了一个先验误差估计。数值实验证实了理论预测并突出了INF-SUP条件的重要性。我们的结果表明,以某种方式违反直觉,对于平滑解决方案,实现高衰减率的最佳策略在选择最低多项式程度的测试功能方面,同时使用适当高精度的正交公式。
translated by 谷歌翻译
我们提出了一种在多孔培养基中使用物理知识的神经网络(PINNS)中多相热力学(THM)过程中的参数鉴定的解决方案策略。我们采用无量纲的理事方程式,特别适合逆问题,我们利用了我们先前工作中开发的顺序多物理Pinn求解器。我们在多个基准问题上验证了所提出的反模型方法,包括Terzaghi的等温固结问题,Barry-Mercer的等温注射产生问题以及非饱和土壤层的非等热整合。我们报告了提出的顺序PINN-THM逆求器的出色性能,从而为将PINNS应用于复杂非线性多物理问题的逆建模铺平了道路。
translated by 谷歌翻译
在本文中,我们开发了一种物理知识的神经网络(PINN)模型,用于具有急剧干扰初始条件的抛物线问题。作为抛物线问题的一个示例,我们考虑具有点(高斯)源初始条件的对流 - 分散方程(ADE)。在$ d $维的ADE中,在初始条件衰减中的扰动随时间$ t $ as $ t^{ - d/2} $,这可能会在Pinn解决方案中造成较大的近似错误。 ADE溶液中的局部大梯度使该方程的残余效率低下的(PINN)拉丁高立方体采样(常见)。最后,抛物线方程的PINN解对损耗函数中的权重选择敏感。我们提出了一种归一化的ADE形式,其中溶液的初始扰动不会降低幅度,并证明该归一化显着降低了PINN近似误差。我们提出了与通过其他方法选择的权重相比,损耗函数中的权重标准更准确。最后,我们提出了一种自适应采样方案,该方案可显着减少相同数量的采样(残差)点的PINN溶液误差。我们证明了提出的PINN模型的前进,反向和向后ADE的准确性。
translated by 谷歌翻译
概率密度演化的推导提供了对许多随机系统及其性能的行为的宝贵洞察力。但是,对于大多数实时应用程序,对概率密度演变的数值确定是一项艰巨的任务。后者是由于所需的时间和空间离散方案引起的,这些方案使大多数计算解决方案过于效率和不切实际。在这方面,有效的计算替代模型的开发至关重要。关于物理受限网络的最新研究表明,可以通过编码对深神经网络的物理洞察力来实现合适的替代物。为此,目前的工作介绍了Deeppdem,它利用物理信息网络的概念通过提出深度学习方法来解决概率密度的演变。 Deeppdem了解随机结构的一般密度演化方程(GDEE)。这种方法为无网格学习方法铺平了道路,该方法可以通过以前的模拟数据解决密度演化问题。此外,它还可以作为优化方案或实时应用程序中任何其他时空点的溶液的有效替代物。为了证明所提出的框架的潜在适用性,研究了两个具有不同激活功能的网络体系结构以及两个优化器。关于三个不同问题的数值实施验证了所提出方法的准确性和功效。
translated by 谷歌翻译
Recent years have witnessed a growth in mathematics for deep learning--which seeks a deeper understanding of the concepts of deep learning with mathematics, and explores how to make it more robust--and deep learning for mathematics, where deep learning algorithms are used to solve problems in mathematics. The latter has popularised the field of scientific machine learning where deep learning is applied to problems in scientific computing. Specifically, more and more neural network architectures have been developed to solve specific classes of partial differential equations (PDEs). Such methods exploit properties that are inherent to PDEs and thus solve the PDEs better than classical feed-forward neural networks, recurrent neural networks, and convolutional neural networks. This has had a great impact in the area of mathematical modeling where parametric PDEs are widely used to model most natural and physical processes arising in science and engineering, In this work, we review such methods and extend them for parametric studies as well as for solving the related inverse problems. We equally proceed to show their relevance in some industrial applications.
translated by 谷歌翻译
Despite great progress in simulating multiphysics problems using the numerical discretization of partial differential equations (PDEs), one still cannot seamlessly incorporate noisy data into existing algorithms, mesh generation remains complex, and high-dimensional problems governed by parameterized PDEs cannot be tackled. Moreover, solving inverse problems with hidden physics is often prohibitively expensive and requires different formulations and elaborate computer codes. Machine learning has emerged as a promising alternative, but training deep neural networks requires big data, not always available for scientific problems. Instead, such networks can be trained from additional information obtained by enforcing the physical laws (for example, at random points in the continuous space-time domain). Such physics-informed learning integrates (noisy) data and mathematical models, and implements them through neural networks or other kernel-based regression networks. Moreover, it may be possible to design specialized network architectures that automatically satisfy some of the physical invariants for better accuracy, faster training and improved generalization. Here, we review some of the prevailing trends in embedding physics into machine learning, present some of the current capabilities and limitations and discuss diverse applications of physics-informed learning both for forward and inverse problems, including discovering hidden physics and tackling high-dimensional problems.
translated by 谷歌翻译
我们引入了一种实用方法,以实施由神经网络(NNS)定义的功能的线性偏微分方程(PDE)约束,直至所需的公差。通过将隐式函数定理的可区分物理和应用中的应用结合到NN模型中,我们开发了可区分的PDE受限的NN层。在培训期间,我们的模型学习了一个功能系列,每个功能都定义了从PDE参数到PDE解决方案的映射。在推理时,该模型通过解决PDE受限的优化问题来发现学到家族中功能的最佳线性组合。我们的方法提供了有关感兴趣领域的连续解决方案,这些解决方案完全满足了所需的物理约束。我们的结果表明,与对无约束目标的训练相比,将硬约束直接纳入NN体系结构的测试错误要低得多。
translated by 谷歌翻译
科学和工程学中的一个基本问题是设计最佳的控制政策,这些政策将给定的系统转向预期的结果。这项工作提出了同时求解给定系统状态和最佳控制信号的控制物理信息的神经网络(控制PINNS),在符合基础物理定律的一个阶段框架中。先前的方法使用两个阶段的框架,该框架首先建模然后按顺序控制系统。相比之下,控制PINN将所需的最佳条件纳入其体系结构和损耗函数中。通过解决以下开环的最佳控制问题来证明控制PINN的成功:(i)一个分析问题,(ii)一维热方程,以及(iii)二维捕食者捕食者问题。
translated by 谷歌翻译