计算机代数可以使用符号算法回答有关部分微分方程的各种问题。但是,在计算机代数中将数据包含在方程式中很少。因此,最近,计算机代数模型与高斯流程(机器学习中的回归模型)相结合,以描述数据下某些微分方程的行为。尽管可以在这种情况下描述多项式边界条件,但我们将这些模型扩展到分析边界条件。此外,我们描述了具有某些分析系数的Weyl代数的gr \ obner和Janet碱基的必要算法。使用这些算法,我们提供了由分析功能界定并适应观察结果的域中无差流流的示例。
translated by 谷歌翻译
Partial differential equations (PDEs) are important tools to model physical systems, and including them into machine learning models is an important way of incorporating physical knowledge. Given any system of linear PDEs with constant coefficients, we propose a family of Gaussian process (GP) priors, which we call EPGP, such that all realizations are exact solutions of this system. We apply the Ehrenpreis-Palamodov fundamental principle, which works like a non-linear Fourier transform, to construct GP kernels mirroring standard spectral methods for GPs. Our approach can infer probable solutions of linear PDE systems from any data such as noisy measurements, or initial and boundary conditions. Constructing EPGP-priors is algorithmic, generally applicable, and comes with a sparse version (S-EPGP) that learns the relevant spectral frequencies and works better for big data sets. We demonstrate our approach on three families of systems of PDE, the heat equation, wave equation, and Maxwell's equations, where we improve upon the state of the art in computation time and precision, in some experiments by several orders of magnitude.
translated by 谷歌翻译
Linear partial differential equations (PDEs) are an important, widely applied class of mechanistic models, describing physical processes such as heat transfer, electromagnetism, and wave propagation. In practice, specialized numerical methods based on discretization are used to solve PDEs. They generally use an estimate of the unknown model parameters and, if available, physical measurements for initialization. Such solvers are often embedded into larger scientific models or analyses with a downstream application such that error quantification plays a key role. However, by entirely ignoring parameter and measurement uncertainty, classical PDE solvers may fail to produce consistent estimates of their inherent approximation error. In this work, we approach this problem in a principled fashion by interpreting solving linear PDEs as physics-informed Gaussian process (GP) regression. Our framework is based on a key generalization of a widely-applied theorem for conditioning GPs on a finite number of direct observations to observations made via an arbitrary bounded linear operator. Crucially, this probabilistic viewpoint allows to (1) quantify the inherent discretization error; (2) propagate uncertainty about the model parameters to the solution; and (3) condition on noisy measurements. Demonstrating the strength of this formulation, we prove that it strictly generalizes methods of weighted residuals, a central class of PDE solvers including collocation, finite volume, pseudospectral, and (generalized) Galerkin methods such as finite element and spectral methods. This class can thus be directly equipped with a structured error estimate and the capability to incorporate uncertain model parameters and observations. In summary, our results enable the seamless integration of mechanistic models as modular building blocks into probabilistic models.
translated by 谷歌翻译
许多应用程序中的数据遵循普通微分方程(ODE)的系统。本文提出了一种新型的算法和符号结构,用于高斯过程的协方差函数(GPS),其实现严格遵循具有恒定系数的线性均匀ODES系统,我们称之为lode-gps。将这种强的感应偏置引入GP,可以改善此类数据的建模。使用史密斯正常形式算法,一种符号技术,我们克服了技术状态中的两个当前限制:(1)在一组解决方案中需要某些唯一性条件的需求,通常在经典的ODE求解器及其概率求解器及其概率对应物中假定,以及(2)对可控系统的限制,通常在编码协方差函数中的微分方程时假设。我们显示了Lode-GP在许多实验中的有效性,例如通过最大化的可能性来学习物理解释的参数。
translated by 谷歌翻译
Interacting particle or agent systems that display a rich variety of swarming behaviours are ubiquitous in science and engineering. A fundamental and challenging goal is to understand the link between individual interaction rules and swarming. In this paper, we study the data-driven discovery of a second-order particle swarming model that describes the evolution of $N$ particles in $\mathbb{R}^d$ under radial interactions. We propose a learning approach that models the latent radial interaction function as Gaussian processes, which can simultaneously fulfill two inference goals: one is the nonparametric inference of {the} interaction function with pointwise uncertainty quantification, and the other one is the inference of unknown scalar parameters in the non-collective friction forces of the system. We formulate the learning problem as a statistical inverse problem and provide a detailed analysis of recoverability conditions, establishing that a coercivity condition is sufficient for recoverability. Given data collected from $M$ i.i.d trajectories with independent Gaussian observational noise, we provide a finite-sample analysis, showing that our posterior mean estimator converges in a Reproducing kernel Hilbert space norm, at an optimal rate in $M$ equal to the one in the classical 1-dimensional Kernel Ridge regression. As a byproduct, we show we can obtain a parametric learning rate in $M$ for the posterior marginal variance using $L^{\infty}$ norm, and the rate could also involve $N$ and $L$ (the number of observation time instances for each trajectory), depending on the condition number of the inverse problem. Numerical results on systems that exhibit different swarming behaviors demonstrate efficient learning of our approach from scarce noisy trajectory data.
translated by 谷歌翻译
本论文主要涉及解决深层(时间)高斯过程(DGP)回归问题的状态空间方法。更具体地,我们代表DGP作为分层组合的随机微分方程(SDES),并且我们通过使用状态空间过滤和平滑方法来解决DGP回归问题。由此产生的状态空间DGP(SS-DGP)模型生成丰富的电视等级,与建模许多不规则信号/功能兼容。此外,由于他们的马尔可道结构,通过使用贝叶斯滤波和平滑方法可以有效地解决SS-DGPS回归问题。本论文的第二次贡献是我们通过使用泰勒力矩膨胀(TME)方法来解决连续离散高斯滤波和平滑问题。这诱导了一类滤波器和SmooThers,其可以渐近地精确地预测随机微分方程(SDES)解决方案的平均值和协方差。此外,TME方法和TME过滤器和SmoOthers兼容模拟SS-DGP并解决其回归问题。最后,本文具有多种状态 - 空间(深)GPS的应用。这些应用主要包括(i)来自部分观察到的轨迹的SDES的未知漂移功能和信号的光谱 - 时间特征估计。
translated by 谷歌翻译
高斯工艺是能够以代表不确定性的方式学习未知功能的机器学习模型,从而促进了最佳决策系统的构建。由于渴望部署新颖的科学领域的高斯过程,一种迅速增长的研究线路集中于建设性地扩展这些模型来处理非欧几里德域,包括黎曼歧管,例如球形和托尔。我们提出了概括这一类的技术,以模拟黎曼歧管上的矢量字段,这在物理科学中的许多应用领域都很重要。为此,我们介绍了构建规范独立核的一般配方,它诱导高斯矢量字段,即矢量值高斯工艺与几何形状相干,从标量值riemannian内核。我们扩展了标准高斯过程培训方法,例如变分推理,以此设置。这使得旨在使用标准方法培训的Riemannian歧管上的矢量值高斯流程,并使它们可以访问机器学习从业者。
translated by 谷歌翻译
我们为特殊神经网络架构,称为运营商复发性神经网络的理论分析,用于近似非线性函数,其输入是线性运算符。这些功能通常在解决方案算法中出现用于逆边值问题的问题。传统的神经网络将输入数据视为向量,因此它们没有有效地捕获与对应于这种逆问题中的数据的线性运算符相关联的乘法结构。因此,我们介绍一个类似标准的神经网络架构的新系列,但是输入数据在向量上乘法作用。由较小的算子出现在边界控制中的紧凑型操作员和波动方程的反边值问题分析,我们在网络中的选择权重矩阵中促进结构和稀疏性。在描述此架构后,我们研究其表示属性以及其近似属性。我们还表明,可以引入明确的正则化,其可以从所述逆问题的数学分析导出,并导致概括属性上的某些保证。我们观察到重量矩阵的稀疏性改善了概括估计。最后,我们讨论如何将运营商复发网络视为深度学习模拟,以确定诸如用于从边界测量的声波方程中重建所未知的WAVESTED的边界控制的算法算法。
translated by 谷歌翻译
高斯过程可以说是空间统计中最重要的模型类别。他们编码有关建模功能的先前信息,可用于精确或近似贝叶斯推断。在许多应用中,尤其是在物理科学和工程中,以及在诸如地统计和神经科学等领域,对对称性的不变性是人们可以考虑的先前信息的最基本形式之一。高斯工艺与这种对称性的协方差的不变性导致了对此类空间平稳性概念的最自然概括。在这项工作中,我们开发了建设性和实用的技术,用于在在对称的背景下产生的一大批非欧基人空间上构建固定的高斯工艺。我们的技术使(i)以实用的方式计算(i)计算在此类空间上定义的先验和后高斯过程中的协方差内核和(ii)。这项工作分为两部分,每个部分涉及不同的技术考虑:第一部分研究紧凑的空间,而第二部分研究的非紧密空间具有某些结构。我们的贡献使我们研究的非欧亚人高斯流程模型与标准高斯流程软件包中可用的良好计算技术兼容,从而使从业者可以访问它们。
translated by 谷歌翻译
我们介绍了Hida-Mat'Ern内核的班级,这是整个固定式高斯 - 马尔可夫流程的整个空间的规范家庭协方差。它在垫子内核上延伸,通过允许灵活地构造具有振荡组件的过程。任何固定内核,包括广泛使用的平方指数和光谱混合核,要么直接在该类内,也是适当的渐近限制,展示了该类的一般性。利用其Markovian Nature,我们展示了如何仅使用内核及其衍生物来代表状态空间模型的过程。反过来,这使我们能够更有效地执行高斯工艺推论,并且侧面通常计算负担。我们还表明,除了进一步减少计算复杂性之外,我们还显示了如何利用状态空间表示的特殊属性。
translated by 谷歌翻译
本文提出了一个身体一致的高斯过程(GP),以识别不确定的拉格朗日系统。该功能空间是根据拉格朗日和微分方程结构的能量成分量身定制的,可以在分析上保证物理和数学特性,例如能量保护和二次形式。Cholesky分解矩阵内核的新型配方可允许概率保留正定性。在扭矩,速度和加速度中允许高斯噪声时,仅需要进行函数图的差分输入测量值。我们证明了该方法在数值模拟中的有效性。
translated by 谷歌翻译
Physical law learning is the ambiguous attempt at automating the derivation of governing equations with the use of machine learning techniques. The current literature focuses however solely on the development of methods to achieve this goal, and a theoretical foundation is at present missing. This paper shall thus serve as a first step to build a comprehensive theoretical framework for learning physical laws, aiming to provide reliability to according algorithms. One key problem consists in the fact that the governing equations might not be uniquely determined by the given data. We will study this problem in the common situation that a physical law is described by an ordinary or partial differential equation. For various different classes of differential equations, we provide both necessary and sufficient conditions for a function to uniquely determine the differential equation which is governing the phenomenon. We then use our results to devise numerical algorithms to determine whether a function solves a differential equation uniquely. Finally, we provide extensive numerical experiments showing that our algorithms in combination with common approaches for learning physical laws indeed allow to guarantee that a unique governing differential equation is learnt, without assuming any knowledge about the function, thereby ensuring reliability.
translated by 谷歌翻译
本文通过引入几何深度学习(GDL)框架来构建通用馈电型型模型与可区分的流形几何形状兼容的通用馈电型模型,从而解决了对非欧国人数据进行处理的需求。我们表明,我们的GDL模型可以在受控最大直径的紧凑型组上均匀地近似任何连续目标函数。我们在近似GDL模型的深度上获得了最大直径和上限的曲率依赖性下限。相反,我们发现任何两个非分类紧凑型歧管之间始终都有连续的函数,任何“局部定义”的GDL模型都不能均匀地近似。我们的最后一个主要结果确定了数据依赖性条件,确保实施我们近似的GDL模型破坏了“维度的诅咒”。我们发现,任何“现实世界”(即有限)数据集始终满足我们的状况,相反,如果目标函数平滑,则任何数据集都满足我们的要求。作为应用,我们确认了以下GDL模型的通用近似功能:Ganea等。 (2018)的双波利馈电网络,实施Krishnan等人的体系结构。 (2015年)的深卡尔曼 - 滤波器和深度玛克斯分类器。我们构建了:Meyer等人的SPD-Matrix回归剂的通用扩展/变体。 (2011)和Fletcher(2003)的Procrustean回归剂。在欧几里得的环境中,我们的结果暗示了Kidger和Lyons(2020)的近似定理和Yarotsky和Zhevnerchuk(2019)无估计近似率的数据依赖性版本的定量版本。
translated by 谷歌翻译
Many applications, such as system identification, classification of time series, direct and inverse problems in partial differential equations, and uncertainty quantification lead to the question of approximation of a non-linear operator between metric spaces $\mathfrak{X}$ and $\mathfrak{Y}$. We study the problem of determining the degree of approximation of such operators on a compact subset $K_\mathfrak{X}\subset \mathfrak{X}$ using a finite amount of information. If $\mathcal{F}: K_\mathfrak{X}\to K_\mathfrak{Y}$, a well established strategy to approximate $\mathcal{F}(F)$ for some $F\in K_\mathfrak{X}$ is to encode $F$ (respectively, $\mathcal{F}(F)$) in terms of a finite number $d$ (repectively $m$) of real numbers. Together with appropriate reconstruction algorithms (decoders), the problem reduces to the approximation of $m$ functions on a compact subset of a high dimensional Euclidean space $\mathbb{R}^d$, equivalently, the unit sphere $\mathbb{S}^d$ embedded in $\mathbb{R}^{d+1}$. The problem is challenging because $d$, $m$, as well as the complexity of the approximation on $\mathbb{S}^d$ are all large, and it is necessary to estimate the accuracy keeping track of the inter-dependence of all the approximations involved. In this paper, we establish constructive methods to do this efficiently; i.e., with the constants involved in the estimates on the approximation on $\mathbb{S}^d$ being $\mathcal{O}(d^{1/6})$. We study different smoothness classes for the operators, and also propose a method for approximation of $\mathcal{F}(F)$ using only information in a small neighborhood of $F$, resulting in an effective reduction in the number of parameters involved.
translated by 谷歌翻译
这项工作提出了一个新的程序,可以在高斯过程(GP)建模的背景下获得预测分布,并放松了一些感兴趣的范围之外的插值约束:预测分布的平均值不一定会在观察到的值时插入观察值的值。感兴趣的外部范围,但仅限于留在外面。这种称为放松的高斯工艺(REGP)插值的方法在感兴趣的范围内提供了更好的预测分布,尤其是在GP模型的平稳性假设不合适的情况下。它可以被视为一种面向目标的方法,并且在贝叶斯优化中变得特别有趣,例如,对于目标函数的最小化,低功能值的良好预测分布很重要。当将预期改进标准和REGP用于依次选择评估点时,从理论上保证了所得优化算法的收敛性(前提)。实验表明,在贝叶斯优化中使用REGP代替固定的GP模型是有益的。
translated by 谷歌翻译
The polynomial kernels are widely used in machine learning and they are one of the default choices to develop kernel-based classification and regression models. However, they are rarely used and considered in numerical analysis due to their lack of strict positive definiteness. In particular they do not enjoy the usual property of unisolvency for arbitrary point sets, which is one of the key properties used to build kernel-based interpolation methods. This paper is devoted to establish some initial results for the study of these kernels, and their related interpolation algorithms, in the context of approximation theory. We will first prove necessary and sufficient conditions on point sets which guarantee the existence and uniqueness of an interpolant. We will then study the Reproducing Kernel Hilbert Spaces (or native spaces) of these kernels and their norms, and provide inclusion relations between spaces corresponding to different kernel parameters. With these spaces at hand, it will be further possible to derive generic error estimates which apply to sufficiently smooth functions, thus escaping the native space. Finally, we will show how to employ an efficient stable algorithm to these kernels to obtain accurate interpolants, and we will test them in some numerical experiment. After this analysis several computational and theoretical aspects remain open, and we will outline possible further research directions in a concluding section. This work builds some bridges between kernel and polynomial interpolation, two topics to which the authors, to different extents, have been introduced under the supervision or through the work of Stefano De Marchi. For this reason, they wish to dedicate this work to him in the occasion of his 60th birthday.
translated by 谷歌翻译
内核方法是机器学习中最流行的技术之一,使用再现内核希尔伯特空间(RKHS)的属性来解决学习任务。在本文中,我们提出了一种新的数据分析框架,与再现内核Hilbert $ C ^ * $ - 模块(rkhm)和rkhm中的内核嵌入(kme)。由于RKHM包含比RKHS或VVRKHS)的更丰富的信息,因此使用RKHM的分析使我们能够捕获和提取诸如功能数据的结构属性。我们向RKHM展示了rkhm理论的分支,以适用于数据分析,包括代表性定理,以及所提出的KME的注射性和普遍性。我们还显示RKHM概括RKHS和VVRKHS。然后,我们提供采用RKHM和提议的KME对数据分析的具体程序。
translated by 谷歌翻译
内核Stein差异(KSD)是一种基于内核的广泛使用概率指标之间差异的非参数量度。它通常在用户从候选概率度量中收集的样本集合的情况下使用,并希望将它们与指定的目标概率度量进行比较。 KSD的一个有用属性是,它可以仅从候选度量的样本中计算出来,并且不知道目标度量的正常化常数。 KSD已用于一系列设置,包括合适的测试,参数推断,MCMC输出评估和生成建模。当前KSD方法论的两个主要问题是(i)超出有限维度欧几里得环境之外的适用性以及(ii)缺乏影响KSD性能的清晰度。本文提供了KSD的新频谱表示,这两种补救措施都使KSD适用于希尔伯特(Hilbert)评估数据,并揭示了内核和Stein oterator Choice对KSD的影响。我们通过在许多合成数据实验中对各种高斯和非高斯功能模型进行拟合优度测试来证明所提出的方法的功效。
translated by 谷歌翻译
我们使用仅使用独立和相同分布的样本的有限集合来估计用于估计动态系统的正向可迁移装置的算法。产生的估计是称为经验逆克里斯科特的函数的诸如函数的Sublevel组:已知经验逆Christoffel功能,以提供对概率分布的支持的良好近似。除了可达性分析之外,可以应用于估计随机变量支持的一般问题,这在数据科学中具有数据科学中的应用程序,可以应用于数据集中的Novelties和异常值。在安全是一个问题的应用中,保证在有限数据集上保持的准确性至关重要。在本文中,我们在可能大致正确(PAC)框架下证明了我们算法的界限。除了应用古典VAPnik-Chervonenkis(VC)维度绑定参数之外,我们除了利用核化经验逆克里斯科特函数和高斯进程回归模型之间的正式连接,我们还应用PAC-Bayes定理。基于Pac-Bayes的界限适用于比VC维度参数更一般的Christoffel功能,并在实验中实现了更大的样本效率。
translated by 谷歌翻译
本文研究了无限二维希尔伯特空间之间线性算子的学习。训练数据包括希尔伯特空间中的一对随机输入向量以及在未知的自我接合线性运算符下的嘈杂图像。假设操作员在已知的基础上是对角线化的,则该工作解决了给定数据估算操作员特征值的等效反问题。采用贝叶斯方法,理论分析在无限的数据限制中建立了后部收缩率,而高斯先验者与反向问题的正向图没有直接相关。主要结果还包括学习理论的概括错误保证了广泛的分配变化。这些收敛速率分别量化了数据平滑度和真实特征值衰减或生长的影响,分别是紧凑或无界操作员对样品复杂性的影响。数值证据支持对角线和非对角性环境中的理论。
translated by 谷歌翻译