我们介绍了一种从电磁(EM)采样量热计收集的数据重建多个淋浴的第一算法。这种探测器广泛用于高能量物理中,以测量进入粒子的能量和运动学。在这项工作中,我们考虑许多电子通过乳液云室(ECC)砖的情况,启动电子诱导的电磁淋浴,这可以是长曝光时间或大输入粒子通量的情况。例如,船舶实验计划使用乳液检测器进行暗物质搜索和中微子物理调查。船舶实验的预期完整通量约为10 ^ 20颗粒。为了降低与替换ECC砖和离线数据的实验的成本(乳液扫描),决定增加暴露时间。因此,我们希望观察大量重叠阵雨,将EM淋浴重建变为挑战的点云分割问题。我们的重建管线包括图形神经网络,其预测邻接矩阵和聚类算法。我们提出了一种新的层型(乳液CONV),其考虑了ECC砖中淋浴开发的几何特性。对于重叠阵雨的聚类,我们使用修改后的基于分层密度的聚类算法。我们的方法不使用有关进入粒子的任何先前信息,并识别乳液检测器中的高达87%的电磁淋浴。用于重建电磁淋浴的算法的主要测试台将是SND @ LHC。
translated by 谷歌翻译
在这项工作中,我们提出了一种神经方法,用于重建描述层次相互作用的生根树图,使用新颖的表示,我们将其称为最低的共同祖先世代(LCAG)矩阵。这种紧凑的配方等效于邻接矩阵,但是如果直接使用邻接矩阵,则可以单独从叶子中学习树的结构,而无需先前的假设。因此,采用LCAG启用了第一个端到端的可训练解决方案,该解决方案仅使用末端树叶直接学习不同树大小的层次结构。在高能量粒子物理学的情况下,粒子衰减形成了分层树结构,只能通过实验观察到最终产物,并且可能的树的大型组合空间使分析溶液变得很棘手。我们证明了LCAG用作使用变压器编码器和神经关系编码器编码器图神经网络的模拟粒子物理衰减结构的任务。采用这种方法,我们能够正确预测LCAG纯粹是从叶子特征中的LCAG,最大树深度为$ 8 $ in $ 92.5 \%\%的树木箱子,最高$ 6 $叶子(包括)和$ 59.7 \%\%\%\%的树木$在我们的模拟数据集中$ 10 $。
translated by 谷歌翻译
最近的工作已经证明了图形神经网络(GNN)等几何深度学习方法非常适合于在高能粒子物理学中解决各种重建问题。特别地,粒子跟踪数据通过识别硅跟踪器命中作为节点和粒子轨迹作为边缘来自然表示为曲线图;给定一组假设的边缘,边缘分类GNN标识与真实粒子轨迹相对应的那些。在这项工作中,我们将物理激励的相互作用网络(IN)GNN调整为与高亮度大强子撞机的预期相似的填充条件中的粒子跟踪问题。假设在各种粒子矩阈值下进行理想化的击中过滤,我们通过在基于GNN的跟踪的每个阶段进行了一系列测量来展示了优异的边缘分类精度和跟踪效率:图形结构,边缘分类和轨道建筑。建议的建筑基本上比以前研究的GNN跟踪架构小幅小;这尤其希望,因为大小的减小对于在受约束的计算环境中实现基于GNN的跟踪至关重要。此外,可以将其表示为一组显式矩阵操作或传递GNN的消息。正在进行努力,以通过异构计算资源朝向高级和低延迟触发应用程序加速每个表示。
translated by 谷歌翻译
ICECUBE是一种用于检测1 GEV和1 PEV之间大气和天体中微子的光学传感器的立方公斤阵列,该阵列已部署1.45 km至2.45 km的南极的冰盖表面以下1.45 km至2.45 km。来自ICE探测器的事件的分类和重建在ICeCube数据分析中起着核心作用。重建和分类事件是一个挑战,这是由于探测器的几何形状,不均匀的散射和冰中光的吸收,并且低于100 GEV的光,每个事件产生的信号光子数量相对较少。为了应对这一挑战,可以将ICECUBE事件表示为点云图形,并将图形神经网络(GNN)作为分类和重建方法。 GNN能够将中微子事件与宇宙射线背景区分开,对不同的中微子事件类型进行分类,并重建沉积的能量,方向和相互作用顶点。基于仿真,我们提供了1-100 GEV能量范围的比较与当前ICECUBE分析中使用的当前最新最大似然技术,包括已知系统不确定性的影响。对于中微子事件分类,与当前的IceCube方法相比,GNN以固定的假阳性速率(FPR)提高了信号效率的18%。另外,GNN在固定信号效率下将FPR的降低超过8(低于半百分比)。对于能源,方向和相互作用顶点的重建,与当前最大似然技术相比,分辨率平均提高了13%-20%。当在GPU上运行时,GNN能够以几乎是2.7 kHz的中位数ICECUBE触发速率的速率处理ICECUBE事件,这打开了在在线搜索瞬态事件中使用低能量中微子的可能性。
translated by 谷歌翻译
Graph classification is an important area in both modern research and industry. Multiple applications, especially in chemistry and novel drug discovery, encourage rapid development of machine learning models in this area. To keep up with the pace of new research, proper experimental design, fair evaluation, and independent benchmarks are essential. Design of strong baselines is an indispensable element of such works. In this thesis, we explore multiple approaches to graph classification. We focus on Graph Neural Networks (GNNs), which emerged as a de facto standard deep learning technique for graph representation learning. Classical approaches, such as graph descriptors and molecular fingerprints, are also addressed. We design fair evaluation experimental protocol and choose proper datasets collection. This allows us to perform numerous experiments and rigorously analyze modern approaches. We arrive to many conclusions, which shed new light on performance and quality of novel algorithms. We investigate application of Jumping Knowledge GNN architecture to graph classification, which proves to be an efficient tool for improving base graph neural network architectures. Multiple improvements to baseline models are also proposed and experimentally verified, which constitutes an important contribution to the field of fair model comparison.
translated by 谷歌翻译
在2015年和2019年之间,地平线的成员2020年资助的创新培训网络名为“Amva4newphysics”,研究了高能量物理问题的先进多变量分析方法和统计学习工具的定制和应用,并开发了完全新的。其中许多方法已成功地用于提高Cern大型Hadron撞机的地图集和CMS实验所执行的数据分析的敏感性;其他几个人,仍然在测试阶段,承诺进一步提高基本物理参数测量的精确度以及新现象的搜索范围。在本文中,在研究和开发的那些中,最相关的新工具以及对其性能的评估。
translated by 谷歌翻译
Recently methods of graph neural networks (GNNs) have been applied to solving the problems in high energy physics (HEP) and have shown its great potential for quark-gluon tagging with graph representation of jet events. In this paper, we introduce an approach of GNNs combined with a HaarPooling operation to analyze the events, called HaarPooling Message Passing neural network (HMPNet). In HMPNet, HaarPooling not only extract the features of graph, but also embed additional information obtained by clustering of k-means of different particle observables. We construct Haarpooling from three different observables: absolute energy $\log E$, transverse momentum $\log p_T$ , and relative coordinates $(\Delta\eta,\Delta\phi)$, then discuss their impacts on the tagging and compare the results with those obtained via MPNN and ParticleNet (PN). The results show that an appropriate selection of information for HaarPooling enhance the accuracy of quark-gluon tagging, for adding extra information of $\log P_T$ to the HMPNet outperforms all the others, meanwhile adding relative coordinates information $(\Delta\eta,\Delta\phi)$ is not very beneficial.
translated by 谷歌翻译
即使机器学习算法已经在数据科学中发挥了重要作用,但许多当前方法对输入数据提出了不现实的假设。由于不兼容的数据格式,或数据集中的异质,分层或完全缺少的数据片段,因此很难应用此类方法。作为解决方案,我们提出了一个用于样本表示,模型定义和培训的多功能,统一的框架,称为“ Hmill”。我们深入审查框架构建和扩展的机器学习的多个范围范式。从理论上讲,为HMILL的关键组件的设计合理,我们将通用近似定理的扩展显示到框架中实现的模型所实现的所有功能的集合。本文还包含有关我们实施中技术和绩效改进的详细讨论,该讨论将在MIT许可下发布供下载。该框架的主要资产是其灵活性,它可以通过相同的工具对不同的现实世界数据源进行建模。除了单独观察到每个对象的一组属性的标准设置外,我们解释了如何在框架中实现表示整个对象系统的图表中的消息推断。为了支持我们的主张,我们使用框架解决了网络安全域的三个不同问题。第一种用例涉及来自原始网络观察结果的IoT设备识别。在第二个问题中,我们研究了如何使用以有向图表示的操作系统的快照可以对恶意二进制文件进行分类。最后提供的示例是通过网络中实体之间建模域黑名单扩展的任务。在所有三个问题中,基于建议的框架的解决方案可实现与专业方法相当的性能。
translated by 谷歌翻译
我们描述了作为黑暗机器倡议和LES Houches 2019年物理学研讨会进行的数据挑战的结果。挑战的目标是使用无监督机器学习算法检测LHC新物理学的信号。首先,我们提出了如何实现异常分数以在LHC搜索中定义独立于模型的信号区域。我们定义并描述了一个大型基准数据集,由> 10亿美元的Muton-Proton碰撞,其中包含> 10亿美元的模拟LHC事件组成。然后,我们在数据挑战的背景下审查了各种异常检测和密度估计算法,我们在一组现实分析环境中测量了它们的性能。我们绘制了一些有用的结论,可以帮助开发无监督的新物理搜索在LHC的第三次运行期间,并为我们的基准数据集提供用于HTTPS://www.phenomldata.org的未来研究。重现分析的代码在https://github.com/bostdiek/darkmachines-unsupervisedChallenge提供。
translated by 谷歌翻译
跟踪是大型强子撞机(LHC)的事件重建最耗时的方面之一及其高亮度升级(HL-LHC)。通过在模式识别和参数估计中包括时序,创新的探测器技术将跟踪到四维。然而,现在和未来的硬件已经具有通过现有的轨道播种算法主要未使用的附加信息。簇的形状为轨道播种提供了额外的尺寸,这可以显着降低轨道发现的组合挑战。我们使用神经网络来表明群集形状可以显着降低假组合背景的速度,同时保持高效率。我们使用集群单曲,双峰和三胞胎中的信息来展示这一点。来自TrackML挑战的仿真呈现了数值结果。
translated by 谷歌翻译
喷气标记是粒子物理学中的一项关键但具有挑战性的分类任务。尽管深度学习已经改变了喷气标记并显着提高了性能,但缺乏大规模的公共数据集阻碍了进一步的增强。在这项工作中,我们提出了JetClass,这是一种用于喷气标记的新综合数据集。 JETCLASS数据集由100 M喷气机组成,比现有公共数据集大约两个数量级。总共模拟了10种类型的喷气机,包括到目前为止未探索用于标记的几种类型。基于大型数据集,我们提出了一种用于喷射标记的新的基于变压器的体系结构,称为“粒子变压器”(部分)。通过将成对的粒子相互作用纳入注意机制,部分可以达到比普通变压器更高的标记性能,并超过了先前最新的颗粒,颗粒的幅度很大。一旦进行了微调,预先训练的零件模型也大大提高了两个广泛采用的喷气标记基准的性能。数据集,代码和模型可在https://github.com/jet-universe/particle_transformer上公开获得。
translated by 谷歌翻译
The goal of graph summarization is to represent large graphs in a structured and compact way. A graph summary based on equivalence classes preserves pre-defined features of a graph's vertex within a $k$-hop neighborhood such as the vertex labels and edge labels. Based on these neighborhood characteristics, the vertex is assigned to an equivalence class. The calculation of the assigned equivalence class must be a permutation invariant operation on the pre-defined features. This is achieved by sorting on the feature values, e. g., the edge labels, which is computationally expensive, and subsequently hashing the result. Graph Neural Networks (GNN) fulfill the permutation invariance requirement. We formulate the problem of graph summarization as a subgraph classification task on the root vertex of the $k$-hop neighborhood. We adapt different GNN architectures, both based on the popular message-passing protocol and alternative approaches, to perform the structural graph summarization task. We compare different GNNs with a standard multi-layer perceptron (MLP) and Bloom filter as non-neural method. For our experiments, we consider four popular graph summary models on a large web graph. This resembles challenging multi-class vertex classification tasks with the numbers of classes ranging from $576$ to multiple hundreds of thousands. Our results show that the performance of GNNs are close to each other. In three out of four experiments, the non-message-passing GraphMLP model outperforms the other GNNs. The performance of the standard MLP is extraordinary good, especially in the presence of many classes. Finally, the Bloom filter outperforms all neural architectures by a large margin, except for the dataset with the fewest number of $576$ classes.
translated by 谷歌翻译
我们提出了一种新型的图形神经网络(GNN)方法,用于高通量显微镜视频中的细胞跟踪。通过将整个延时序列建模为直接图,其中细胞实例由其节点及其边缘表示,我们通过查找图中的最大路径来提取整个细胞轨迹。这是由纳入端到端深度学习框架中的几个关键贡献来完成的。我们利用深度度量学习算法来提取细胞特征向量,以区分不同生物细胞的实例并组装相同的细胞实例。我们引入了一种新的GNN块类型,该类型可以对节点和边缘特征向量进行相互更新,从而促进基础消息传递过程。消息传递概念的范围由GNN块的数量确定,这是至关重要的,因为它可以在连续的框架中实现节点和边缘之间的“节点和边缘”之间的“流动”。最后,我们解决了边缘分类问题,并使用已确定的活动边缘来构建单元格的轨道和谱系树。我们通过将其应用于不同细胞类型,成像设置和实验条件的2D和3D数据集,来证明所提出的细胞跟踪方法的强度。我们表明,我们的框架在大多数评估的数据集上都优于当前最新方法。该代码可在我们的存储库中获得:https://github.com/talbenha/cell-tracker-gnn。
translated by 谷歌翻译
在过去十年中,图形内核引起了很多关注,并在结构化数据上发展成为一种快速发展的学习分支。在过去的20年中,该领域发生的相当大的研究活动导致开发数十个图形内核,每个图形内核都对焦于图形的特定结构性质。图形内核已成功地成功地在广泛的域中,从社交网络到生物信息学。本调查的目标是提供图形内核的文献的统一视图。特别是,我们概述了各种图形内核。此外,我们对公共数据集的几个内核进行了实验评估,并提供了比较研究。最后,我们讨论图形内核的关键应用,并概述了一些仍有待解决的挑战。
translated by 谷歌翻译
Clustering is a fundamental problem in network analysis that finds closely connected groups of nodes and separates them from other nodes in the graph, while link prediction is to predict whether two nodes in a network are likely to have a link. The definition of both naturally determines that clustering must play a positive role in obtaining accurate link prediction tasks. Yet researchers have long ignored or used inappropriate ways to undermine this positive relationship. In this article, We construct a simple but efficient clustering-driven link prediction framework(ClusterLP), with the goal of directly exploiting the cluster structures to obtain connections between nodes as accurately as possible in both undirected graphs and directed graphs. Specifically, we propose that it is easier to establish links between nodes with similar representation vectors and cluster tendencies in undirected graphs, while nodes in a directed graphs can more easily point to nodes similar to their representation vectors and have greater influence in their own cluster. We customized the implementation of ClusterLP for undirected and directed graphs, respectively, and the experimental results using multiple real-world networks on the link prediction task showed that our models is highly competitive with existing baseline models. The code implementation of ClusterLP and baselines we use are available at https://github.com/ZINUX1998/ClusterLP.
translated by 谷歌翻译
X-ray imaging technology has been used for decades in clinical tasks to reveal the internal condition of different organs, and in recent years, it has become more common in other areas such as industry, security, and geography. The recent development of computer vision and machine learning techniques has also made it easier to automatically process X-ray images and several machine learning-based object (anomaly) detection, classification, and segmentation methods have been recently employed in X-ray image analysis. Due to the high potential of deep learning in related image processing applications, it has been used in most of the studies. This survey reviews the recent research on using computer vision and machine learning for X-ray analysis in industrial production and security applications and covers the applications, techniques, evaluation metrics, datasets, and performance comparison of those techniques on publicly available datasets. We also highlight some drawbacks in the published research and give recommendations for future research in computer vision-based X-ray analysis.
translated by 谷歌翻译
通过使用机器学习技术的异常检测已成为一种新型强大的工具,可以在标准模型之外寻找新物理学。从历史上看,与JET可观察物的发展相似,理论一致性并不总是在算法和神经网络体系结构的快速发展中扮演核心角色。在这项工作中,我们通过使用能量加权消息传递来构建基于图神经网络的红外和共线安全自动编码器。我们证明,尽管这种方法具有理论上有利的特性,但它也对非QCD结构表现出强大的敏感性。
translated by 谷歌翻译
地震阶段关联将地震到达时间测量连接到其致病来源。有效的关联必须确定离散事件的数量,其位置和起源时间,并且必须将实际到达与测量工件区分开。深度学习采摘者的出现,从紧密重叠的小地震中提供了高率的速度,它激发了重新审视相关问题并使用深度学习方法来解决它。我们已经开发了一个图形神经网络关联器,该协会同时预测源时空定位和离散的源源 - 边界关联可能性。该方法适用于任意几何形状,数百个电台的时变地震网络,并且具有可变噪声和质量的高源和输入选拔速率。我们的图形地震神经解释引擎(Genie)使用一个图来表示站点,另一个图表示空间源区域。 Genie从数据中从数据中学习了关系,使其能够确定可靠的源和源源联想。我们使用Phasenet Deep Learth Learning Phase Phase Picker生成的输入来培训合成数据,并测试来自北加州(NC)地震网络的真实数据的方法。我们成功地重新检测了USGS在2000年$ \ unicode {x2013} $ 2022之间的500天报告中报告的所有事件M> 1的96%。在2017年的100天连续处理间隔中,$ \ unicode {x2013} $ 2018,我们检测到〜4.2x USGS报告的事件数量。我们的新事件的估计值低于USGS目录的完整性幅度,并且位于该地区的活动故障和采石场附近。我们的结果表明,精灵可以在复杂的地震监测条件下有效解决关联问题。
translated by 谷歌翻译
Deep learning has revolutionized many machine learning tasks in recent years, ranging from image classification and video processing to speech recognition and natural language understanding. The data in these tasks are typically represented in the Euclidean space. However, there is an increasing number of applications where data are generated from non-Euclidean domains and are represented as graphs with complex relationships and interdependency between objects. The complexity of graph data has imposed significant challenges on existing machine learning algorithms. Recently, many studies on extending deep learning approaches for graph data have emerged. In this survey, we provide a comprehensive overview of graph neural networks (GNNs) in data mining and machine learning fields. We propose a new taxonomy to divide the state-of-the-art graph neural networks into four categories, namely recurrent graph neural networks, convolutional graph neural networks, graph autoencoders, and spatial-temporal graph neural networks. We further discuss the applications of graph neural networks across various domains and summarize the open source codes, benchmark data sets, and model evaluation of graph neural networks. Finally, we propose potential research directions in this rapidly growing field.
translated by 谷歌翻译
In this article, we use artificial intelligence algorithms to show how to enhance the resolution of the elementary particle track fitting in inhomogeneous dense detectors, such as plastic scintillators. We use deep learning to replace more traditional Bayesian filtering methods, drastically improving the reconstruction of the interacting particle kinematics. We show that a specific form of neural network, inherited from the field of natural language processing, is very close to the concept of a Bayesian filter that adopts a hyper-informative prior. Such a paradigm change can influence the design of future particle physics experiments and their data exploitation.
translated by 谷歌翻译