scispace - formally typeset
Open AccessPosted Content

Graph Neural Networks for Particle Tracking and Reconstruction

Reads0
Chats0
TLDR
This chapter recapitulate the mathematical formalism of GNNs and highlight aspects to consider when designing these networks for HEP data, including graph construction, model architectures, learning objectives, and graph pooling.
Abstract
Machine learning methods have a long history of applications in high energy physics (HEP). Recently, there is a growing interest in exploiting these methods to reconstruct particle signatures from raw detector data. In order to benefit from modern deep learning algorithms that were initially designed for computer vision or natural language processing tasks, it is common practice to transform HEP data into images or sequences. Conversely, graph neural networks (GNNs), which operate on graph data composed of elements with a set of features and their pairwise connections, provide an alternative way of incorporating weight sharing, local connectivity, and specialized domain knowledge. Particle physics data, such as the hits in a tracking detector, can generally be represented as graphs, making the use of GNNs natural. In this chapter, we recapitulate the mathematical formalism of GNNs and highlight aspects to consider when designing these networks for HEP data, including graph construction, model architectures, learning objectives, and graph pooling. We also review promising applications of GNNs for particle tracking and reconstruction in HEP and summarize the outlook for their deployment in current and future experiments.

read more

Citations
More filters
Journal ArticleDOI

MLPF: efficient machine-learned particle-flow reconstruction using graph neural networks

TL;DR: In this paper, an end-to-end trainable, machine-learned particle-flow algorithm based on parallelizable, computationally efficient, and scalable graph neural network optimized using a multi-task objective on simulated events is presented.
Journal ArticleDOI

Recent advances in utility of artificial intelligence towards multiscale colloidal based materials design

TL;DR: A review of supervised and unsupervised strategies for colloidal material design can be found in this paper , where a collection of computer approaches ranging from quantum chemistry to molecular dynamics and continuum modeling are discussed.
Journal ArticleDOI

Graph Neural Networks in Particle Physics: Implementations, Innovations, and Challenges

TL;DR: A range of graph neural networks capabilities that allow a wide variety of high- and low-level physical features to be attached to measurements and, by the same token, to best match unique GNN capabilities to unique HEP obstacles are presented.
Posted Content

Applications and Techniques for Fast Machine Learning in Science

TL;DR: In this article, the authors discuss applications and techniques for fast machine learning (ML) in science, the concept of integrating power ML methods into the real-time experimental data processing loop to accelerate scientific discovery.
References
More filters
Proceedings Article

Attention is All you Need

TL;DR: This paper proposed a simple network architecture based solely on an attention mechanism, dispensing with recurrence and convolutions entirely and achieved state-of-the-art performance on English-to-French translation.
Posted Content

MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications

TL;DR: This work introduces two simple global hyper-parameters that efficiently trade off between latency and accuracy and demonstrates the effectiveness of MobileNets across a wide range of applications and use cases including object detection, finegrain classification, face attributes and large scale geo-localization.
Proceedings ArticleDOI

PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation

TL;DR: This paper designs a novel type of neural network that directly consumes point clouds, which well respects the permutation invariance of points in the input and provides a unified architecture for applications ranging from object classification, part segmentation, to scene semantic parsing.
Proceedings ArticleDOI

FaceNet: A unified embedding for face recognition and clustering

TL;DR: A system that directly learns a mapping from face images to a compact Euclidean space where distances directly correspond to a measure offace similarity, and achieves state-of-the-art face recognition performance using only 128-bytes perface.
Proceedings ArticleDOI

Non-local Neural Networks

TL;DR: In this article, the non-local operation computes the response at a position as a weighted sum of the features at all positions, which can be used to capture long-range dependencies.
Related Papers (5)