Open Access
Multivariate Analysis from a Statistical Point of View
Kyle Cranmer
- pp 211
Reads0
Chats0
TLDR
In this paper, the Neyman-Pearson theory was translated into the language of statistical learning theory, and a formalism for a learning machine was introduced, which is general enough to encompass all of the techniques used within high energy physics.Abstract:
Multivariate Analysis is an increasingly common tool in experimental high energy physics; however, many of the common approaches were borrowed from other elds. We clarify what the goal of a multivariate algorithm should be for the search for a new particle and compare dieren t approaches. We also translate the Neyman-Pearson theory into the language of statistical learning theory. Multivariate Analysis is an increasingly common tool in experimental high energy physics; however, most of the common approaches were borrowed from other elds. Each of these algorithms were developed for their own particular task, thus they look quite different at their core. It is not obvious that what these dieren t algorithms do internally is optimal for the the tasks which they perform within high energy physics. It is also quite dicult to compare these dieren t algorithms due to the dierences in the formalisms that were used to derive and/or document them. In Section 2 we introduce a formalism for a Learning Machine, which is general enough to encompass all of the techniques used within high energy physics. In Sections 3 & 4 we review the statistical statements relevant to new particle searches and translate them into the formalism of statistical learning theory. In the remainder of the note, we look at the main results of statistical learning theory and their relevance to some of the common algorithms used within high energy physics.read more
Citations
More filters
Journal ArticleDOI
PhysicsGP: A Genetic Programming approach to event selection
Kyle Cranmer,R. Sean Bowman +1 more
TL;DR: A novel multivariate classification technique based on Genetic Programming that optimizes a set of human-readable classifiers with respect to some user-defined performance measure is presented and offers several advantages compared to Neural Networks and Support Vector Machines.
Journal ArticleDOI
Estimating a Signal In the Presence of an Unknown Background
Wolfgang Rolke,Alan D. Lopez +1 more
TL;DR: In this paper, the authors describe a method for fitting distributions to data which only requires knowledge of the parametric form of either the signal or the background but not both, and the unknown distribution is fit using a non-parametric kernel density estimator.
References
More filters
Book
The Nature of Statistical Learning Theory
TL;DR: Setting of the learning problem consistency of learning processes bounds on the rate of convergence ofLearning processes controlling the generalization ability of learning process constructing learning algorithms what is important in learning theory?
Book
Genetic Programming: On the Programming of Computers by Means of Natural Selection
TL;DR: This book discusses the evolution of architecture, primitive functions, terminals, sufficiency, and closure, and the role of representation and the lens effect in genetic programming.
Journal ArticleDOI
Multivariate Density Estimation, Theory, Practice and Visualization
TL;DR: Representation and Geometry of Multivariate Data.