scispace - formally typeset
Open Access

Multivariate Analysis from a Statistical Point of View

Kyle Cranmer
- pp 211
Reads0
Chats0
TLDR
In this paper, the Neyman-Pearson theory was translated into the language of statistical learning theory, and a formalism for a learning machine was introduced, which is general enough to encompass all of the techniques used within high energy physics.
Abstract
Multivariate Analysis is an increasingly common tool in experimental high energy physics; however, many of the common approaches were borrowed from other elds. We clarify what the goal of a multivariate algorithm should be for the search for a new particle and compare dieren t approaches. We also translate the Neyman-Pearson theory into the language of statistical learning theory. Multivariate Analysis is an increasingly common tool in experimental high energy physics; however, most of the common approaches were borrowed from other elds. Each of these algorithms were developed for their own particular task, thus they look quite different at their core. It is not obvious that what these dieren t algorithms do internally is optimal for the the tasks which they perform within high energy physics. It is also quite dicult to compare these dieren t algorithms due to the dierences in the formalisms that were used to derive and/or document them. In Section 2 we introduce a formalism for a Learning Machine, which is general enough to encompass all of the techniques used within high energy physics. In Sections 3 & 4 we review the statistical statements relevant to new particle searches and translate them into the formalism of statistical learning theory. In the remainder of the note, we look at the main results of statistical learning theory and their relevance to some of the common algorithms used within high energy physics.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

PhysicsGP: A Genetic Programming approach to event selection

TL;DR: A novel multivariate classification technique based on Genetic Programming that optimizes a set of human-readable classifiers with respect to some user-defined performance measure is presented and offers several advantages compared to Neural Networks and Support Vector Machines.
Journal ArticleDOI

Estimating a Signal In the Presence of an Unknown Background

TL;DR: In this paper, the authors describe a method for fitting distributions to data which only requires knowledge of the parametric form of either the signal or the background but not both, and the unknown distribution is fit using a non-parametric kernel density estimator.
References
More filters
Book

The Nature of Statistical Learning Theory

TL;DR: Setting of the learning problem consistency of learning processes bounds on the rate of convergence ofLearning processes controlling the generalization ability of learning process constructing learning algorithms what is important in learning theory?
Book

Genetic Programming: On the Programming of Computers by Means of Natural Selection

TL;DR: This book discusses the evolution of architecture, primitive functions, terminals, sufficiency, and closure, and the role of representation and the lens effect in genetic programming.
Journal ArticleDOI

Multivariate Density Estimation, Theory, Practice and Visualization

R. H. Glendinning
- 01 Mar 1994 - 
TL;DR: Representation and Geometry of Multivariate Data.
Related Papers (5)