scispace - formally typeset
Search or ask a question
Topic

Statistical learning theory

About: Statistical learning theory is a research topic. Over the lifetime, 1618 publications have been published within this topic receiving 158033 citations.


Papers
More filters
Posted Content
TL;DR: It is shown how a probabilistic program can be automatically represented in a theorem prover using the concept ofreparameterization, and how some of the tedious proofs of measurability can be generated automatically from the probabilism program.
Abstract: As machine learning is increasingly used in essential systems, it is important to reduce or eliminate the incidence of serious bugs. A growing body of research has developed machine learning algorithms with formal guarantees about performance, robustness, or fairness. Yet, the analysis of these algorithms is often complex, and implementing such systems in practice introduces room for error. Proof assistants can be used to formally verify machine learning systems by constructing machine checked proofs of correctness that rule out such bugs. However, reasoning about probabilistic claims inside of a proof assistant remains challenging. We show how a probabilistic program can be automatically represented in a theorem prover using the concept of \emph{reparameterization}, and how some of the tedious proofs of measurability can be generated automatically from the probabilistic program. To demonstrate that this approach is broad enough to handle rather different types of machine learning systems, we verify both a classic result from statistical learning theory (PAC-learnability of decision stumps) and prove that the null model used in a Bayesian hypothesis test satisfies a fairness criterion called demographic parity.

2 citations

Posted Content
TL;DR: In this article, the empirical duality gap is defined as the difference between an approximate, tractable solution and the solution of the original (nonconvex)~statistical problem.
Abstract: Though learning has become a core technology of modern information processing, there is now ample evidence that it can lead to biased, unsafe, and prejudiced solutions. The need to impose requirements on learning is therefore paramount, especially as it reaches critical applications in social, industrial, and medical domains. However, the non-convexity of most modern learning problems is only exacerbated by the introduction of constraints. Whereas good unconstrained solutions can often be learned using empirical risk minimization (ERM), even obtaining a model that satisfies statistical constraints can be challenging, all the more so a good one. In this paper, we overcome this issue by learning in the empirical dual domain, where constrained statistical learning problems become unconstrained, finite dimensional, and deterministic. We analyze the generalization properties of this approach by bounding the empirical duality gap, i.e., the difference between our approximate, tractable solution and the solution of the original (non-convex)~statistical problem, and provide a practical constrained learning algorithm. These results establish a constrained counterpart of classical learning theory and enable the explicit use of constraints in learning. We illustrate this algorithm and theory in rate-constrained learning applications.

2 citations

Book ChapterDOI
01 Jan 2012
TL;DR: A new method called support vector estimating and selecting (SVES) is presented, which improves SVM for large sample training speed by remove the sample, which help smeller, redundancy or obvious noise.
Abstract: As a kind of statistical learning theory, in solving the small data set, nonlinear and high dimensional problems, support vector machine (SVM) has shown many advantages. It has been widely applied in recent years. However, with the increase of the training samples, normal SVM training speed becomes the bottleneck of restricting its application. Therefore, this paper presents a new method called support vector estimating and selecting (SVES). It improves SVM for large sample training speed by remove the sample, which help smeller, redundancy or obvious noise.

2 citations

Proceedings ArticleDOI
13 Jun 2005
TL;DR: A kind of classification model for document classification based on statistical learning theory is presented, which adopts organized vectors as the eigenvector of documents, trains classifier by means of SVM algorithm, and gets satisfactory experiment results.
Abstract: Document classification is one of important steps in document mining. With the statistical learning theory, they proposed a kind of machine learning method based on small sample set. This paper presents a kind of classification model for document classification based on statistical learning theory. In this model, we adopt organized vectors as the eigenvector of documents, trains classifier by means of SVM algorithm, and obtain satisfactory experiment results.

2 citations

Book ChapterDOI
12 Apr 2015
TL;DR: A framework to analyze the sample complexity of problems that arise in the study of genomic datasets based on tools from combinatorial analysis and statistical learning theory, showing that sample sizes much larger than currently available may be required to identify all the cancer genes in a pathway.
Abstract: In this work we propose a framework to analyze the sample complexity of problems that arise in the study of genomic datasets. Our framework is based on tools from combinatorial analysis and statistical learning theory that have been used for the analysis of machine learning and probably approximately correct (PAC) learning. We use our framework to analyze the problem of the identification of cancer pathways through mutual exclusivity analysis of mutations from large cancer sequencing studies. We analytically derive matching upper and lower bounds on the sample complexity of the problem, showing that sample sizes much larger than currently available may be required to identify all the cancer genes in a pathway. We also provide two algorithms to find a cancer pathway from a large genomic dataset. On simulated and cancer data, we show that our algorithms can be used to identify cancer pathways from large genomic datasets.

2 citations


Network Information
Related Topics (5)
Artificial neural network
207K papers, 4.5M citations
86% related
Cluster analysis
146.5K papers, 2.9M citations
82% related
Feature extraction
111.8K papers, 2.1M citations
81% related
Optimization problem
96.4K papers, 2.1M citations
80% related
Fuzzy logic
151.2K papers, 2.3M citations
79% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
20239
202219
202159
202069
201972
201847