scispace - formally typeset
Open AccessProceedings ArticleDOI

Self bounding learning algorithms

Yoav Freund
- pp 247-258
TLDR
A self-bounding learning algorithm is an algorithm which, in addition to the hypothesis that it outputs, outputs a reliable upper bound on the generalization error of this hypothesis.
Abstract
Most of the work which attempts to give bounds on the generalization error of the hypothesis generated by a learning algorithm is based on methods from the theory of uniform convergence. These bounds are a-priori bounds that hold for any distribution of examples and are calculated before any data is observed. In this paper we propose a different approach for bounding the generalization error after the data has been observed. A self-bounding learning algorithm is an algorithm which, in addition to the hypothesis that it outputs, outputs a reliable upper bound on the generalization error of this hypothesis. We first explore the idea in the statistical query learning framework of Keams [lo]. After that we give an explicit self bounding algorithm for learning algorithms that are based on local search. Permission to make digital or h,ard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for prolit or commercial advantage and that copies bear this notice and the full citation on the first page. To COPY otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific pemlission .and/or a fee. COLT 98 Madison WI USA Copyright ACM 1998 l-581 13-057--0/9X/ 7...$5.00

read more

Content maybe subject to copyright    Report

Citations
More filters
Book

Adaptive computation and machine learning

TL;DR: This book attempts to give an overview of the different recent efforts to deal with covariate shift, a challenging situation where the joint distribution of inputs and outputs differs between the training and test stages.
Journal ArticleDOI

On the generalization ability of on-line learning algorithms

TL;DR: This paper proves tight data-dependent bounds for the risk of this hypothesis in terms of an easily computable statistic M/sub n/ associated with the on-line performance of the ensemble, and obtains risk tail bounds for kernel perceptron algorithms interms of the spectrum of the empirical kernel matrix.
Journal ArticleDOI

Theory of classification : a survey of some recent advances

TL;DR: The last few years have witnessed important new developments in the theory and practice of pattern classification, see as discussed by the authors for a survey of the main new ideas that have lead to these important recent developments.
Journal ArticleDOI

Prediction games and arcing algorithms

TL;DR: The theory behind the success of adaptive reweighting and combining algorithms (arcing) such as Adaboost and others in reducing generalization error has not been well understood, and an explanation of whyAdaboost works in terms of its ability to produce generally high margins is offered.
References
More filters
Book

C4.5: Programs for Machine Learning

TL;DR: A complete guide to the C4.5 system as implemented in C for the UNIX environment, which starts from simple core learning methods and shows how they can be elaborated and extended to deal with typical problems such as missing data and over hitting.
Journal ArticleDOI

Bagging predictors

Leo Breiman
TL;DR: Tests on real and simulated data sets using classification and regression trees and subset selection in linear regression show that bagging can give substantial gains in accuracy.
Book

Classification and regression trees

Leo Breiman
TL;DR: The methodology used to construct tree structured rules is the focus of a monograph as mentioned in this paper, covering the use of trees as a data analysis method, and in a more mathematical framework, proving some of their fundamental properties.