scispace - formally typeset
Open AccessBook ChapterDOI

Using diversity in preparing ensembles of classifiers based on different feature subsets to minimize generalization error

Reads0
Chats0
TLDR
This paper presents a process for producing ensembles of classifiers based on different feature subsets that emphasizes diversity (ambiguity) in the ensemble members and finds that the ensemble based on ambiguity have lower generalization error.
Abstract
It is well known that ensembles of predictors produce better accuracy than a single predictor provided there is diversity in the ensemble. This diversity manifests itself as disagreement or ambiguity among the ensemble members. In this paper we focus on ensembles of classifiers based on different feature subsets and we present a process for producing such ensembles that emphasizes diversity (ambiguity) in the ensemble members. This emphasis on diversity produces ensembles with low generalization errors from ensemble members with comparatively high generalization error. We compare this with ensembles produced focusing only on the error of the ensemble members (without regard to overall diversity) and find that the ensembles based on ambiguity have lower generalization error. Further, we find that the ensemble members produced focusing on ambiguity have less features on average that those based on error only. We suggest that this indicates that these ensemble members are local learners.

read more

Citations
More filters
Journal ArticleDOI

Diversity creation methods: a survey and categorisation

TL;DR: This paper reviews the varied attempts to provide a formal explanation of error diversity, including several heuristic and qualitative explanations in the literature, and introduces the idea of implicit and explicit diversity creation methods, and three dimensions along which these may be applied.
Journal ArticleDOI

A survey of multiple classifier systems as hybrid systems

TL;DR: An up-to-date survey on multiple classifier system (MCS) from the point of view of Hybrid Intelligent Systems is presented, providing a vision of the spectrum of applications that are currently being developed.
Journal ArticleDOI

Ensemble learning for data stream analysis

TL;DR: This paper surveys research on ensembles for data stream classification as well as regression tasks and discusses advanced learning concepts such as imbalanced data streams, novelty detection, active and semi-supervised learning, complex data representations and structured outputs.
Journal ArticleDOI

Classifier selection for majority voting

TL;DR: This work provides a revision of the classifier selection methodology and evaluates the practical applicability of diversity measures in the context of combining classifiers by majority voting, and proposes a novel design of multiple classifier systems in which selection and fusion are recurrently applied to a population of best combinations of classifiers.
Journal ArticleDOI

Ensemble approaches for regression: A survey

TL;DR: Different approaches to each of these phases that are able to deal with the regression problem are discussed, categorizing them in terms of their relevant characteristics and linking them to contributions from different fields.
References
More filters
Journal ArticleDOI

Bagging predictors

Leo Breiman
TL;DR: Tests on real and simulated data sets using classification and regression trees and subset selection in linear regression show that bagging can give substantial gains in accuracy.
Journal ArticleDOI

The random subspace method for constructing decision forests

TL;DR: A method to construct a decision tree based classifier is proposed that maintains highest accuracy on training data and improves on generalization accuracy as it grows in complexity.
Journal ArticleDOI

Neural network ensembles

TL;DR: It is shown that the remaining residual generalization error can be reduced by invoking ensembles of similar networks, which helps improve the performance and training of neural networks for classification.
Proceedings Article

Neural Network Ensembles, Cross Validation, and Active Learning

TL;DR: It is shown how to estimate the optimal weights of the ensemble members using unlabeled data and how the ambiguity can be used to select new training data to be labeled in an active learning scheme.
Journal ArticleDOI

Ensemble learning via negative correlation

TL;DR: The experimental results show that negative correlation learning can produce neural network ensembles with good generalisation ability.