Open AccessProceedings Article
Feature selection for ensembles
David W. Opitz
- pp 379-384
Reads0
Chats0
TLDR
This paper presents an ensemble feature selection approach that is based on genetic algorithms and shows improved performance over the popular and powerful ensemble approaches of AdaBoost and Bagging and demonstrates the utility of ensemble features selection.Abstract:
The traditional motivation behind feature selection algorithms is to find the best subset of features for a task using one particular learning algonthm. Given the recent success of ensembles, however, we investigate the notion of ensemble feature selection in this paper. This task is harder than traditional feature selection in that one not only needs to find features germane to the learning task and learning algorithm, but one also needs to find a set of feature subsets that will promote disagreement among the ensemble's classifiers. In this paper, we present an ensemble feature selection approach that is based on genetic algorithms. Our algorithm shows improved performance over the popular and powerful ensemble approaches of AdaBoost and Bagging and demonstrates the utility of ensemble feature selection.read more
Citations
More filters
MonographDOI
Combining Pattern Classifiers
TL;DR: This combining pattern classifiers methods and algorithms helps people to enjoy a good book with a cup of coffee in the afternoon, instead they cope with some harmful virus inside their computer.
Journal ArticleDOI
Ensemble-based classifiers
TL;DR: This paper reviews existing ensemble techniques and can be served as a tutorial for practitioners who are interested in building ensemble based systems.
Proceedings ArticleDOI
A streaming ensemble algorithm (SEA) for large-scale classification
W. Nick Street,Yong Seog Kim +1 more
TL;DR: A fast algorithm for large-scale or streaming data that classifies as well as a single decision tree built on all the data, requires approximately constant memory, and adjusts quickly to concept drift is presented.
Journal ArticleDOI
Feature selection in machine learning: A new perspective
TL;DR: This study discusses several frequently-used evaluation measures for feature selection, and surveys supervised, unsupervised, and semi-supervised feature selection methods, which are widely applied in machine learning problems, such as classification and clustering.
Journal ArticleDOI
Diversity creation methods: a survey and categorisation
TL;DR: This paper reviews the varied attempts to provide a formal explanation of error diversity, including several heuristic and qualitative explanations in the literature, and introduces the idea of implicit and explicit diversity creation methods, and three dimensions along which these may be applied.
References
More filters
Book
Adaptation in natural and artificial systems
TL;DR: Names of founding work in the area of Adaptation and modiication, which aims to mimic biological optimization, and some (Non-GA) branches of AI.
Journal ArticleDOI
Bagging predictors
TL;DR: Tests on real and simulated data sets using classification and regression trees and subset selection in linear regression show that bagging can give substantial gains in accuracy.
Book
An Introduction to Genetic Algorithms
TL;DR: An Introduction to Genetic Algorithms focuses in depth on a small set of important and interesting topics -- particularly in machine learning, scientific modeling, and artificial life -- and reviews a broad span of research, including the work of Mitchell and her colleagues.
Journal ArticleDOI
Wrappers for feature subset selection
Ron Kohavi,George H. John +1 more
TL;DR: The wrapper method searches for an optimal feature subset tailored to a particular algorithm and a domain and compares the wrapper approach to induction without feature subset selection and to Relief, a filter approach tofeature subset selection.