scispace - formally typeset
Proceedings ArticleDOI

Evaluating the Importance of each Feature in Classification task

TLDR
This research used two techniques namely Data partition and K Fold, in evaluating the importance of each feature from the randomly generated dataset with 5399 instances and 20 attributes to improve the classification accuracy by knowing the most important feature from any given dataset.
Abstract
In Machine Learning and statistics attribute/feature selection is used in predictive model construction. This help the Machine in interpreting the features easily by discovering good insight and improves efficiency in predictive modeling. The objective of our research is to improve the classification accuracy by knowing the most important feature from any given dataset. In this research, we used two techniques namely Data partition and K Fold, in evaluating the importance of each feature from the randomly generated dataset with 5399 instances and 20 attributes. In Data partitioning, the attribute with lowest accuracy is filtered out. Where as in K Fold cross validation, attributes with biggest error is removed from the original dataset. In our experiments, the evaluation parameters considered are Recall. Precision and F-Measure. Finally the accuracy rate of both the techniques are compared. The finding in our research stats that K Fold approach achieves better accuracy of 97.03% than Data partitioning(96.11%) in estimating the importance of features in classification.

read more

References
More filters
Journal ArticleDOI

A survey on feature selection methods

TL;DR: The objective is to provide a generic introduction to variable elimination which can be applied to a wide array of machine learning problems and focus on Filter, Wrapper and Embedded methods.
Journal ArticleDOI

Online selection of discriminative tracking features

TL;DR: This paper presents an online feature selection mechanism for evaluating multiple features while tracking and adjusting the set of features used to improve tracking performance, and notes susceptibility of the variance ratio feature selection method to distraction by spatially correlated background clutter.
Journal ArticleDOI

Support vector machines combined with feature selection for breast cancer diagnosis

TL;DR: The results show that the highest classification accuracy (99.51%) is obtained for the SVM model that contains five features, and this is very promising compared to the previously reported results.
Journal ArticleDOI

Robust Joint Graph Sparse Coding for Unsupervised Spectral Feature Selection

TL;DR: This paper proposes a new unsupervised spectral feature selection model by embedding a graph regularizer into the framework of joint sparse regression for preserving the local structures of data by proposing a novel joint graph sparse coding (JGSC) model.
Journal ArticleDOI

A feature selection model based on genetic rank aggregation for text sentiment classification

TL;DR: An ensemble approach for feature selection is presented, which aggregates the several individual feature lists obtained by the different feature selection methods so that a more robust and efficient feature subset can be obtained.
Related Papers (5)
Trending Questions (1)
What are the advantages and disadvantages of different feature importance calculation methods?

The paper does not provide specific advantages and disadvantages of different feature importance calculation methods.