scispace - formally typeset
Open AccessJournal ArticleDOI

Attribute selection with fuzzy decision reducts

TLDR
This paper introduces the concept of fuzzy decision reducts, dependent on an increasing attribute subset measure, and presents a generalization of the classical rough set framework for data-based attribute selection and reduction using fuzzy tolerance relations.
About
This article is published in Information Sciences.The article was published on 2010-01-01 and is currently open access. It has received 227 citations till now. The article focuses on the topics: Attribute domain & Fuzzy set operations.

read more

Citations
More filters
Journal ArticleDOI

Machine learning

TL;DR: Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis.
Journal ArticleDOI

mr2PSO: A maximum relevance minimum redundancy feature selection method based on swarm intelligence for support vector machine classification

TL;DR: This hybrid algorithm, called maximum relevance minimum redundancy PSO (mr^2PSO), is novel in the sense that it uses the mutual information available from the filter model to weigh the bit selection probabilities in the discrete PSO.
Journal ArticleDOI

Feature selection in mixed data

TL;DR: It is proved that the newly-defined entropy meets the common requirement of monotonicity and can equivalently characterize the existing attribute reductions in the fuzzy rough set theory.
Journal ArticleDOI

A Fitting Model for Feature Selection With Fuzzy Rough Sets

TL;DR: A parameterized fuzzy relation is introduced to characterize the fuzzy information granules, using which the fuzzy lower and upper approximations of a decision are reconstructed and a new fuzzy rough set model is introduced.
Journal ArticleDOI

Feature subset selection based on fuzzy neighborhood rough sets

TL;DR: This paper constructs a novel rough set model for feature subset selection, and defines the dependency between fuzzy decision and condition attributes and employ the dependency to evaluate the significance of a candidate feature, using which a greedyfeature subset selection algorithm is designed.
References
More filters
Journal ArticleDOI

Random Forests

TL;DR: Internal estimates monitor error, strength, and correlation and these are used to show the response to increasing the number of features used in the forest, and are also applicable to regression.
Book

Fuzzy sets

TL;DR: A separation theorem for convex fuzzy sets is proved without requiring that the fuzzy sets be disjoint.
Book

Data Mining: Practical Machine Learning Tools and Techniques

TL;DR: This highly anticipated third edition of the most acclaimed work on data mining and machine learning will teach you everything you need to know about preparing inputs, interpreting outputs, evaluating results, and the algorithmic methods at the heart of successful data mining.
Book

The Elements of Statistical Learning: Data Mining, Inference, and Prediction

TL;DR: In this paper, the authors describe the important ideas in these areas in a common conceptual framework, and the emphasis is on concepts rather than mathematics, with a liberal use of color graphics.
Frequently Asked Questions (9)
Q1. What contributions have the authors mentioned in the paper "Aberystwyth university attribute selection with fuzzy decision reducts" ?

Therefore, within the context of fuzzy rough set theory, the authors present a generalization of the classical rough set framework for databased attribute selection and reduction using fuzzy tolerance relations. The paper unifies existing work in this direction, and introduces the concept of fuzzy decision reducts, dependent on an increasing attribute subset measure. Experimental results demonstrate the potential of fuzzy decision reducts to discover shorter attribute subsets, leading to decision models with a better coverage and with comparable, or even higher accuracy. 

For each of the fuzzy-rough measures introduced, the authors ran QuickReduct once with α = 1, and a second time with a fixed α < 1; in particular, a value of α = 0.95 was deemed a suitable overall choice for most measures, except for g, which requires a much higher threshold, and for which α = 0.9999 was selected. 

When interpreting these results, one should always keep in mind the trade-off between accuracy (RMSE) and attribute subset size: a higher accuracy (lower RSME) is of course desirable, but so is a smaller subset size, i.e., the less conditional attributes there are in the reduced data set, the stronger its generalization capacity. 

In the general definition that the authors propose, the authors require an increasing [0, 1]-valued measure, so as to guarantee that the larger an attribute subset, the higher its degree of fuzzy decision reducthood (monotonicity), which is in analogy to other approaches to define a degree of approximating decision classes [43, 44]. 

As the authors have shown, the evaluation measures γ, γ′, δ, δ′, f and g introduced in the previous subsections all give rise to corresponding fuzzy decision reducts. 

as seen in Figure 2b), if a 1% accuracy drop is permissible, fuzzy γ-decision reducts manage to reduce the subset size by over 40%, while with g a reduction of the data set by more than 63% is possible. 

This affects QuickReduct’s operation adversely; when all of the considered subsets in a given iteration evaluate to 0, the heuristic is forced to select one without any information about its true merit. 

In their experiments, the authors have used the very simple K-nearest neighbour classifier [1], implemented in Weka [53] as IBk, with default parameters (K = 1, no distance weighting). 

Their experiments clearly endorse the benefit of using fuzzy decision reducts, showing a greater flexibility and better potential to produce good-sized, highquality attribute subsets than the crisp decision reducts that have been used so far in fuzzy-rough data analysis.