scispace - formally typeset
Journal ArticleDOI

A new fusion of grey wolf optimizer algorithm with a two-phase mutation for feature selection

Reads0
Chats0
TLDR
A new Grey Wolf Optimizer algorithm integrated with a Two-phase Mutation to solve the feature selection for classification problems based on the wrapper methods to reduce the number of selected features while preserving high classification accuracy.
Abstract
Because of their high dimensionality, dealing with large datasets can hinder the data mining process. Thus, the feature selection is a pre-process mandatory phase for reducing the dimensionality of datasets through using the most informative features and at the same time maximizing the classification accuracy. This paper proposes a new Grey Wolf Optimizer algorithm integrated with a Two-phase Mutation to solve the feature selection for classification problems based on the wrapper methods. The sigmoid function is used to transform the continuous search space to the binary one in order to match the binary nature of the feature selection problem. The two-phase mutation enhances the exploitation capability of the algorithm. The purpose of the first mutation phase is to reduce the number of selected features while preserving high classification accuracy. The purpose of the second mutation phase is to attempt to add more informative features that increase the classification accuracy. As the mutation phase can be time-consuming, the two-phase mutation can be done with a small probability. The wrapper methods can give high-quality solutions so we use one of the most famous wrapper methods which called k-Nearest Neighbor (k-NN) classifier. The Euclidean distance is computed to search for the k-NN. Each dataset is split into training and testing data using K-fold cross-validation to overcome the overfitting problem. Several comparisons with the most famous and modern algorithms such as flower algorithm, particle swarm optimization algorithm, multi-verse optimizer algorithm, whale optimization algorithm, and bat algorithm are done. The experiments are done using 35 datasets. Statistical analyses are made to prove the effectiveness of the proposed algorithm and its outperformance.

read more

Citations
More filters
Journal ArticleDOI

Orthogonal learning covariance matrix for defects of grey wolf optimizer: Insights, balance, diversity, and feature selection

TL;DR: This paper develops a GWO variant enhanced with a covariance matrix adaptation evolution strategy (CMAES), levy flight mechanism, and orthogonal learning (OL) strategy named GWOCMALOL, which could reach higher classification accuracy and fewer feature selections than other optimization algorithms.
Journal ArticleDOI

Review of swarm intelligence-based feature selection methods

TL;DR: A comparative analysis of different feature selection methods is presented, and a general categorization of these methods is performed, which shows the strengths and weaknesses of the different studied swarm intelligence-based feature selection Methods.
Journal ArticleDOI

An efficient Henry Gas Solubility Optimization for Feature Selection

TL;DR: A novel approach to dimensionality reduction is proposed by using Henry gas solubility optimization (HGSO) algorithm for selecting significant features, to enhance the classification accuracy by producing 100% accuracy on classification problems with more than 11,000 features.
Journal ArticleDOI

Dynamic Salp swarm algorithm for feature selection

TL;DR: The proposed Dynamic Salp swarm algorithm (DSSA) outperformed the original SSA and the other well-known optimization algorithms over the 23 datasets in terms of classification accuracy, fitness function values, the number of selected features, and convergence speed.
Journal ArticleDOI

A hyper learning binary dragonfly algorithm for feature selection: A COVID-19 case study

TL;DR: A novel Hyper Learning Binary Dragonfly Algorithm is proposed as a wrapper-based method to find an optimal subset of features for a given classification problem and demonstrates the superiority of HLBDA in increasing prediction accuracy and reducing the number of selected features.
References
More filters
Proceedings Article

A study of cross-validation and bootstrap for accuracy estimation and model selection

TL;DR: The results indicate that for real-word datasets similar to the authors', the best method to use for model selection is ten fold stratified cross validation even if computation power allows using more folds.
Journal ArticleDOI

Grey Wolf Optimizer

TL;DR: The results of the classical engineering design problems and real application prove that the proposed GWO algorithm is applicable to challenging problems with unknown search spaces.
Proceedings ArticleDOI

A discrete binary version of the particle swarm algorithm

TL;DR: The paper reports a reworking of the particle swarm algorithm to operate on discrete binary variables, where trajectories are changes in the probability that a coordinate will take on a zero or one value.
Journal ArticleDOI

An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression

TL;DR: Kernel and nearest-neighbor regression estimators are local versions of univariate location estimators, and so they can readily be introduced to beginning students and consulting clients who are familiar with such summaries as the sample mean and median.
Journal ArticleDOI

A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms

TL;DR: The basics are discussed and a survey of a complete set of nonparametric procedures developed to perform both pairwise and multiple comparisons, for multi-problem analysis are given.
Related Papers (5)