An efficient hybrid sine-cosine Harris hawks optimization for low and high-dimensional feature selection

•Every growing data volume also incorporate extraordinary dimensions.•Hybrid sine-cosine Harris hawks optimization (SCHHO) is proposed for feature selection.•Results of numerical problems and feature selection reveal efficacy of SCHHO.•SCHHO outperforms other counterparts and hybrid approached from...

Full description

Saved in:
Bibliographic Details
Published inExpert systems with applications Vol. 176; p. 114778
Main Authors Hussain, Kashif, Neggaz, Nabil, Zhu, William, Houssein, Essam H.
Format Journal Article
LanguageEnglish
Published New York Elsevier Ltd 15.08.2021
Elsevier BV
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:•Every growing data volume also incorporate extraordinary dimensions.•Hybrid sine-cosine Harris hawks optimization (SCHHO) is proposed for feature selection.•Results of numerical problems and feature selection reveal efficacy of SCHHO.•SCHHO outperforms other counterparts and hybrid approached from recent literature. Feature selection, an optimization problem, becomes an important pre-process tool in data mining, which simultaneously aims at minimizing feature-size and maximizing model generalization. Because of large search space, conventional optimization methods often fail to generate global optimum solution. A variety of hybrid techniques merging different search strategies have been proposed in feature selection literature, but mostly deal with low dimensional datasets. In this paper, a hybrid optimization method is proposed for numerical optimization and feature selection, which integrates sine-cosine algorithm (SCA) in Harris hawks optimization (HHO). The goal of SCA integration is to cater ineffective exploration in HHO, moreover exploitation is enhanced by dynamically adjusting candidate solutions for avoiding solution stagnancy in HHO. The proposed method, namely SCHHO, is evaluated by employing CEC’17 test suite for numerical optimization and sixteen datasets with low and high-dimensions exceeding 15000 attributes, and compared with original SCA and HHO, as well as, other well-known optimization methods like dragonfly algorithm (DA), whale optimization algorithm (WOA), grasshopper optimization algorithm (GOA), Grey wolf optimization (GWO), and salp swarm algorithm (SSA); in addition to state-of-the-art methods. Performance of the proposed method is also validated against hybrid methods proposed in recent related literature. The extensive experimental and statistical analyses suggest that the proposed hybrid variant of HHO is able to produce efficient search results without additional computational cost. With increased convergence speed, SCHHO reduced feature-size up to 87% and achieved accuracy up to 92%. Motivated from the findings of this study, various potential future directions are also highlighted.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:0957-4174
1873-6793
DOI:10.1016/j.eswa.2021.114778