1. Optimal Feature Selection Based on Discrete Grasshopper Optimization Algorithm and K-nearest Neighbor Classifier.
- Author
-
Yu-Liang Qi, Jie-Sheng Wang, Yu-Wei Song, Yu-Cai Wang, Hao-Ming Song, and Jia-Ning Hou
- Subjects
- *
OPTIMIZATION algorithms , *METAHEURISTIC algorithms , *K-nearest neighbor classification , *FEATURE selection , *HEURISTIC algorithms , *DATA reduction , *ALGORITHMS - Abstract
In the majority of data mining tasks, feature selection serves as an essential pre-processing step. The most important attributes are selected to lower the dimensionality reduction of data set and enhance the precision of classification. Natural heuristic algorithms are extensively employed in the realm of encapsulated feature selection. Based on the wrapper feature selection method, seven natural heuristic algorithms are used to solve feature selection problems and perform performance comparison, which include Slime Mold Algorithm (SMA), Whale Optimization Algorithm (WOA), Harris Hawks Optimization Algorithm (HHO), Marine Predator Algorithm (MPA), Butterfly Optimization Algorithm (BOA), Cuckoo Search (CS) and Firefly Algorithm (FA). At the same time, performance tests are carried out on 21 standard UCI data sets to verily the functionality of various algorithms, and the convergence curves and accuracy boxplots of 7 natural heuristic algorithms on 21 data sets are given. The simulation outcomes were assessed utilizing the mean and standard deviation of fitness, as well as the number of chosen features, and the running time, with the optimal value in bold. By comparing the comprehensive performance indexes, MPA obtained the maximum mean fitness value in most data sets (16 data sets), followed by FA (6 data sets). SMA obtained the best performance and finds the minimum eigenvalues (20 data sets) in multiple data sets and has an advantage in computing time. [ABSTRACT FROM AUTHOR]
- Published
- 2024