Hybrid binary bat enhanced particle swarm optimization algorithm for solving feature selection problems

Mohamed A. Tawhid (Department of Mathematics and Statistics, Faculty of Science, Thompson Rivers University, Kamloops, Canada) (Department of Mathematics and Computer Science, Faculty of Science, Alexandria University, Alexandria, Egypt)
Kevin B. Dsouza (Department of Electrical and Computer Engineering, The University of British Columbia, Vancouver, Canada)

Applied Computing and Informatics

ISSN: 2634-1964

Article publication date: 11 April 2018

979

Abstract

In this paper, we present a new hybrid binary version of bat and enhanced particle swarm optimization algorithm in order to solve feature selection problems. The proposed algorithm is called Hybrid Binary Bat Enhanced Particle Swarm Optimization Algorithm (HBBEPSO). In the proposed HBBEPSO algorithm, we combine the bat algorithm with its capacity for echolocation helping explore the feature space and enhanced version of the particle swarm optimization with its ability to converge to the best global solution in the search space. In order to investigate the general performance of the proposed HBBEPSO algorithm, the proposed algorithm is compared with the original optimizers and other optimizers that have been used for feature selection in the past. A set of assessment indicators are used to evaluate and compare the different optimizers over 20 standard data sets obtained from the UCI repository. Results prove the ability of the proposed HBBEPSO algorithm to search the feature space for optimal feature combinations.

Keywords

Citation

Tawhid, M.A. and Dsouza, K.B. (2018), "Hybrid binary bat enhanced particle swarm optimization algorithm for solving feature selection problems", Applied Computing and Informatics, Vol. 16 No. 1/2, pp. 117-136. https://doi.org/10.1016/j.aci.2018.04.001

Publisher

:

Emerald Publishing Limited

Copyright © 2018, Mohamed A. Tawhid and Kevin B. Dsouza

License

Published in Applied Computing and Informatics. Published by Emerald Publishing Limited. This article is published under the Creative Commons Attribution (CC BY 4.0) license. Anyone may reproduce, distribute, translate and create derivative works of this article (for both commercial and non-commercial purposes), subject to full attribution to the original publication and authors. The full terms of this license may be seen at http://creativecommons.org/licences/by/4.0/legalcode


1. Introduction

Feature selection is a way for identifying the independent features and removing expendable ones from the dataset [1]. The objectives of feature selection are dimensionality reduction of the data, improving accuracy of prediction, and understanding data for different machine learning applications such as clustering, classification, regression and computer vision [2]. It is also widely used in the analysis of economic and trade markets. In the real world, data representation often uses too many features, which means certain independent features can fill in for others and the redundant features can be removed. Moreover, the output is influenced by the relevant features because they contain important information about the data and the results will be obscure if any of them are left out [3]. The classical optimization techniques have some limitations in solving the feature selection problems [4] and hence evolutionary computation (EC) algorithms are the alternative for solving these limitations and searching for the best solution [5]. Evolutionary Computation (EC) algorithms are inspired by nature, group dynamics, social behavior, and biological interaction of species in a group. The binary version of these algorithms allow us to investigate problems like feature selection and arrive at superior results.

Many heuristic algorithms have been used in an attempt to solve the feature selection problem. A survey on evolutionary computation approaches to feature selection is explained in [5]. A binary PSO based method with a mutation operator is introduced in [6] to achieve spam detection using decision trees. A wavelet entropy based feature selection approach is used in [7] to detect abnormal MR brains. Ref. [8] delineates about a Firefly based feature selection approach. A binary bat based feature selection method is shed light upon in [9]. Ref. [35] elaborates a feature subset selection approach by Grey wolf optimization. Even hybrid algorithms have been used to solve feature selection problems. A hybrid genetic algorithm on mutual information is presented in [10]. Ref. [11] expounds a hybrid flower pollination algorithms for feature selection.

Bat algorithm was recently developed by Yang which is based on the ability of bats to use echolocation to sense distance and also to distinguish between prey and background barriers [12]. The bat algorithm and its variants have been used in many computing applications. A binary bat algorithm is suggested in [37] to solve unconstrained optimization bench test problems and compared with binary GA and binary PSO. Also, a binary bat algorithm for feature selection is presented in [9]. A combination of K-means and bat algorithm is used for efficient clustering in [13]. A variant fuzzy bat algorithm is proposed in [14]. Multi-objective optimization problems are dealt with to solve engineering design benchmarks in [15]. A variant of bat algorithm using differential operator and Levy flights to solve function optimization problems is delineated in [16].

Particle swarm optimization (PSO) is a population based stochastic optimization technique developed by Eberhart and Kennedy in 1995 [17], inspired by social behavior of bird flocking and fish schooling. A comprehensive survey about PSO and its applications can be found at [18]. In past several years, even though PSO has been successfully applied in many research and application areas like the constrained non-linear optimization problems [19], for optimal design of combinational logic circuits [20] and also to real world hydraulic problems [21], little work is seen in the domain of feature selection [22]. A multi-objective approach using PSO is introduced in [23,24]. Also, a bare bones PSO technique is delineated in [25]. It is demonstrated that PSO gets better results in a faster, cheaper way compared with other methods. Another reason that PSO is attractive is that there are few parameters to tweak. One version, with slight variations, works well in a wide variety of real world applications. Here an enhanced version of the standard PSO is used [26] to solve the feature selection problem.

Hybridization of different algorithmic concepts is a method to obtain better performing systems and is believed to benefit from synergy, i.e. usually it exploits and unites advantages of the individual pure strategies. It is mostly due to the no free lunch theorems [27], that the generalized view of metaheuristics changed and people recognized that there cannot exist a general optimization strategy which is globally better than any other. In fact, to solve a problem at hand most effectively, it almost always requires a specialized algorithm that needs to be compiled of adequate parts. Hybridization is classified into many categories [28,29]. Hybridization of one metaheuristic with another is a popular method to enhance the performance of both the algorithms.

The aim of this work is to propose a new hybrid binary version of bat and enhanced PSO algorithm in order to solve feature selection problems effectively. The hybridization allows us to combine the best aspects of both these algorithms and obtain better performance. In this paper, we propose a new hybrid algorithm, which is called HBBEPSO Algorithm by combining the bat Algorithm with the enhanced PSO algorithm in order to obtain superior results when compared to the respective individual algorithms. The binary HBBEPSO algorithm is tested on 20 standard data sets obtained from the UCI repository [30]. The algorithm is also compared with the HBEPSOB, where the PSO is carried out first and then given to the bat algorithm. A set of assessment indicators is used to evaluate and compare the different optimizers. The experimental results show the ability of the proposed binary HBBEPSO algorithm to search the feature space for optimal feature combinations.

The reminder of this paper is organized as follows. Section 2 presents the definition of the feature selection problem. Section 3 summarizes the main concepts of the bat algorithm. Section 4 describes the main concepts of the enhanced PSO algorithm. Section 5 presents the main structure of the proposed binary HBBEPSO algorithm. The Section 6 provides details about the feature selection problem, evaluation criteria and an insight about the classifier used. Section 7 reports the experimental results and finally, the conclusion and some future work make up Section 8.

2. Definition of the feature selection problem

In real life machine learning applications thousand of features are measured while only handful of them contain useful information. Therefore, we need methods to reduce the dimensionality of our feature set. This can be achieved by two ways, feature reduction and feature selection. Feature reduction is when we apply some sort of transformation on our original feature set of dimension d to produce a new feature set of dimension m with m<d. Techniques like Linear Discriminant Analysis (LDA), and Principal Component Analysis (PCA) come under this category. Feature selection is the process of selecting a subset of the original features. In this section, we present the definition of the feature selection problem as follows.

The feature selection problem can be defined as the selection of certain number of features out of the total number of available features in such a way that the classification performance is maximum and the number of selected features is minimum.

(1)fitness=αγR(D)+β|CR||C|
where γR(D) is the classification quality of set R relative to decision D, R is the length of selected feature subset, C is the total number of features, α and β are two parameters corresponding to the importance of classification quality and subset length, α[0,1] and β=1α. The fitness function maximizes the classification quality, γR(D), and the ratio of the unselected features to the total number of features is described by |CR||C|. The above equation can be easily converted into a minimization problem by using error rate rather than classification quality and using selected features ratio rather than using unselected feature size. The minimization problem can be formulated as in Eq. (2).
(2)fitness=αER(D)+β|R||C|
where ER(D) is the error rate of the classifier, R is the length of the selected feature subset and C is the total number of features. α[0,1] and β=1α are constants used to control the weights of classification accuracy and feature reduction.

3. Overview of binary bat algorithm

In the following subsection, we will give an overview of the main concepts and structure of the binary bat algorithm.

3.1 Main concepts and inspiration

The binary bat algorithm mimics the concept of echolocation of bats to sense distance and distinguish between prey and background barriers. The bats send out loud, short pulses of sound and can sense the distance by the time it takes for the echo to return to them [31]. This fascinating mechanism also allows bats to distinguish between barrier and prey, thus allowing them to hunt in complete darkness [32].

3.2 Definition of concepts

  • 1.

    Loudness: This parameter (a) is used to eliminate solutions that are too loud and will hinder the bat from reaching the optimum. This mimics the loudness of the bat pulse.

  • 2.

    Pulse rate: This parameter (r) mimics the rate of pulsing of the bat. It randomly assigns the best solution in the previous iteration to the present solution.

  • 3.

    Frequency: This parameter (Qi) is used to represent the frequency of the bat wave. It varies from a minimum to a maximum and the changes occur randomly. This parameter gives the weight to the separation of the current solution from the best solution in the space. The frequency is represented by a D dimensional vector and is initialized to zero.

(3)Qi=(Qi1,Qi2,,QiD)
  • 4.

    Velocity: This parameter (vi) is the resultant velocity of the bat at every iteration. The velocity is represented by a D dimensional vector and is initialized to zero.

(4)vi=(vi1,vi2,,viD)

3.3 Binary bat algorithm

Using the concepts mentioned in the Section 3.2, the binary bat algorithm distinguishes between barrier and prey. It should also be noted that the bats can change the wavelength of their emitted pulses and the rate of emission based on their relative position with respect to the targets. In the context of feature selection, this gives the algorithm flexibility to adapt the changes in the feature space and explore better solutions.

The details of the algorithm are mentioned in Algorithm 1. We provide a brief overview of the binary bat algorithm. After initializing the position, frequency and velocity vectors, the best solution is noted and updated throughout the algorithm. This is done mainly using the following equations

(5)Qi=Qmin+(Qmin-Qmax)·rand
(6)v(i,j)=v(i,j)+(x(i,j)-best(j))·Qi
(7)x(i,j)=x(i,j)+v(i,j)
where rand denotes a randomly generated number in the interval (0,1) and x represents the new solutions. These new solutions may not always be adopted and are updated depending on certain other parameters in the algorithm. A threshold is selected depending on the value of the velocity of the bat which will control the amount of exploration, the bat is capable of achieving as is given in Eq. (8). If a particular random number is less than this threshold value, the new solutions are updated and the bat moves on to the new solution space.
(8)V_value=2πarctanπ2v(i,j).

The rate of pulse emission decides whether the bat will stick to the previous best solution obtained or adopt the newly updated solution. This is similar to the best global solution adoption step in most meta heuristics and helps to steer and clear off too much unnecessary exploration. The loudness parameter introduces a further filter to the adoption of the solution as the new accepted solution. The solution is only accepted if a random number chosen is lesser than the loudness value and the fitness of the new solution is better than the old solution.

As this solutions will be in the continuous space, a binary map needs to be applied to the solutions so as to make them compatible to feature selection. This map could be a regular squashing function like a sigmoid function or any other function that is capable of taking continuous values into the logistic space.

4. Overview of binary enhanced particle swarm optimization

In the following section, we will give an overview of the main concepts and structure of the binary enhanced PSO algorithm.

4.1 Main concepts and inspiration

The PSO is a population based search method inspired from the swarm behavior (information interchange) of birds [33]. In PSO, initially a random population of particles is initialized and these particles move with certain velocity based on their interaction with other particles in the population. At each iteration the personal best achieved by each particle and the global best of all the particles is tracked and the velocity of all the particles is updated based on this information. Certain parameters are used to give weights to the global and personal best. In the enhanced version of the binary PSO [26], special type of S shaped transfer functions is used to convert a continuous value to a binary value instead of a simple hyperbolic tangent function.

4.2 Movement of particles

Each of the particles is represented by D dimensional vectors and they are randomly initialized with each individual value being binary.

(9)xi=(xi1,xi2,,xiD)S
where S is the available search space. The velocity is represented by a D dimensional vector and is initialized to zero,
(10)vi=(vi1,vi2,,viD).
The best personal(local) position recorded by each particle is maintained as
(11)pi=(pi1,pi2,,piD)S.

At each iteration, each particle changes its position according to its personal best(Pbest) and the global best(gbest) as follows

(12)vi(t+1)=wvi(t)+c1ri1·(Pbesti(t)-xi(t))+c2ri2·(gbest-xi(t))
where c1 and c2 are acceleration constants called cognitive and social parameters respectively. r1 and r2 are random values [0,1]. w is called as the inertia weight. It determines how the previous velocity of the particle influences the velocity in the next iteration. The value of w is determined by the following expression
(13)w=wmax-iteration·wmax-wminMax_iteration
where wmax and wmin are constants. Max_iteration is the maximum number of iterations.

4.3 The continuous to binary map

The position of each particle is determined by the S shaped as a transfer function that maps the continuous velocity value to the position of the particle. This is a special sigmoid function that enhances the PSO.

(14)s=11+e-vi,j,i=1,,SS,j=1,,D
X(i,j)=1if(rand<s)0otherwise

4.4 Enhanced PSO algorithm

In this section, we present in details the main steps of the binary enhanced PSO algorithm as shown in Algorithm 2.

After initializing the solutions and the velocity values, the personal and the global best of the particles are noted throughout the algorithm. These values play an important role in directing the velocity values of the particles and need to be revisited in each iteration. The velocity values are updated according Eq. (12) at each stage and these new values are used to change the positions of the particles. It should be noted that, to make sure that the values don’t cross the given thresholds its necessary to restrict the velocity values to their assigned maximum and minimum values. These new velocity values are then crushed into the logistic space by applying the binary map as mentioned in Section 4.3 and a threshold is used to update the positions.

This particular algorithm makes the best use of both the personal and the global solutions to arrive at globally optimum solutions. The inertia weight that’s updated in every iteration also helps to control the convergence of the algorithm as it progresses. The inertia towards the previous direction is pretty high initially when the algorithm starts but it starts to explore new directions during its progress. This value can be tuned to arrive at better solutions and needs to be tried with grid search over the possible parameter values.

5. Hybrid Binary Bat Enhanced Particle Swarm Optimization (HBBEPSO) algorithm

The main steps of the proposed HBBEPSO algorithm for feature selection are shown in Algorithm 3 and summarized in this section.

The combination of the concepts in the binary bat and the binary PSO algorithms that are described in the previous sections are combined in this section to arrive at an algorithm that can benefit from their amalgamation. In the HBBEPSO algorithm, the decoupling of the velocity vectors of the bats and the particles leads us to a novel formulation. The velocity vectors are updated independently for both the particles according to the weighted combination of the personal and global best solutions and the velocity of the bats is arrived in an instantaneous manner. This is done in order to allow both the algorithms to explore the search space in an alternating fashion and not direct one algorithm with the results obtained from the other algorithm. This form of decoupling is also why the personal and global solutions are not updated after the binary bat algorithm but only after the particle swarm iteration update i.e. once per the whole iteration.

This leads us to some interesting insights that decoupling these variables that accomplish the same goal but in different ways is in fact beneficial for the hybrid algorithm because it benefits from the diversity of the solutions in each iteration which is also the main philosophy behind hybridizing algorithms. It has to be noted that choosing the hyperparameters is important for getting good solutions and can be achieved by a simple grid search or a random search over the hyperparameter space.

6. Feature selection

The feature selection problem is as defined in Section 2. For a feature vector of size N the number of different feature combinations would be 2N, which is a huge space to search exhaustively. So the proposed hybrid metaheuristic algorithm is used to adaptively search the feature space and produce the best feature combination. The fitness function used is the one given in Eq. (2).

6.1 Classifier

K-nearest neighbor (KNN) [34] is a common simple method used for classification. KNN is a supervised learning algorithm that classifies an unknown sample instance based on the majority vote of its K-nearest neighbors. Here, a wrapper approach to feature selection is used which uses KNN classifier as a guide for the same. Classifiers do not use any model for K-nearest neighbors and are determined solely based on the minimum distance from the current query instance to the neighboring training samples. In this proposed system,the KNN is used as a classifier to ensure robustness to noisy training data and obtain best feature combinations. A single dimension in the search space represents individual feature and hence the position of a particle represents a single feature combination or solution.

7. Experimental results

The proposed binary HBBEPSO algorithm is tested against 20 data sets in Table 1 taken from the UCI machine learning repository [30] and is compared with other algorithms like binary versions of dragonfly, enhanced PSO, GA, bat and greywolf. The algorithm is also compared with HBEPSOB, where the order of implementation of the two algorithms is reversed. The datasets are chosen to have variety in number of instances and features to test for varied data. The datasets are divided into three equal sets: training, validation and testing. The training and validation sets are used for a two fold cross validation on the data. We note that other ways of implementing validation do exist and can be used like stratified K-fold cross validation, group K fold, shuffle split and many others. We take accuracy of the classifier as our main metric and rely on the 2-fold cross validation for weak statistical validation. It should be noted that metrics like uncertainty coefficient which is more robust to the relative sizes of the classes can also be used as a metric along with additional supplements like precision, recall and receiver operating characteristics to analyze the true and false positives of each of the classes.

The value of K (the number of nearest neighbors) is selected as 5 based on the 2-fold cross-validation results of the model. The training set is used to evaluate the KNN on the validation set through this algorithm to guide the feature selection process. The test data is only used for the final evaluation of the best selected feature combination. The global and optimizer-specific parameter setting is given in Table 2. The parameters are set according random search implemented over the hyperparameter space. It should be noted that better values for the hyperparameters are possible and can be obtained by using exhaustive grid search over sufficiently large parameter space assuming computational power is not an issue. The evaluation criteria is explained in Section 7.1.

7.1 Evaluation criteria

The datasets are divided into 3 sets of training, validation and testing. The algorithm is run repeatedly for M=10 times for statistical significance of the results. The following measures [35] are recorded from the validation data:

  • 1.

    Mean fitness function is the average of the fitness function value obtained from running the algorithm M times. The mean fitness function is calculated as shown in Eq. (15).

(15)Mean=1Mi=1Mgi
where gi is the best fitness value obtained at run i.
  • 2.

    Best fitness function is the minimum of the fitness function value obtained from running the algorithm M times. The best fitness function is calculated as shown in Eq. (16).

(16)Best=mini=1Mgi
where gi is the best fitness value obtained at run i.
  • 3.

    Worst fitness function is the maximum of the fitness function value obtained from running the algorithm M times. The worst fitness function is calculated as shown in Eq. (17).

(17)Worst=maxi=1Mgi
where gi is the best fitness value obtained at run i.
  • 4.

    Standard deviation gives the variation of the fitness function value obtained from running the algorithm M times. It is an indicator of the stability and robustness of the algorithm. Larger values of standard deviation would suggest wandering results where as smaller value suggests the algorithm converges to the same value most of the times. The Standard deviation is calculated as shown in Eq. (18).

(18)Std=1M-1i=1M(gi-Mean)2
where gi is the best fitness value obtained at run i.
  • 5.

    Average Performance (CA) is the mean of the classification accuracy values when an algorithm is run M times. The average performance is calculated as shown in Eq. (19).

(19)CA=1Mi=1MCAi
where CAi is the accuracy value obtained at run i
  • 6.

    Mean Feature selection ratio (FSR) is the mean of the ratio of the number of selected features to the total number of features when an algorithm is run M times. The Mean Feature selection ratio is calculated as shown in Eq. (20).

(20)FSR=1Mi=1Msize(gi)D
where gi is the best fitness value obtained at run i, size(gi) gives the number of features selected and D is the total number of features.
  • 7.

    Average F-score is a measure that evaluates the performance of a chosen feature subset. It requires that in the data spanned by the feature combination the distance between data points in different classes be large and of those in the same class be as small as possible. The Fischer index for a given feature is calculated as in Eq. (21) [36].

(21)Fj=k=1Cnk(μkj-μj)2(σj)2
(22)(σj)2=k=1Cnk(σkj)2
where Fj is the Fischer index for j, μj is the mean of the entire data for feature j, (σj)2 is defined as in Eq. (22), nk is the size of class k, μkj is the mean of class k for feature j, (σkj)2 is the variance of class k for feature j. The average F-score is calculated by taking the average of values obtained from M runs for only the selected features.

7.2 Results

The proposed binary version of the HBBEPSO algorithm is compared with the binary bat algorithm, the Enhanced PSO and other optimizers. The results are tabulated as follows.

Table 3 outlines the performance of the algorithms using the fitness function mentioned in Eq. (2) in the minimization mode. The table shows the average fitness obtained over M runs and is calculated using Eq. (15). The best performance is achieved by the proposed binary version of the HBBEPSO algorithm proving its ability to search the feature space effectively.

For testing the stability, robustness and the repeatability of convergence of these stochastic algorithms the standard deviation of the fitness values over M runs is recorded as per Eq. (18) in Table 4. The table shows that the HBBEPSO algorithm has the ability to converge repeatedly irrespective of the random initialization.

The Best selected feature combinations by the algorithms are also allowed to run on the test data and the average classification accuracy and the average feature selection ratio over M runs is recorded using Eqs. (19) and (20), respectively as shown in Tables 5 and 6. As seen from these tables, the HBBEPSO algorithm is able to select the minimum number of features and yet maintain the classification accuracy. This shows the capability of the HBBEPSO algorithm to satisfy both the objectives of optimization.

To analyze the separability and closeness of the selected features Fischer score of these features is calculated as shown in Eq. (21). The average over M runs is recorded in Table 7. As shown in the table, HBBEPSO algorithm achieves superior data compactness in comparison with the other algorithms.

It should be noted that datasets with comparable number of features and instances outperform the other datasets. Mainly we notice that datasets, if the number of features is f and the number of instances is n, then, datasets with f<n and n=10f outperform the other datasets. This is expected as datasets with too few training samples compared to the number of features will substantially underfit. The tables given above show that the HBBEPSO algorithm outperforms the other algorithms with respect to all of the assessment indicators. It can also be seen that it performs much better when compared to its switched version HBEPSOB algorithm. This leads us to believe that the bat algorithm is powerful in exploring the search space and the enhanced PSO algorithm aids in exploiting the reduced feature space (see Figures 1 and 2).

8. Conclusion and future work

In this paper, a new hybrid binary metaheuristic algorithm with bat algorithm and enhanced PSO algorithm is proposed in order to solve feature selection problems. The proposed algorithm is called hybrid Binary Bat Enhanced Particle Swarm Optimization (HBBEPSO) algorithm. The two algorithms come together to give better solutions than each of them individually. In order to verify the robustness and the effectiveness of the proposed algorithm, we apply it on 20 feature selection problems. The evaluation is performed using a set of evaluation criteria to assess different aspects of the proposed system. The experimental results show that the proposed algorithm is a promising algorithm with its ability to search the feature space effectively. The given algorithm is also run on test data and observations show higher performance of the selected features when compared to the other optimizers. The Fischer index table reveals better separability. It is also noted from the values of standard deviation that the algorithm has the robustness to repeatedly converge to similar solutions therefore a powerful ability to solve feature selection problems better than other algorithms in most cases. This research motivates us for further investigations as future work as follows. The authors in [37] suggested a binary bat algorithm to solve unconstrained optimization bench test problems and compared with binary GA and binary PSO. We would like to apply our proposed algorithm on solving unconstrained optimization problems [39,40], large scale problems and molecular potential energy function [41], and engineering optimization problems [38]. Also, further comparison studies for various binary variants of bat algorithm in the literature need to be done on feature selection problem as suggested by one of the referees.

Figures

The Comparison of performance the HBBEPSO algorithm with other optimizers through main objectives of feature selection. The values are averaged over all the datasets.

Figure 1

The Comparison of performance the HBBEPSO algorithm with other optimizers through main objectives of feature selection. The values are averaged over all the datasets.

The Comparison of performance the HBBEPSO algorithm with other optimizers through few assessment indicators. The values are averaged over all the datasets.

Figure 2

The Comparison of performance the HBBEPSO algorithm with other optimizers through few assessment indicators. The values are averaged over all the datasets.

Binary bat algorithm

Enhanced PSO Algorithm

HBBEPSO Algorithm

Datasets.

Dataset# of attributes# of instances
Zoo16101
WineEW13178
IonosphereEW34351
WaveformEW405000
BreastEW30569
Breastcancer9699
Congress16435
Exactly131000
Exactly2131000
HeartEW13270
KrvskpEW363196
M-of-n131000
SonarEW60208
SpectEW60208
Tic-tac-toe9958
Lymphography18148
Dermatology34366
Echocardiogram12132
hepatitis19155
LungCancer5632

Parameter setting.

ParameterValue
# of iterations (maxiter)70
# of search agents (n)5
Dimension (D)# of features in the data
Search domain[0 1]
# of runs (M)10
a0.4
r0.1
Qmin0
Qmax2
wmax0.9
wmin0.4
c12
c22
vmax6
β in fitness function0.01
α in fitness function0.99

Mean fitness function obtained from the different algorithms.

DatasetHBBESPOBBAEPSOBGWO2BDABGAHBEPSOB
Zoo0.0140.0940.0310.110.0670.1240.053
Wine EW0.0270.1280.0420.0920.0500.0650.038
IonosphereEW0.0890.1460.1370.1720.1300.1430.117
WaveformEW0.1790.1930.1750.1850.1830.1860.176
BreastEW0.0500.0700.0500.0800.0570.1060.052
Breastcancer0.0310.0350.0320.0420.0320.0360.039
Congress0.0310.0530.0330.0730.0420.0590.035
Exactly0.1550.3030.1040.3160.1780.2690.064
Exactly20.2240.2430.2340.2630.2400.2430.230
HeartEW0.1400.2400.1530.2680.1530.2500.157
KrvskpEW0.0410.1080.0430.0800.0410.0890.043
M-of-n0.0240.1670.0240.1540.0480.1080.047
SonarEW0.1790.2770.1920.2900.1940.2620.166
SpectEW0.1320.1670.1600.2050.1330.1680.117
Tic-tac-toe0.2140.2700.2220.2620.2230.2410.231
Lymphography0.3430.4870.5310.4870.4120.4660.438
Dermatology0.0150.0810.0160.0990.0170.0310.025
Echocardiogram0.0470.1120.0830.2000.0580.0720.080
Hepatitis0.1370.1750.1230.1920.1010.1520.122
LungCancer0.1290.4270.2200.4550.2550.3180.165
Average0.1100.1890.1230.2040.1310.1690.120

Standard deviation of the fitness function obtained from the different algorithms.

DatasetHBBESPOBBAEPSOBGWO2BDABGAHBEPSOB
Zoo0.0200.0700.0330.0670.0750.0660.041
Wine EW0.0140.0800.0180.0570.0300.0260.016
IonosphereEW0.0160.0400.0160.0570.0180.0250.019
WaveformEW0.0030.0120.0080.0070.0060.0080.009
BreastEW0.0110.0170.0190.0180.0070.7550.011
Breastcancer0.0060.0090.0070.0080.0050.0070.007
Congress0.0060.0190.0080.0150.0160.0130.009
Exactly0.0640.0400.0820.0160.1190.0780.043
Exactly20.0090.0170.0090.0180.0150.0190.018
HeartEW0.0330.0640.0550.0690.0360.0620.038
KrvskpEW0.0050.0440.0070.0120.0070.0510.006
M-of-n0.0250.0360.0220.0190.0510.0320.034
SonarEW0.0250.0590.0290.0430.0330.0300.030
SpectEW0.0130.0280.0270.0240.0220.0290.027
Tic-tac-toe0.0120.0250.0200.0170.0120.0210.020
Lymphography0.0380.0470.0480.0440.0490.0620.038
Dermatology0.0070.0750.0080.0500.0140.0140.007
Echocardiogram0.0230.0550.0300.2280.0260.0240.036
Hepatitis0.0540.0430.0280.0520.0250.0380.038
LungCancer0.1130.1940.1800.0930.1510.2330.092
Average0.0250.0490.0330.0460.0360.0800.027

Average performance of the selected features by different algorithms.

DatasetHBBESPOBBAEPSOBGWO2BDABGAHBEPSOB
Zoo0.8730.7990.7910.8510.7880.8630.852
Wine EW0.9010.7260.8810.8960.9230.8860.908
IonosphereEW0.8310.8170.8290.8240.7990.8280.819
WaveformEW0.8290.7790.8090.8190.8070.8060.809
BreastEW0.9450.8420.9310.9080.9440.8920.94
Breastcancer0.9580.9570.9560.9570.9560.9570.952
Congress0.9440.8930.9430.9280.9310.9150.927
Exactly0.8350.6470.8840.6800.7980.6870.952
Exactly20.7600.7110.7380.7320.7390.7340.726
HeartEW0.8130.6480.7760.7020.810.7110.783
KrvskpEW0.9590.7720.9580.9170.9540.9060.956
M-of-n0.9800.7190.9750.8430.9490.8920.942
SonarEW0.7040.6780.6820.6820.6580.6940.694
SpectEW0.7640.7550.7570.7770.7520.7500.759
Tic-tac-toe0.7450.6470.7400.7130.7450.7340.733
Lymphography0.4220.4220.3540.3790.4170.4160.422
Dermatology0.9330.8020.9520.9080.9400.9500.957
Echocardiogram0.8880.8610.9060.8770.8930.8520.859
Hepatitis0.8250.7880.8130.7880.7880.7980.792
LungCancer0.4540.3430.3900.3450.4270.4090.484
Average0.8180.7300.8030.7760.8010.7840.813

Average selected feature ratio by different algorithms.

DatasetHBBESPOBBAEPSOBGWO2BDABGAHBEPSOB
Zoo0.3180.5120.3560.4730.3310.4120.356
Wine EW0.3920.5380.40.5160.3380.3150.292
IonosphereEW0.3760.5260.3880.5410.3970.4020.426
WaveformEW0.6610.6640.70910.6660.6760.742
BreastEW0.3250.4800.2410.4700.2830.2900.270
Breastcancer0.4770.5110.5110.6440.4220.5660.500
Congress0.3180.4930.3250.5750.3370.4120.343
Exactly0.5000.5380.5070.5760.5070.5610.561
Exactly20.3920.5460.4920.8000.3920.4000.407
HeartEW0.3840.4920.4070.4300.4070.4150.391
KrvskpEW0.4670.5130.5020.6330.4750.5300.541
M-of-n0.4760.4460.4760.9230.5300.5760.546
SonarEW0.4130.5210.4630.5330.4130.420.45
SpectEW0.4240.4810.4630.5290.4540.4250.415
Tic-tac-toe0.5110.5770.6660.8660.5550.5110.622
Lymphography0.3550.4610.4160.5350.4380.40.388
Dermatology0.4700.4940.5110.5440.4110.4790.485
Echocardiogram0.2160.5080.2660.4830.2330.2830.218
Hepatitis0.2310.5150.3210.4310.2730.2310.273
LungCancer0.3480.4980.4230.5260.3530.3800.362
Average0.4030.5160.4420.6010.4110.4340.429

Average Fischer index of the selected features by different algorithms.

DatasetHBBESPOBBAEPSOBGWO2BDABGAHBEPSOB
Zoo165112143130105156138
Wine EW1043.611,784648.4820,939374.67540.5290.078
IonosphereEW5.7834.1543.8705.0423.9864.7695.011
WaveformEW3.5912.0292.0663.4562.3142.1652.515
BreastEW7.8E+135.7E+123.3E+136.97E+122.5E+111.4E+136.2E+11
Breastcancer1.160.9421.0701.1050.7480.9230.740
Congress23.05011.00313.99618.31731.79713.04516.870
Exactly0.2120.3500.1310.2820.2590.3780.106
Exactly20.3990.2000.2670.2370.2400.2870.354
HeartEW4.742161.6223.357430.073.424140.644.721
KrvskpEW496.94639.91940.241187.5544.211023.2569.11
M-of-n1.8281.6521.7351.3731.7111.7861.586
SonarEW5E+68.2E+68.2E+61.2E+77.3E+65.5E+65.6E+6
SpectEW0.0080.0060.0050.0060.0060.0040.004
Tic-tac-toe0.1610.1360.1610.1170.0900.1190.120
Lymphography9.534.419.182.733.132.437.51
Dermatology384210343174269148173
Echocardiogram21,452130,939137653,037579.0662,931152.24
Hepatitis48.616132.03151.8053,0373.49114.4203.078
LungCancer47.55930.81040.20333.61531.14829.40531.574
Average3.9E+122.8E+111.6E+123.4E+121.3E+106.9E+113.1E+10

References

[1]B. Chizi, L. Rokach, O. Maimon, A Survey of Feature Selection Techniques, Encyclopedia of Data Warehousing and Mining, second ed., IGI Global, 2009, pp. 18881895.

[2]G. Chandrashekar, F. Sahin, A Survey on Feature Selection Methods, Electrical and Microelectronic Engineering, Rochester Institute of Technology, Rochester, NY 14623, USA, 2013.

[3]D. Bell, H. Wang, A formalism for relevance and its application in feature subset selection, Mach. Learn. 41 (2) (2000) 175195.

[4]Samina Khalid, A survey of feature selection and feature extraction techniques in machine learning, in: Science and Information Conference (SAI), 2014.

[5]Bing Xue, Mengjie Zhang, Will Browne, Xin Yao, A survey on evolutionary computation approaches to feature selection, IEEE Trans. Evolution. Comput. (2015).

[6]Y. Zhang, W. Shuihua, P. Preetha, J. Genlin, Binary PSO with mutation operator for feature selection using decision tree applied to spam detection, Knowl.-Based Syst. 64 (2014) 2231.

[7]X. Zhou, Z. Yudong, J. Genlin, Y. Jiquan, D. Zhengchao, W. Shuihua, Z. Guangshuai, P. Preetha, Detection of abnormal MR brains based on wavelet entropy and feature selection, IEEJ Trans. Electr. Electron. Eng. 11 (3) (2016) 364373.

[8]H. Banati, M. Bajaj, Fire fly based feature selection approach, IJCSI Int. J. Comp. Sci. Iss. 8 (4) (2011) 2.

[9]R.Y.M. Nakamura, L.A.M. Pereira, K.A. Costa, D. Rodrigues, J.P. Papa, X.-S. Yang, Binary bat algorithm for feature selection, in: Conference on Graphics, Patterns and Images, 2012, pp. 291297.

[10]Jinjie Huang, Yunze Cai, Xiaoming Xu, A hybrid genetic algorithm for feature selection wrapper based on mutual information, Patt. Recog. Lett. Arch. 28(13) (2007).

[11]Hossam M. Zawbaa, Aboul Ella Hassanien, E. Emary, Waleed Yamany, B. Parv, Hybrid flower pollination algorithm with rough sets for feature selection, in: 11th International Computer Engineering Conference (ICENCO), IEEE, Cairo, Egypt, December, 2015, pp. 278283.

[12]X.S. Yang, Bat algorithm for multi-objective optimisation, Int. J. Bio-Insp. Comput. 3 (5) (2011) 267274.

[13]G. Komarasamy, A. Wahi, An optimized K-means clustering technique using bat algorithm, Euro. J. Scient. Res. 84 (2) (2012) 263273.

[14]K. Khan, A. Nikov, A. Sahai, A fuzzy bat clustering method for ergonomic screening of office workplaces, in: S3T 2011, Adv. Intell. Soft Comput., vol. 101, 2011, pp. 5966.

[15]X.S. Yang, Bat algorithm for multi-objective optimisation, Int. J. BioInsp. Comput. 3 (5) (2011) 267274.

[16]J. Xie, Y.Q. Zhou, H. Chen, A novel bat algorithm based on differential operator and Levy flights trajectory, Comput. Intell. Neurosci. 2013, 453812 <http://www.hindawi.com/journals/cin/aip/453812.pdf>.

[17]R.C. Eberhart, J. Kennedy, A new optimizer using particle swarm theory, in: Proceedings of the Sixth International Symposium on Micromachine and Human Science, Nagoya, Japan, 1995, pp. 3943.

[18]Y. Zhang, W. Shuihua, J. Genlin, A comprehensive survey on particle swarm optimization algorithm and its applications, Math. Prob. Eng. (2015) 2015.

[19]G. Coath, S.K. Halgamuge, A comparison of constraint-handling methods for the application of particle swarm optimization to constrained nonlinear optimization problems, in: Proceedings of IEEE Congress on Evolutionary Computation 2003 (CEC 2003), Canbella, Australia, 2003, pp. 24192425.

[20]C.A. Coello Coello, E.H.N. Luna, A.H.N. Aguirre, Use of particle swarm optimization to design combinational logic circuits, in: Lecture Notes in Computer Science (LNCS), vol. 2606, 2003, pp. 398409.

[21]R.A. Krohling, H. Knidel, Y. Shi, Solving numerical equations of hydraulic problems using particle swarm optimization, in: Proceedings of the IEEE Congress on Evolutionary Computation (CEC 2002), Honolulu, Hawaii, USA, 2002.

[22]D.K. Agrafiotis, W. Cedeno, Feature selection for structure-activity correlation using binary particle swarms, J. Med. Chem. 45 (5) (2002) 10981107.

[23]B. Xue, Z. Mengjie, N.B. Will, Particle swarm optimization for feature selection in classification: a multi-objective approach, IEEE Trans. Cybernet. 43 6 (2013) 16561671.

[24]Y. Zhang, G. Dun-Wei, C. Jian, Multi-objective particle swarm optimization approach for cost-based feature selection in classification, IEEE/ACM Trans. Comput. Biol. Bioinf. (TCBB) 14(1) (2017) 6475.

[25]Y. Zhang, G. Dunwei, H. Ying, Z. Wanqiu, Feature selection algorithm based on bare bones particle swarm optimization, Neurocomputing 148 (2015) 150157.

[26]S. Mirjalili, A. Lewis, S-shaped versus V-shaped transfer functions for binary particle swarm optimization, Swarm Evolution. Comput. 9 (2012) 114.

[27]D. Wolpert, W. Macready, No free lunch theorems for optimization, IEEE Trans. Evolution. Comput. 1 (1) (1997) 6782.

[28]C. Cotta, A study of hybridisation techniques and their application to the design of evolutionary algorithms, AI Commun. 11 (3–4) (1998) 223224.

[29]E.G. Talbi, A taxonomy of hybrid metaheuristics, J. Heur. 8 (5) (2002) 541565.

[30]A. Frank, A. Asuncion, UCI Machine Learning Repository, 2010.

[31]D.R. Griffin, F.A. Webster, C.R. Michael, The echolocation of flying insects by bats, Animal Behav. 8(34) (1960) 141154.

[32]H.-U. Schnitzler, E.K.V. Kalko, Echolocation by insect-eating bats, BioScience 51 (7) (2001) 557569.

[33]J. Kennedy, R.C. Eberhart, Y. Shi, Swarm Intelligence, Morgan Kaufmann, SanMateo, CA, 2001.

[34]L.Y. Chuang, H.W. Chang, C.J. Tu, C.H. Yang, Improved binary PSO for feature selection using gene expression data, Comput. Biol. Chem. 32 (2008) 2938.

[35]E. Emary, Hossam M. Zawbaa, C. Grosan, A.E. Hassanien, Binary grey wolf optimization approaches for feature selection, Neurocomputing, vol. 172, Elsevier, 2016, pp. 371381, January.

[36]Quanquan Gu, L. Zhenhui, H. Jiawei, Generalized fisher score for feature selection, in: Proceedings of the 27th Conference on Uncertainty in Artificial Intelligence (UAI), Barcelona, Spain, 2011.

[37]S. Mirjalili, S.M. Mirjalili, X.S. Yang, Binary bat algorithm, Neural Comput. Appl. 25 (3–4) (2014) 663681.

[38]A.F. Ali, M.A. Tawhid, Hybrid simulated annealing and pattern search method for solving minimax and integer programming problems, Pacific J. Optim. 12 (1) (2016) 151184.

[39]A.F. Ali, M.A. Tawhid, Hybrid particle swarm optimization with a modified arithmetical crossover for solving unconstrained optimization problems, INFOR: Inf. Syst. Operat. Res. 53 (3) (2015) 125141.

[40]M.A. Tawhid, A.F. Ali, Multi-directional bat algorithm for solving unconstrained optimization problems, OPSEARCH 54 (4) (2017) 684705.

[41]M.A. Tawhid, A.F. Ali, A hybrid social spider optimization and genetic algorithm for minimizing molecular potential energy function, Soft Comput. 21 (21) (2017) 64996514.

Acknowledgements

We are grateful to the anonymous 4 reviewers for constructive feedback and insightful suggestions which greatly improved this article. This research was supported partially by Mitacs Canada. The research of the 1st author is supported in part by the Natural Sciences and Engineering Research Council of Canada (NSERC). Publishers note: The publisher wishes to inform readers that the article “Hybrid Binary Bat Enhanced Particle Swarm Optimization Algorithm for solving feature selection problems” was originally published by the previous publisher of Applied Computing and Informatics and the pagination of this article has been subsequently changed. There has been no change to the content of the article. This change was necessary for the journal to transition from the previous publisher to the new one. The publisher sincerely apologises for any inconvenience caused. To access and cite this article, please use Tawhid, M.A., Dsouza, K.B. (2020), “Hybrid Binary Bat Enhanced Particle Swarm Optimization Algorithm for solving feature selection problems”, Applied Computing and Informatics. Vol. 16 No. 1/2, pp. 117-136. The original publication date for this paper was 11/04/2018.

Corresponding author

Mohamed A. Tawhid can be contacted at: mtawhid@tru.ca

Related articles