Professional Documents
Culture Documents
: 978-1-4799-4949-6
This paper is organized as follows: Section II (i) All fireflies within a population are
gives a brief description about feature selection and unisex, so that one firefly will be attracted to other
classification related to feature selection. The fireflies irrespective of their sex;
concept of Firefly algorithm is explained in Section (ii) Attractiveness between fireflies is
III. Details of the proposed methods are given in proportional to their brightness, implying that for
Section IV. Computations and results are discussed any two flashing fireflies, the less bright one will
in Section V and Section VI concludes the paper. move towards the brighter one. Attractiveness and
brightness both decrease as the distance between
II. FEATURE SELECTION AND
fireflies increases. If there is no brighter firefly
CLASSIFICATION
within its visible vicinity, then a particular firefly
A. Feature Selection will move randomly;
(iii) The brightness of a firefly is
Feature Selection (FS) is a commonly
determined by the landscape of the objective
used pre processing step in data mining, especially function.
when dealing with high dimensional space of
features. The main objective is to choose a subset IV. OPTIMIZATION OF FEATURE
of features from the original set of features as a SELECTION USING FIREFLY
representation of the entire domain [1]. The process ALGORITHM (FA-FS)
of FS is extensive and it spreads throughout many
In FA-FS algorithm, FA searches through the
fields which includes text categorization, machine
feature space and generates possible feature subset
learning, pattern recognition, and signal processing.
combinations; each time the feature subsets are
Considering the entire features may slowdown the
generated, they are evaluated by the Prediction
learning process and may reduce the performance
Accuracy given by the classifier. The detailed
of the classifier because of redundant and irrelevant
description of FA-FS algorithm is as follows:
features. Thus it is essential to reduce the number
The number of fireflies is set equal to the
of features by selecting the most relevant features
number of features in the dataset and each firefly is
to represent a dataset. FS allows the reduction of
assigned a single feature from the dataset. A binary
feature space, which is crucial in reducing the training
time and improving the prediction accuracy. This is
string (of length equal to the number of features) is
achieved by removing irrelevant, redundant, and noisy assigned to each firefly to represent the selection of
features [3]. features. The support to classification extended by
each feature (i.e. Predictive Accuracy (PA)) is
B. Classification considered as the brightness ( ) of the firefly
A classifier works on the entire feature set of holding that particular feature. PA is the percentage
an application to perform the task of classification. of instances that have been correctly classified as
Features of the application domain affect the instances of their original category [22]. The firefly
performance of the classifier in different manners. with the highest PA is termed as the “BRIGHTEST
So, there are higher chances for the irrelevant and FIREFLY” (BF) of the iteration.
noisy features to degrade the performance of the When the search procedure begins, each firefly
classifier and hence they should be removed [1 and selects the feature pointed by the BF for
2]. Not only the irrelevant, the redundancies in the combination. This is because of the principle of the
features should also be removed for effective FA, “fireflies get attracted towards the brightest
classification [6]. firefly in its vicinity”. So the feature subset
combinations that may result will definitely have
III. FIREFLY ALGORITHM the feature with the highest accuracy. Thus the
It is a wonderful sight to view the flashing of principle of the FA refines the search concentrating
lights from the fireflies in the summer sky. There only on more promising features than the others.
are a number of species of fireflies and most of the The feature of the BF has only been considered
species have a sort of unique pattern in the flash and before forming the feature subset combinations,
behavior. The fundamental principle behind this the new value of needs to be computed by using
flashing behavior is either to attract other fireflies the equation (1) as follows:
either for mating or as a prey [19 and 20]. This
principle is exploited and variants of FA have been = + exp − ( − ) + (1)
proposed.
The FA employed in the proposed work where, is the solution pointed by the current
depends on the variation of light intensity and the firefly (Classification Accuracy) and is the
formulation of attractiveness [adopted from [21]]. It solution pointed by the BF. is the attractiveness
depends on three idealized rules and uses them in measure between 0 and 1. γ is the variation of
framing the algorithm. The rules are: attractiveness whose value is chosen between 0.1
and 10. Distance is set to 1. α is a randomization
parameter normally selected within the range [0,1] TABLE I DATASETS DESCRIPTION
and is a vector of random numbers drawn from
either a Gaussian or uniform (generally [-0.5,0.5]) Dataset Instances Features Classes
distribution [21]. Heart-C 303 14 2
If the newly found is greater than the Dermatology 366 34 6
previous value, the feature of the BF and the Hepatitis 155 19 2
present firefly are combined, passed on to the Lung Cancer 32 56 2
classifier for evaluation and the PA extended by the Pima Indian Diabetes 768 8 2
combined features is now set as the brightness of Iris 150 4 3
the firefly; otherwise nothing is done and the firefly Wisconsin 699 9 2
will just be holding the feature it was previously Lymphography 148 18 4
Diabetes 768 9 2
holding. Once, all the fireflies have completed one
Heart-Stalog 270 13 2
iteration, the value of BF is reset based on the
highest value of PA computed. The above
As the algorithm proceeds, the possible
procedure is repeated for a pre-determined number
feature subsets that are generated are passed on to
of times or until an optimal configuration of feature
the J48 algorithm for evaluation. After every
subset with a reasonably good accuracy is reached.
iteration, the optimal feature subsets are chosen and
As we have considered the number of fireflies
the firefly is made to point to them. When the
equivalent to the number of features, they are
algorithm comes to a halt, FA-FS algorithm would
distributed throughout the available space and
have yielded optimal feature subset along with
hence the FA has an extremely quicker
maximum PA.
convergence [21]. The steps of the proposed FA-FS
The results obtained from FA-FS
are summarized and given in Fig. 1.
algorithm are presented in comparison with ABC-
1.Cycle =1
FS [12] in Fig. 2 and Fig. 3. The comparison of size
2. Initialize FA parameters of the feature subset obtained for each of the dataset
3. Evaluate the fitness of each individual feature from both FA-FS and ABC-FS is given in Fig. 2. It
4. Repeat could be seen from Fig.2 that, for some datasets,
5. Construct solutions by the fireflies ( Xi ) ABC-FS holds good and for the others FA-FS holds
Select the BRIGHTEST FIREFLY good in substantially reducing the size of the
Assign feature subset configurations (binary bit feature set. The PA obtained for all the datasets
string) to each Firefly from both ABC-FS and FA-FS are compared in
Produce new feature subsets
Fig. 3. Except for Hepatitis and Lung Cancer
Pass the produced feature subset to the classifier
Evaluate the fitnes of the feature subset by datasets, FA-FS has performed better than ABC-FS
yielding more PA. Thus, the proposed FA-FS
computing the new value of Xi algorithm has promising behavior and has selected
Reset the value of Xi based on either consideration the features in a way so as to maximize the
of rejection of the BRIGHTEST FIREFLY classification accuracy.
6. Calculate the best feature subset of the cycle
7. Cycle = Cycle + 1
8. Until pre-determined number of cycles is reached 80
9. Employ the same searching procedure of fireflies to generate 70
the optimal feature subset configurations 60
Fig.1 Steps of FA-FS algorithm 50 No.of
40 Features ABC
V. EXPERIMENTAL RESULTS AND 30
DISCUSSION No.of
20
Features FA
10
The performance of the proposed FA-FS
0 Total
algorithm has been tested with 10 different UCI
Dermatol…
Heart…
Lung…
Features
Lymph
Wisconsin
Pima
Diabetes
Hepatitis
Iris
Sonar
Heart C
Audiology