Professional Documents
Culture Documents
=
c
i
ik
e<)4+
[ ] 1 , 0
ik
e<)6+
The wor!ing of fu'' K$nearest 5eigh%or Algorithm .//0 is as follows&
1or each test sample
= repeat steps a)d
a+ Compute distance %etween test sample E and each of the training sample
%+ 1ind ! nearest neigh%ors for test sample E
c+ Compute mem%ership of test sample E for each class ci )i J / to num%er of class la%els+ i.e i(x)
using following e<uation )7+
=
k
j
m
j
m
j
k
j
ij
i
x x
x x
x
1
)
1
2
(
)
1
2
(
1
|| ||
1
|| ||
1
) (
e<)7+
where Ii> is the mem%ership of >
th
neigh%or of test sample x for each class ci and m is the
fu''ifier #alue usuall set to 4.
d+ The results of the fu'' classification for test sample x is specified using simple crisp partition,
where a test sample is assigned to the class of maEimum mem%ership.
/ndfor
There are %asicall three different techni<ues of mem%ership assignment Ii> used in e<uation 7 for the
training samples ./40. The first method, uses a crisp la%eling, and assigns each training sample complete
mem%ership of one in its !nown class and 'ero mem%ership in all other classes. The second techni<ue
wor!s onl on two class data sets. The procedure assigns mem%ership in its !nown class %ased on its
distance from the mean of the training sample class. The third method assigns mem%ership to the
training samples according to a K$nearest neigh%or rule using e<uation 8. The K$nearest neigh%ors to
each sample E )sa E %elonging to class ci+ are found, and then mem%ership in each class is assigned
according to the following e<uation&
i j if
k
i j if
k
else
j
n
n
x
j
j
= +
=
=
, 49 . 0 * ) ( 51 . 0
! , 49 . 0 * ) (
) (
)> J / to c+ e<)8+
where, #alue n> is the num%er of the neigh%ors %elonging to the class c> and k is the total num%er of
neigh%ors of test sample x.
III$ 3A F4R I>/&7IF@I&3 %I3&IFICA&7 F/A7<R/% A&> F/A7<R/ DI/3A7% F4R CRI%0 A&> F<??@
1&& CBA%%IFI/R
International Journal of Advance Foundation and Research in Computer (IJAFRC)
Volume 1, Issue 5, Ma !"1#$ I%%& !'#( ) #(5'
!E * + !"1#, IJAFRC All Ri,hts Reserved ---$i.afrc$or,
GA is a stochastic general search method, capa%le of effecti#el eEploring large search spaces. The %asic
techni<ues of the GAs, follow the Charles Darwin principle of Gsur#i#al of the fittestK. The reproduction,
crosso#er, and mutation operations are applied to parent chromosomes to generate the neEt generation
offspring. Authors ha#e used GA for optimi'ing the connection weights of feed forward neural networ!,
finding the significant features for different classifier using %oth filter and wrapper approach for #arious
classifiers and for finding the optimal centroids for !$means clustering and fu'' !$means clustering .7,
/6$/;0.
Authors ha#e applied Binar encoded GA for identifing the significant features. In Binar (ncoded GA,
the chromosome can ha#e either Ds or /s as gene #alue. The /s and Ds in the %inar encoded GA
represent that the feature is significant and not significant respecti#el. In the proposed method, with
%inar encoding GA, the length of the chromosome is e<ual to total num%er of features sa 1. 1or eEample
with dia%etic data set, total num%er of features 1 J B. The chromosome length is B. 2ith /DD//DD/
%inar encoded chromosomes, /st, 7th, 8th and Bth features are significant features and 4nd, 6rd, ;th and
=th features are not significant. The wor!ing of %inar encoded GA for finding the significant feature
su%set for K$5earest 5eigh%or is as follows&
i. initiali'e the chromosome population randoml using %inar encoding )each chromosome length
is e<ual to total num%er of features 1 for a gi#en dataset+
ii. Fepeat the steps a$d till terminating condition )maEimum num%er of generations+ is reached.
a. Appl KL5earest 5eigh%or using indi#idual chromosome representing the significant features
and find the classification accurac as fitness of the chromosome.
%. Select the chromosome resulting in highest classification accurac of K55 classifier as the
fittest chromosome and replace the low fit chromosome % highest fit chromosome
)reproduction+.
c. Select an two chromosomes randoml and appl one point crosso#er operation.
d. Appl mutation operation % randoml selecting an random chromosome and randoml
change the %it / to D and %it D to /.
The positions of %it / in the %est$fit chromosome are considered as significant attri%utes for %oth fu''
and crisp K$5earest 5eigh%or classifier.
In addition to significant feature selection using Binar encoded GA, authors ha#e applied Feal (ncoded
GA for finding feature weights for K55 classifier. 2ith the real encoded GA, chromosomes represent the
weights of the features in contrast to %inar encoded GA which is used to find the su%set of significant
features from original feature set. The %inar encoded GA can %e considered as special method for finding
the weights of features, where, 'ero weight is assigned to non significant features and weight one is
assigned to significant features. A Fepair algorithm is used with real (ncoded GA algorithms to guarantee
the feasi%ilit of chromosomes )i.e. sum of weights of all features must %e e<ual to one+. This is done %
finding the sum of all feature weights and di#iding each feature weight % total weight.
The functioning of real encoded GA for finding the feature weights for K$5earest 5eigh%or is as follows&
i. Initiali'e the chromosome population randoml using real encoding. )(ach chromosome length is
e<ual to total num%er of features 1+
ii. Fepeat the steps a$d till terminating condition )maEimum num%er of generations+ is reached.
International Journal of Advance Foundation and Research in Computer (IJAFRC)
Volume 1, Issue 5, Ma !"1#$ I%%& !'#( ) #(5'
!( * + !"1#, IJAFRC All Ri,hts Reserved ---$i.afrc$or,
a+ Appl weighted KL5earest 5eigh%or to indi#idual chromosome representing the feature
weights and find the classification accurac as fitness of chromosome.
%+ Select the chromosome resulting in highest classification accurac of weighted K55 as the
fittest chromosome and replace the low fit chromosome % highest fit chromosome.
c+ Select an two chromosomes randoml and appl one point crosso#er operation )call repair
algorithm if the sum of the weights of all feature eEceeds one+
d+ Appl mutation operation % randoml selecting an chromosome and alter an one of the
weights % multipling with a random real num%er. )call repair algorithm if the sum of the
weights of all features eEceeds one+
iii. The weights in the fittest chromosome are considered as features weights for %oth crisp and
fu'' K$5earest 5eigh%or classifier.
IV$ 0%4 F4R FI&>I&3 F/A7<R/% D/I3A7% F4R CRI%0 A&> F<??@ 1&& CBA%%IFI/R
9S: are population %ased search method, inspired % the social %eha#ior of a Moc! of migrating %irds. A
particle is analogous to a chromosome )population mem%er+ in GAs. In GA the neEt generation
chromosomes are generated using crosso#er, mutation and reproduction process using parent
chromosomes. As opposed to GAs, the e#olutionar process in the 9S: does not create new %irds from
parent ones, instead, each particle flies in the search space with a #elocit ad>usted % its own !nowledge&
9%est )local search+ and %est among the companionAs !nowledge& G%est )glo%al search+ .;0. Authors ha#e
applied 9S: to optimi'e the connection weights of feed forward neural networ! ./;0 and to find the !$
means centroids ./=0.
This paper proposes real encoded 9S: algorithm to find the weights of features for K55 classifier. The
dimension of each particle is e<ual to total num%er of features for a gi#en dataset. The wor!ing of real
encoded 9S: for finding the significant feature weights for K$5earest 5eigh%or is as follows&
i. Initiali'e the particle population randoml using real encoding )(ach particle length is e<ual to
total num%er of features 1+
ii. Fepeat the steps a$e following till terminating condition )MaEimum num%er of iterations+ is
reached.
a+ Appl weighted KL5earest 5eigh%or using indi#idual particle representing the
significant feature weights and find the classification accurac.
%+ 1ind local %est for each particle. )Best accurac of indi#idual particle+
c+ 1ind glo%al %est from the population of particles. )Best accurac among all particles+
d+ Compute new #elocit using local and glo%al %est for each particle using e<uation );+.
e+
#i>)t N /+ J w#i> )t+ N c/F/)p%esti> $ Ei> )t++ N c4F4)g%esti> $ Ei> )t++ e< );+
f+ Opdate each particle position using old position and new #elocit using e<uation )=+
Ei> )t N /+ J Ei> )t+ N #i> )t N /+ e< )=+
)Call repair algorithm )as mentioned is Section 6+ if the sum of weights of all features of particle
eEceeds /+
iii. The weights represented % the glo%al %est particle are considered as final features weights for
%oth crisp and fu'' K$5earest 5eigh%or classifier.
International Journal of Advance Foundation and Research in Computer (IJAFRC)
Volume 1, Issue 5, Ma !"1#$ I%%& !'#( ) #(5'
!F * + !"1#, IJAFRC All Ri,hts Reserved ---$i.afrc$or,
In e<uation ;, #i>)t N /+ is the new #elocit, w is the inertia weight, #i> )t+ is the old #elocit, c/ and c4 are
constants usuall set to 4, F/ and F4 are randoml generated num%ers, p%esti> is the particles local %est
and g%esti> is the population glo%al %est, Ei> is the old position of particle.
V$ /G0/RIM/&7AB R/%<B7%
(Eperiments ha#e %een carried out using se#en different datasets namel& Peart stat log, dia%etes, wine,
Indian li#er, #ehicle, iris and ionosphere a#ailed from OCI machine learning dataset. The data has %een
partitioned % means of holdout method with ;D$7D ratio as training set and test set. The data is
normali'ed using min$maE normali'ation method. To enhance the performance the K55 classifier, %inar
encoded GA has %een eEperimented to disco#er the significant features as eEplained in section III. In
addition real encoded GA and real encoded 9S: ha#e %een eEperimented for identifing the weights of
features for K55 classifier. The results of crisp K55 are compared with anticipated %inar encoded GA
identified significant features, and with weights identified % GA and 9S: #ersus weights identified %
information gain, gain ratio and Felief method as shown in 1igure /. (Eperiments with GA and ha#e %een
carried out % #aring the populationAs si'e and num%er of generations, in addition to changing the #alue
of K )/$/D+ for crisp K55 and fu'' K55 classifier.
1igure / illustrates, thatAs the classification accurac of crisp K55 is impro#ed % GA identified features
and GA and 9S: identified weights for all the se#en datasets. 1or heart stat log dataset and #ehicle
dataset, the crisp K55 performance was preeminent with weights identified % real encoded GA. 9S:
identified weights pro#ed to %e superlati#e for crisp K55 classifier for dia%etes and Indian li#er datasets.
1or iris dataset, the %inar encoded GA features pro#ed to top for crisp K55 classifier. 1or wine and
ionosphere datasets, %inar encoded GA features and real encoded GA and 9S: weights resulted in top
and same accurac of crisp K55 classifier.
1urther GA identified feature and GA and 9S: identified weights ha#e %een used to further enhance the
performance of fu'' K55 classifier. 1or fu'' K55 classifier, the mem%ership of training samples is
computed using two methods )a+ crisp assignment method )assign each training sample complete
mem%ership in its !nown class and 'ero mem%ership in all other classes+ and is named as 1/2K55 and
)%+ using e<uation 8 and is named as 142K55. The relati#e results of 1/2K55 classifier )with crisp
assignment method+ and 142K55 classifier )with assignment using e<uation 8+ using significant
features identified % GA and using weights identified % GA and 9S: #ersus weights identified %
information Gain, gain ratio and Felief method are shown in 1igure 4 and 1igure 6 respecti#el with
fu''ifier #alue m e<ual to 4. 2ith 1/2K55 classifier, weights identified % real encoded GA pro#ed to %e
%est for heart$statlog, wine, Indian li#er and iris datasets. 1or dia%etes dataset, 1/2K55 showed an
augmented accurac % order of 4$ 6 Q with weights identified % Gain ratio method when compared to
proposed method. 1or #ehicle and ionosphere datasets, weights gi#en information gain pro#ed to %e to
some eEtent %etter for 1/2K55 when compared to proposed method.
The classification accurac of 142K55 was paramount with real encoded GA identified weights for heart
statlog, Indian li#er, #ehicle and iris datasets when compared to other methods. Both %inar encoded GA
features and real encoded GA weights resulted in same classification accurac of 142K55 for wine and
ionosphere datasets. 1or dia%etes datasets, the %inar encoded GA pro#ided features resulted in %etter
accurac of 142K55 when compared to weights proposed % real encoded GA and 9S:. Powe#er, the
weights identified % information gain and Gain ratio resulted in slight impro#ed performance for
dia%etes dataset when compared to GA identified features. The results of 142K55 are found to enhance
when compared to 1/2K55.
International Journal of Advance Foundation and Research in Computer (IJAFRC)
Volume 1, Issue 5, Ma !"1#$ I%%& !'#( ) #(5'
'" * + !"1#, IJAFRC All Ri,hts Reserved ---$i.afrc$or,
In addition the results of proposed method for impro#ing the crisp and fu'' K55 classifier accurac %
using GA identified features, and 9S: and GA identified weights is compared with fi#e well !nown
classifiers )a#aila%le in 2(KA tool+ namel radial %asis networ! )FB1+, Support #ector machine )S*M+,
Decision tree )C7.8+, Baesian and 5aR#e Baes methods as shown in 1igure 7 L 1igure ;.
1igure 7 illustrates that, %oth %inar encoded GA features and real encoded GA and 9S: weights resulted
in same classification accurac of crisp K55 classifier for wine and ionosphere datasets. 1or iris dataset,
GA identified features for crisp K55 classifier resulted in top accurac for iris dataset when compared to
other classifiers. 9S: identified weights pro#ed to finest for %oth dia%etes and Indian li#er datasets and
GA identified weights pro#ed to superlati#e for Peart stat$statlog and #ehicle datasets for crisp K55
when compared to other classifiers.
1igure 8 depicts that, 9S: identified weights pro#ed %e most eEcellent for 1/2K55 for dia%etes, when
compared to other classifiers. S*M pro#ed to preeminent for Peart$statlog dataset for 1/2K55. 1urther
GA identified weights with 1/2K55 showed %etter performance when compared to other classifiers for
wine, Indian li#er, #ehicle and iris dataset. 1igure 8 illustrates that, %oth %inar encoded GA features and
real encoded GA and 9S: weights resulted in same classification accurac of 1/2K55 ionosphere
datasets. Powe#er, the GA identified features showed a decline in performance for %oth #ehicle and iris
datasets for 1/2K55 classifier.
1igure ; depicts that GA identified weights for 142K55 pro#ed %e %est for all the datasets eEcluding
dia%etes when compared with other classifiers. 1or Dia%etes dataset, Baesian classifier showed
negligi%le enhancement when compared to 142K55 with GA identified features. :n the other hand, the
GA identified features showed a decline in performance for iris datasets with 142K55 classifier. GA and
9S: identified weights pro#ed to #ital with 14 2K55 and resulted in same accurac for ionosphere
dataset, where as GA recogni'ed features and weights pro#ed to paramount with 142KK55 for wine
dataset when compared to other classifiers.
VI$ C4&CB<%I4&%
This paper pro>ected %inar encoded GA for identifing significant features and real encoded GA and 9S:
for finding feature weights for enhancing the performance of %oth crisp and fu'' K55 classifier. Among
the three proposed methods for impro#ing the performance of K55 classifier, there is no one single
method which is superlati#e for all the se#en datasets. :#erall the two e#olutionar methods namel GA
and 9S: ha#e enhanced the performance of K55 classifier when compared to outcome of well !nown
classifier li!e radial %asis function, support #ector machine, decision tree, Baesian and 5aR#e %aes
classifiers as well as with the feature weights identified % information gain, gain ratio and Felief method
for all the eEperimented se#en datasets. As part of future enhancement, authors would li!e to eEtend the
wor! on significant feature selection using %inar 9S: and %inar cuc!oo search algorithms. In addition
to 9S: and GA, the %asic #ersion and modified #ersion of Cuc!oo search algorithms identified feature
weights can %e applied for further impro#ing the performance of K55 classifier.
VII$ R/F/R/&C/%
./0 S. Pan, and M. Kam%er, Data Mining& Concepts and Techni<ues, San 1rancisco, Morgan Kauffmann
9u%lishers, )4DD/+.
International Journal of Advance Foundation and Research in Computer (IJAFRC)
Volume 1, Issue 5, Ma !"1#$ I%%& !'#( ) #(5'
'1 * + !"1#, IJAFRC All Ri,hts Reserved ---$i.afrc$or,
.40 Asha Gowda Karegowda, M.A. Saaram, A.S. Man>unath, G1eature Su%set Selection 9ro%lem using
2rapper Approach in Super#ised -earningH, .International Sournal on Computer
Applications)ISCA+ ,*ol. /, pp. /6$/=, 4D/D.
.60 S. Cost, Sal'%erg, A weighted nearest neigh%or algorithm for learning with sm%olic features,
Machine -earning, *ol. /D, 5o /, San./CC6, pp. 8=$=B.
.70 Asha Gowda Karegowda, M.A. Saaram, A.S. Man>unath, *ida T, Shama, GGenetic Algorithm %ased
Dimensionalit Feduction for Impro#ing 9erformance of !$means and fu'' !$means clustering&
A Case stud for Categori'ation of Medical datasetH, Gwalior, India, 9roceedings of Se#enth
International Conference on Bio$Inspired Computing& Theories and Applications )BIC$TA
4D/4+,Ad#ances in Intelligent Sstems and Computing, *ol.4D/, pp. /;C$/BD, 4D/6.
.80 D. Gold%erg, Genetic Algorithms in Search, :ptimi'ation and Machine learning, Addison 2esle,
)/CBC+.
.;0 F. (%erhart and S. Kenned, GA new optimi'er using particle swarm theorH, 9roceedings of the
SiEth International Smposium on Micro Machine and Puman Science, 5agoa, Sapan, pp.6C$76,
/CC8.
.=0 T.M. Co#er, 9.(. Part, G5earest 5eigh%or pattern classificationH, I((( Transactions on Information
Theor, *ol. IT/6, pp. 4/$4=, /C;=.
.B0 Asha Gowda Karegowda, M.A. Saaram, A.S. Man>unath, GCom%ining A!ai!eAs Information
Criterion )AIC+ and the Golden$Section Search Techni<ue to find :ptimal 5um%ers of 5earest
5eigh%ors, International G, Sournal of Computer Applications *ol. 4, pp. BD$;=, Ma 4D/D.
.C0 Isaac Triguero, Soa<uTn Derrac, Sal#ador GarcTa, 1rancisco Perrera,H A TaEonom and
(Eperimental Stud on 9rototpe Generation for 5earest 5eigh%or ClassificationH, I(((
Transactions on Sstems, Man, and C%ernetics, 9art C *ol. 74)/+, pp. B;$/DD, 4D/4.
./D0 Sal#ador Garcia, Soa<uin Derrac, Sose Famon Cano, 1rancisco Perrera, G9rototpe Selection for
5earest 5eigh%or Classification& TaEonom and (mpirical StudH, I((( Transactions on 9attern
Analsis and Machine Intelligence, *ol. 678)6+, pp. 7/=$768, 4D/4.
.//0 Asha Gowda Karegowda, Fa!esh Kumar Singh, M.A.Saaram, A.S .Man>unath, G Impro#ing
2eighted K$5earest 5eigh%or 1eature 9ro>ections 9erformance with Different 2eights
Assigning Methods G, International Conference on Computational Intelligence )ICCI 4D/D+
Decem%er C L //, 4D/D, Coim%atore, India.
./40 Sames M. Keller, Michael F Gra, Sames A Gi#ens SF, GA fu'' K$nearest neigh%or algorithmH, I(((
Transactions on Sstems, Man, and C%ernetics, *ol. SMC$/8, 5o. 7, pp. 8BD$8B8, /CB8.
./60 Asha Gowda Karegowda, M.A.Saaram, A.S. Man>unath, GApplication of Genetic Algorithm
:ptimi'ed 5eural 5etwor! connection weights for medical diagnosis of 9IMA Indian dia%etesH,
International Sournal of Soft Computing, *ol.4, 5o.4, pp. /8$44, Ma 4D//.
./70 Asha Gowda Karegowda, M.A.Saaram, A.S. Man>unath, G1eature su%set selection using cascaded
GA and C1S& a filter approach in super#ised learning&H, International Sournal on Computer
Applications, *ol. 46)4+, pp./$/D, 4D//.
International Journal of Advance Foundation and Research in Computer (IJAFRC)
Volume 1, Issue 5, Ma !"1#$ I%%& !'#( ) #(5'
'! * + !"1#, IJAFRC All Ri,hts Reserved ---$i.afrc$or,
./80 Asha Gowda Karegowda, Shama, *ida T, M.A.Saaram, A.S .Man>unath, G Impro#ing 9erformance
of K$Means Clustering B Initiali'ing Cluster Centers Osing Genetic Algorithm and (ntrop Based
1u'' Clustering for Categori'ation :f Dia%etic 9atientsH, MSFIT, Bangalore, 9roceedings of
International Conference on Ad#ances in Computing, Ad#ances in Intelligent Sstems and
Computing, *olume /=7, Sul 7$;, pp. BCC$CD7, 4D/4.
./;0 Asha Gowda Karegowda, M.A. Saaram, GSignificant 1eature Set Dri#en, :ptimi'ed 115 for
(nhanced ClassificationH, International Sournal of Computational Intelligence and Informatics
ISS5& 446/$D48B, *ol. 4, 5o 7, Mar 4D/6.
./=0 Asha Gowda Karegowda, Seema Kumari, 9article Swarm :ptimi'ation Algorithm Based !$means
and 1u'' c$means clustering, International Sournal of Ad#anced Fesearch in Computer Science
and Software (ngineering *ol. 6, Issue =, pp. 77B$78/, Sul 4D/6 .
67
70
73
76
79
82
85
88
91
94
97
100
Heart-statlog Diabetes Wine Indian Liver Vehicle Iris ionosphere
Datasets
C
l
a
s
s
i
f
i
c
a
t
i
o
n
A
c
c
u
r
a
c
y
KNN Binary GA-WKNN Real GA-WKNN
Real PSO-WKNN Information Gain-WKNN Gain Ratio-WKNN
Relief-WKNN
Fi,ure 1$ Comparative performance of crisp 1&& classifier usin, 2inar encoded 3A identified features
and real encoded 3A and 0%4 identified feature -ei,hts vs$ information ,ain, ,ain ratio and relief method
identified -ei,hts for different datasets
International Journal of Advance Foundation and Research in Computer (IJAFRC)
Volume 1, Issue 5, Ma !"1#$ I%%& !'#( ) #(5'
'' * + !"1#, IJAFRC All Ri,hts Reserved ---$i.afrc$or,
64
67
70
73
76
79
82
85
88
91
94
97
100
Heart-statlog Diabetes Wine Indian Liver Vehicle Iris Ionosphere
DataSets
C
l
a
s
s
i
f
i
c
a
t
i
o
n
A
c
c
u
r
a
c
y
F1KNN Binary GA-F1WKNN Real GA-F1WKNN
Real PSO-F1KNN Information Gain-F1WKNN Gain Ratio-F1KNN
Releif-F1WKNN
65
70
75
80
85
90
95
100
Heart-statlog Diabetes Wine Indian Liver Vehicle Iris Ionosphere
Datasets
C
l
a
s
s
i
f
i
c
a
t
i
o
n
a
c
c
u
r
a
c
y
F2KNN Binary GA-F2WKNN Real GA-F2WKNN
Real PSO-F2WKNN Information Gain-F2WKNN Gain Ratio-F2WKNN
Relief-F2WKNN
Fi,ure !$ Comparative performance of F1D1&& classifier (-ith crisp method for mem2ership
assi,nment) usin, 2inar encoded 3A identified features and real encoded 3A and 0%4 identified
feature -ei,hts vs$ information ,ain, ,ain ratio and relief method for different datasets
Fi,ure '$ Comparative performance of F!D1&& classifier (-ith e9uation 5 for mem2ership
assi,nment) usin, 2inar encoded 3A identified features and real encoded 3A and 0%4 identified
feature -ei,hts vs$ information ,ain, ,ain ratio and relief method for different datasets
International Journal of Advance Foundation and Research in Computer (IJAFRC)
Volume 1, Issue 5, Ma !"1#$ I%%& !'#( ) #(5'
'# * + !"1#, IJAFRC All Ri,hts Reserved ---$i.afrc$or,
5
50
55
60
65
70
75
80
85
90
95
100
Heart-statlog Diabetes Wine Indian Liver Vehicle Iris ionosphere
DATA SETS
C
l
a
s
s
i
f
i
c
a
t
i
o
n
A
c
c
u
r
a
c
y
KNN Binary GA-WKNN Real GA-WKNN Real PSO-WKNN RBF
SVM DT Baye Nai!eBaye
45
50
55
60
65
70
75
80
85
90
95
100
Heart-statlog Diabetes Wine Indian Liver Vehicle Iris Ionosphere
Datasets
C
l
a
s
s
i
f
i
c
a
t
i
o
n
A
c
c
u
r
a
c
y
F1KNN Binary GA-F1WKNN Real GA-F1WKNN
Real PSO-F1KNN RBF SVM
Decision Tree Bayesian NaiveBayes
Fi,ure 5$ Comparative performance of F1D1&& classifier (-ith crisp method for mem2ership
assi,nment) usin, 2inar encoded 3A identified features and real encoded 3A and 0%4 identified -ei,hts
vs %VM, R6F, >ecision 7ree, 6aesian and &a;ve 6aes classifiers for different datasets
Fi,ure #$ Comparative performance of Crisp 1&& classifier (-ith crisp method for mem2ership
assi,nment) usin, 2inar encoded 3A identified features and real encoded 3A and 0%4 identified
feature -ei,hts vs %VM, R6F, >ecision 7ree, 6aesian and &a;ve 6aes classifiers for different
datasets
International Journal of Advance Foundation and Research in Computer (IJAFRC)
Volume 1, Issue 5, Ma !"1#$ I%%& !'#( ) #(5'
'5 * + !"1#, IJAFRC All Ri,hts Reserved ---$i.afrc$or,
45
50
55
60
65
70
75
80
85
90
95
100
Heart-statlog Diabetes Wine Indian Liver Vehicle Iris Ionosphere
Datasets Datasets Datasets Datasets
C
l
a
s
s
i
f
i
c
a
t
i
o
n
A
c
c
u
r
a
c
y
F2KNN Binary GA-F2WKNN Real GA-F2WKNN
Real PSO-F2WKNN RBF SVM
Decision Tree Bayesian NaiveBayes
ss
Fi,ure C$ Comparative performance of F!D1&& classifier (-ith e9uation 5 for mem2ership assi,nment)
usin, 2inar encoded 3A identified features and real encoded 3A and 0%4 identified -ei,hts vs %VM, R6F,
>ecision 7ree, 6aesian and &a;ve 6aes classifiers for different datasets