Professional Documents
Culture Documents
where, I is domain ( set of inputs) where K is a kernel function. We used Radial Basis
Kernel Function (RBF) [10] here, which is denoted as
I= {D, E} follow:
D= { X, Y} K(xi, xj) = exp(-γ|| xi – xj ||2), γ>0
X= {xi | 1<=i<=n} The non-linearity alter the quadratic form, but the
dual objective function is still quadratic in α,
Y= {yi |1<=i<=n} N N N
min ( ) min 1
y y
i j K ( x i , x j ) i j i,
}
n 2
i.e. D= { (xi, yi) ∈ (X × Y) i 1 j 1 i 1
i 1 0 i C , i,
N
E (set of constants) = {C, e} y
i 1
i i 0.
O is co domain (set of output) Sequential minimal optimization algorithm solves
O = { opi | 1<=i<=n} above quadratic programming problem by repeatedly
finding two Lagrange multipliers that can be
The support vector machine computes a linear optimized with respect to each other.
classifier of the form,
Sequential minimal optimization (SMO) [5] is an
f(x) = WX + b algorithm for efficiently solving the optimization
Where, W is weight vector problem which arises during the training of support
vector machine. At every step, SMO chooses two
X is input vector Lagrange multipliers to jointly optimize, finds the
B is bias optimal values for these multipliers and updates the
SVM to reflect the new optimal values [15].
The separating hyperplane is the plane, f(x) = 0.
Functioning of SMO as in the below algorithm.
update-value is added as shown below: result in speeding up the computation process. They
have high tolerance of noisy data. The major
disadvantage of neural networks is that, they have
poor interpretability. Fully connected networks are
difficult to articulate. Whereas various empirical
studies of Bayesian classifier in comparison with
decision tree and neural network classifiers have
found out that, in theoretical way Bayesian classifiers
have minimum error rate in comparison to all other
wij(t+1) = wij(t) + Δwij(t) classifiers. However, in practice this is not always the
case, owing to inaccuracies in the assumptions made
However, there is one exception. If the for its use, such as class conditional independence
partial derivative changes sign that is the previous and the lack of available probability data.
step was too large and the minimum was missed, the
previous weight-update is reverted: 2. Pre-processed data
Δwij(t) = - Δwij(t-1),
The experiments are carried out on heart dataset
if using Sequential Minimal Optimization in Support
Vector Machine.
To avoid a double penalty of the update-value, set the
above update rule by putting below value in Δij. The experiments are carried out on heart dataset
using Sequential Minimal Optimization in Support
Vector Machine.
Heart disease is diagnosed with the help of some
The partial derivative of the total error is given by the complex pathological data. The heart disease dataset
following formula: used in this experiment is the Cleveland Heart
Disease database taken from UCI machine learning
dataset repository [23]. This database contains 14
attributes as below:
1. Age of patient, 2. Sex of patient, 3. Chest pain
This indicates that the weights are updated only after type, 4. Resting blood pressure, 5. Serum cholesterol,
the presentation of all of the training patterns. 6. Fasting blood sugar, 7. Resting ECG results, 8.
Resilient back-propagation (RPROP) Maximum heart rate achieved, 9. Exercise induced
training algorithm [21] was adopted to train the angina, 10. ST depression induced by exercise
proposed ANN model as mentioned previously. After relative to rest, 11. Slope of the peak exercise ST
the selection of network, the network has been segment, 12. number of major vessels colored by
trained using resilient backpropagation training flourosopy, 13. thal, 14. Diagnosis of heart disease.
scheme. The training parameters have been modified
several times as explained above until the optimum 3. Flow diagram of MDSS
performance has been achieved. Maximum number
The purpose of this proposed model is to diagnose
of iterations has been set to 25000 epochs.
the heart disease by classifying the dataset of heart
disease. This classification process is shown in
IV. PROPOSED MODEL OF MEDICAL Figure 1.
DECISION SUPPORT SYSTEM
1. Related work
Figure 1. Flow diagram of MDSS for heart disease Figure 2 of pie chart shows the performance of the
system with fifth sample from table1, which gives
100% accuracy.
V. RESULT AND DISCUSSION
In the proposed model, we used dataset having 297
2. Multiclass classification by using
total number of patient records. Large part of records
Artificial neural network
in the dataset is used for training and rest of them are
used for testing. The main difference between the Following analysis done with, Input Layer = 13,
dataset given as input to training and testing is that, Hidden Layer = 25 and Output Layer = 1. In this case
the input we are giving to training is the data with training is done till error becomes less than 0.010 or
correct diagnosis (14th field in the dataset) and epochs are less than 25000. Table2 and Figure 3
whereas the input data of testing doesn’t have the gives the result of ANN multiclass classification.
correct diagnosis purposely. The Diagnosis (14 th)
field refers to the presence or absence of heart Following Figure 3 of pie chart shows the
disease of that respective patient. It is integer valued performance of fifth sample from table 2 which gives
field, having value 1(absence of disease) or 65% accuracy.
-1(presence of disease). So that at the end of testing
process we can check the result in the output file
created after testing and verify the efficiency of the
proposed model in terms of accuracy.
In the proposed system two methods of classification
are provided, Support vector machine and Artificial
neural network. Performance of both the
classification techniques is compared in terms of time
needed for classification and accuracy of the system.
From the below analysis we can say that, Sequential
Testing
No. of Class Class Class Class
Class I Accuracy Time Accuracy-
samples II III IV V
(10 samples)
100 57 19 10 10 3 95 31.02 90
[14] Yiqiang Zhan, Dinggang Shen, “Design [24] Mediline Plus: Information related to Heart
efficient support vector machine for fast Diseases.
classification”, The journal of pattern http://www.nlm.gov/medlineplus/heartdiseas
recognition society 38, Pages 157-161,2004. es.html
[15] Peng Peng, Qian-Li Ma, Lei-Ming Hong,
“The Research of the Parallel SMO
algorithm for solving SVM”, Proceedings of
the Eighth International Conference on
Machine Learning and Cybernetics,
Baoding, pp:1271-1274, 12-15 July 2009.
[16] Chin-Jen Lin, “Asymptotic convergence of
an SMO algorithm without any
assumptions”, IEEE Transactions on Neural
Networks, vol.13,issue 1, pp:248-250, 2002.
[17] Baxter Tyson Smith, “Lagrange Multipliers
Tutorial in the Context of Support Vector
Machines”, Memorial University of
Newfoundland St. John’s, Newfoundland,
Canada.
[18] Ya-Zhou Liu; Hong-Xun Yao; Wen Gao;
De-bin Zhao, “Single sequential minimal
optimization: an improved SVMs training
algorithm”, Proceedings of 2005
International Conference on Machine
Learning and Cybernetics, 2005. Pages
4360- 4364.
[19] Kjeldsen, Tinne Hoff. "A contextualized
historical analysis of the Kuhn-Tucker
theorem in nonlinear programming: the
impact of World War II". Historia Math. 27
(2000), no. 4, 331–361.
[20] Deepti Vadicherla, Sheetal Sonawane,
“Decision support system for heart disease
based on sequential minimal optimization in
support vector machine” International
Journal of Engineering Sciences and
Emerging Technologies, Volume 4, Issue 2,
Pages: 19-26, 2013.
[21] Ian H. Witten, Eibe Frank, Data Mining,
Elsevier Inc. 2005.
[22] Jiawei Han and Micheline Kamber, Data
Mining: Concepts and Techniques, 2/e,
Morgan Kaufmann Publishers, Elsevier Inc.
2006.
[23] UCI Machine Learning Repository: Heart
Disease Data Set.
http://archive.ics.uci.edu/ml/datasets/Heart+
Disease