You are on page 1of 4

International Journal of Computer Trends and Technology (IJCTT) volume 4 Issue 6June 2013

ISSN: 2231-2803 http://www.ijcttjournal.org Page 1919



Quantitative Analysis Of Hrct Images Using Knn Classifier
Sinduja.J
1
, S.Prathiba
2

1.
PG Scholar, Bharath University,Chennai,
2
Assistant Professor, Bharath University


AbstractPulmonary Emphysema is a chronic obstructive
lung disease which is characterized by limitation of air
flow. Detection and Quantication of emphysema is
important, since it is the main cause of shortness of breath
and disability in Chronic Obstructive Pulmonary Disease.
Current standard measures like Relative Area method
rely on a single intensity threshold on individual pixels,
ignoring any interrelations between pixels. HRCT is a
sensitive method for diagnosing emphysema, assessing its
severity, and determining its subtype. In this project
texture based classication system is used. A study is
performed by choosing Local binary Pattern (LBP) as
texture feature and classification is performed using KNN
classifier. Classification accuracy and quantification shows
that KNN classifier performs better when LBP is used as
texture feature in determining Pulmonary Emphysema.

Keywords- Emphysema, Local binary patterns,
Quantitative computed tomography, Texture analysis,
Tissue classification.

1. INTRODUCTION

Chronic Obstructive Pulmonary Disease (COPD) is a
major cause of disability, and it's the fourth leading cause of
death in the United States. COPD develops slowly. Symptoms
often worsen over time and can limit the ability to do routine
activities. Severe COPD may prevent from doing even basic
activities like walking, cooking etc. Most of the time, COPD
is diagnosed in middle-aged or older people. The disease isn't
passed from person to person. COPD has no cure yet, and
doctors don't know how to reverse the damage to the airways
and lungs. However, treatments and lifestyle changes can help
feel better, stay more active, and slow the progress of the
disease.
In emphysema, the walls between many of the air sacs are
damaged, causing them to lose their shape and become floppy.
This damage also can destroy the walls of the air sacs, leading
to fewer and larger air sacs instead of many tiny ones. If this
happens, the amount of gas exchange in the lungs is reduced.
In chronic bronchitis, the lining of the airways is constantly
irritated and inflamed. This causes the lining to thicken. Lots
of thick mucus forms in the airways, making it hard to breathe.
Early and accurate diagnosis of emphysema is important for
smoking cessation advice, evaluating the natural history of the
disease, and disease phenoltyping as we improve our


understanding of disease processes and therapeutic interventions.
Generally, then diagnosis of emphysema is based on indirect
features, such as clinical examination, pulmonary function tests,
and subjective visual evaluation of computed tomography (CT)
scans.
Objective

The detection and diagnosis of emphysema mainly
depends on the extend of the disease. The main objective of
this project is to characterize emphysema quantitatively using
texture analysis and also to develop a trained classifier by
extracting the texture feature which efficiently characterizes
the pulmonary Emphysema by taking in to account of the inter
pixel relationship


2. METHODOLOGY

Block Diagram




Training Stage




Testing Stage









3. DESCRIPTION

a) HRCT image

HRCT produces a volume of data which can be
manipulated, through a process known as windowing, in order
to demonstrate various bodily structures based on their ability
to block the X-ray beam. Although historically the images
generated were in the axial or transverse plane, orthogonal to
HRCT
image
slices

Feature

Extraction

Feature
Histogram




KNN

Classifier

Quantitative
measurement

International Journal of Computer Trends and Technology (IJCTT) volume 4 Issue 6June 2013


ISSN: 2231-2803 http://www.ijcttjournal.org Page 1920

the long axis of the body, modern scanners allow this volume
of data to be reformatted in various planes or even as
volumetric (3D) representations of structures. Although most
common in medicine, HRCT is also used in other fields, such
as non destructive materials testing.
HRCT is performed using a conventional CT scanner.
However, imaging parameters are chosen so as to maximize
spatial resolution Depending on the suspected diagnosis, the
scan may be performed in both inspiration and expiration. As
HRCT's aim is to assess a generalized lung disease, the test is
conventionally performed by taking thin sections 1040 mm
apart. The result is a few images that should be representative
of the lungs in general, but that cover only approximately one
tenth of the lungs. Other miscellaneous conditions where
HRCT is useful include:
.Lymphangitis carcinomatosa
Fungal, or other atypical, infections
Chronic pulmonary vascular disease
Lymphangioleiomyomatosis
Sarcoidosis
Organ transplant patients, particularly lung, or heart-lung
transplant recipients, are at relatively high risk of developing
pulmonary complications of the long-term drug and
immunosuppressive treatment. The major pulmonary
complication is bronchiolitis obliterans, which may be a sign
of lung graft rejection. HRCT has better sensitivity for
bronchiolitis obliterans than conventional radiography. Some
transplant centers may arrange annual HRCT to screen for
this.
HRCT IMAGE


ADVANTAGES OVER TRADITIONAL
RADIOGRAPHY
There are several advantages that CT has over
traditional 2D medical radiography. First, CT completely
eliminates the superimposition of images of structures outside
the area of interest. Second, because of the inherent high-
contrast resolution of CT, differences between tissues that
differ in physical density by less than 1% can be distinguished.
Finally, data from a single CT imaging procedure consisting
of either multiple contiguous or one helical scan can be
viewed as images in the axial, coronal, or sagittal planes,
depending on the diagnostic task. This is referred to as
multiplanar reformatted imaging.

b) Feature Extraction

Feature extraction involves simplifying the amount of
resources required to describe a large set of data accurately.
When performing analysis of complex data one of the major
problems stems from the number of variables involved.
Analysis with a large number of variables generally requires a
large amount of memory and computation power or a
classification algorithm which over fits the training sample
and generalizes poorly to new samples. Feature extraction is a
general term for methods of constructing combinations of the
variables to get around these problems while still describing
the data with sufficient accuracy.
Texture is an important characteristic for the
analysis of many types of images. It can be seen in all images
from multispectral scanner images obtained from aircraft or
satellite platforms to microscopic images of cell cultures or
tissue samples. The texture discrimination techniques are, for
the most part, ad hoc. Texture analysis refers to a class of
mathematical procedures and models that characterize the
spatial variations within imagery as a means of extracting
information. Mathematical procedures to characterize texture
fall into four general categories, statistical, geometrical,
model-based methods and signal processing methods and
include Fourier transforms, convolution filters, co-occurrence
matrix, spatial autocorrelation, fractals, etc.
An image texture is described by the number and
types of its primitives and the spatial organization or layout of
its primitives. The spatial organization maybe random, may
have a pair wise dependence of one primitive on a
neighbouring primitive, or may have a dependence of n
primitives at a time.
PROPOSED METHOD: LBP
The local binary pattern (LBP) operator is defined as
a gray-scale invariant texture measure, derived from a general
definition of texture in a local neighbourhood. The most
important property of the LBP operator in real-world
applications is its invariance against monotonic gray level
changes. Another equally important is its computational
simplicity, which makes it possible to analyse images in
challenging real-time settings.
The LBP feature vector, in its simplest form, is created in
the following manner:
International Journal of Computer Trends and Technology (IJCTT) volume 4 Issue 6June 2013


ISSN: 2231-2803 http://www.ijcttjournal.org Page 1921

1. Divide the examined window to cells (e.g. 16x16
pixels for each cell).
2. For each pixel in a cell, compare the pixel to each of
its 8 neighbours (on its left-top, left-middle, left-
bottom, right-top, etc.). Follow the pixels along a
circle, i.e. clockwise or counter-clockwise.
3. Where the centre pixel's value is greater than the
neighbour, write "1". Otherwise, write "0". This
gives an 8-digit binary number (which is usually
converted to decimal for convenience).
5. Compute the histogram, over the cell, of the
frequency of each number occurring (i.e., each
combination of which pixels are smaller and which
are greater than the centre).
6. Optionally normalize the histogram.
7. Concatenate normalized histograms of all cells. This
gives the feature vector for the window.


Fig 3.3 Example for finding LBP
LBP can be obtained using the equation
1
0
( ; , ) ( ( ) ( ))2
P
P
P
p
LBP x R P H I x I x


(3.1)
where H is the Heaviside function, I is the image is the centre
pixel.



Xp are the P local samples taken at a radius R around x.

c) Feature Histogram
In statistics, a histogram is a graphical representation,
showing a visual impression of the distribution of
experimental data. It is an estimate of the probability
distribution of a continuous variable and was first introduced
by Karl Pearson. A histogram consists of tabular frequencies,
shown as adjacent rectangles, erected over discrete intervals
(bins), with an area equal to the frequency of the observations
in the interval. The height of a rectangle is also equal to the
frequency density of the interval, i.e., the frequency divided
by the width of the interval. The total area of the histogram is
equal to the number of data. A histogram may also be
normalized displaying relative frequencies. It then shows the
proportion of cases that fall into each of several categories,
with the total area equalling 1.In an image processing context,
the histogram of an image normally refers to a histogram of
the pixel intensity values. This histogram is a graph showing
the number of pixels in an image at each different intensity
value found in that image. Histograms are used to plot density
of data, and often for density estimation: estimating the
probability density function of the underlying variable. The
total area of a histogram used for probability density is always
normalized to 1. The histogram plots the number of pixels in
the image (vertical axis) with a particular brightness value
(horizontal axis).

d) Quantification

The identification and precise quantification of
pulmonary emphysema in life are an attractive proposition,
but are made virtually impossible by the pathological basis of
its definition. Pulmonary emphysema is a progressive lung
disease mostly affecting elderly people. Because emphysema
is defined as "abnormal permanent enlargement of the air
spaces distal to the terminal bronchioles, accompanied by
destruction of their walls and without obvious fibrosis"
characterization of the disease with morphological and/or
morph metrical parameters is a prerequisite to study the
pathogenesis. Although lung function tests are useful tools to
diagnose chronic obstructive pulmonary disease, it is difficult
to distinguish emphysema from other obstructive lung
diseases.
Prior to classication of the lung eld, the lung
parenchyma pixels are segmented in the HRCT slice, using a
combination of thresholding. Manual editing was needed
afterward in one third of the cases and required simple
outlining of a few of the larger airways. Each segmented lung
parenchyma pixel is classied by classifying the ROI centred
on the pixel.
It should be noted that pixels that are not part of the
lung segmentation are not classied, but they can still
contribute to the classication. For example, part of the
exterior of the lung is in the local neighbourhood when
classifying a pixel at the border of the lung. In this way, all
potentially relevant structural information is incorporated,
such as proximity to the border of the lung or to the large
vessels and airways. The pixel probabilities are fused to obtain
one measure for the complete lung eld that can be used for
emphysema quantication. There are several ways of doing
this, e.g., averaging, voting, or the maximum rule. The
considered quantitative measures for emphysema are the and
is given by
sin(2 / ), cos(2 / )
T
Xp R p P R p P X
International Journal of Computer Trends and Technology (IJCTT) volume 4 Issue 6June 2013


ISSN: 2231-2803 http://www.ijcttjournal.org Page 1922



Where |S |is the number of pixels in the segmentation. P is the
probability of the class W
i
and X
j
is the center pixel.

e) Classifier (KNN)
In pattern recognition, the k-nearest neighbours algorithm (k-
NN) is a method for classifying objects based on closest
training examples in the feature space. k-NN is a type of
instance-based learning, or lazy learning where the function is
only approximated locally and all computation is deferred
until classification. The k-nearest neighbour algorithm is
amongst the simplest of all machine learning algorithms: an
object is classified by a majority vote of its neighbours, with
the object being assigned to the class most common amongst
its k nearest neighbours (k is a positive integer, typically
small). If k =1, then the object is simply assigned to the class
of its nearest neighbour.
The same method can be used for regression, by
simply assigning the property value for the object to be the
average of the values of its k nearest neighbours. It can be
useful to weight the contributions of the neighbours, so that
the nearer neighbours contribute more to the average than the
more distant ones. (A common weighting scheme is to give
each neighbour a weight of 1/d, where d is the distance to the
neighbour. This scheme is a generalization of linear
interpolation.)
The neighbours are taken from a set of objects for
which the correct classification (or, in the case of regression,
the value of the property) is known. This can be thought of as
the training set for the algorithm, though no explicit training
step is required. The k-nearest neighbour algorithm is
sensitive to the local structure of the data.
Nearest neighbour rules in effect compute the
decision boundary in an implicit manner. It is also possible to
compute the decision boundary itself explicitly, and to do so
in an efficient manner so that the computational complexity is
a function of the boundary complexity. The k-NN algorithm
can also be adapted for use in estimating continuous variables.
One such implementation uses an inverse distance weighted
average of the k-nearest multivariate neighbours.
This algorithm functions as follows:
1. Compute Euclidean from target plot to those that
were sampled.
2. Order samples taking for account calculated
distances.
3. Choose heuristically optimal k nearest neighbour
based on RMSE done by cross validation technique.
4. Calculate an inverse distance weighted average with
the k-nearest multivariate neighbours.
KNN is a versatile algorithm and is used in a huge
number of fields.

Texture measures such as LBP and Gaussian can be used for
quantitative analysis of pulmonary emphysema in CT images
of the lung. The ROI classication using LBP showed good
classication performance, compared to Gaussian . Thus LBP
seems to perform better than Gaussian in finding the
quantitative value. Also KNN have a greater sensitivity to
emphysema.LBP analysis is a sensitive method for diagnosing
emphysema, assessing its severity, and determining its
subtype since both visual and quantitative HRCT assessment
are closely correlated with pathological extend of emphysema.

SCOPE FOR FURTHER STUDIES
LBP have shown promising results in various applications in
computer vision and can be successfully applied in other
medical image analysis tasks, e.g., in mammographic mass
detection and magnetic resonance image analysis of the brain.

5.REFERENCES

1. A. H. Mir, M. Hanmandlu, and S. N. Tandon, Texture analysis of
CT images,IEEE Eng. Med. Biol. Mag. ,vol. 14, no. 6, pp. 781
786,Nov./ Dec. 1995
2. A. Bharathi Dr.A.M.Natarajan, Minimal feature selection using
SVM based on anova J ournal of Theoretical and Applied Information
Technology-Mar2009
3. Ching Ming J immy Wang Mamatha Rudrapatna Arcot Sowmya
Lung Disease Detection Using Frequency Spectrum Analysis Amer. J.
Respir. Crit. Care Med.Apr. 2005
4.Daniel I. Morariu, Lucian N. Vintan, and Volker Tresp, Meta-
Classification using SVM Classifiers for sText Documents
International J ournal of Mathematical and Computer Sciences-J an 2005
5. G. Zhao and M. Pietikinen, Dynamic texture recognition using
local binary patterns with an application to facial expressions, IEEE
Trans. Pattern Anal. Mach . Intell., J un. 2007.
6. Lauge Srensen, Saher B. Shaker, and Marleen de Bruijne
Quantitative Analysis of Pulmonary Emphysema Using Local Binary
Patterns IEEE transaction on medical imaging, vol. 29, no. 2, February
2010
7. Lin Zhang, Lei Zhang, Zhenhua Guo, and David
Zhang, monogenic-lbp: a new approach for rotation
invariant texture classification, Proceedings of 2010
IEEE 17th International Conference on Image Processing
September 26-29, 2010, Hong Kong
8. L. Srensen, S. B. Shaker, and M. de Bruijne, Texture
Classication in lung CT using local binary patterns, D.
N. Metaxas, L. Axel, G. Fichtinger, and G. Szkely,
Eds. New York: Springer-Verlag, Sep. 2008
9. .Michael Fitzpatrick, Milan Sonka,Automatic
detection systemfor pulmonary emphysema on 3-D
Chest images". Proceedings of SPIE-mar,2008

You might also like