Professional Documents
Culture Documents
=
0 0 0
0 1 1
0 0 0
,
0 1 0
0 1 0
0 0 0
,
0 0 0
1 1 0
0 0 0
,
0 0 0
0 1 0
0 1 0
3 2 1 0
B B B B
.
X
w was empirically set to 0.25. Eq.(1) says that the
object is iteratively shrinking as i increments until
X area contracts to A w
X
portion of the rectangle size
A . This technique does not preserve object
connectivity, which is not critical for the classification
method.
The OTR somewhat reminds on a skeleton. It is
sensitive to a certain degree to boundary profile and
tends to produce non-desirably spurious shapes. An
algorithm [3] with proper setting of parameters to
inhibit small but spiky boundary behavior could be
tried as an alternative.
3.2. Characteristic Background Spots (CBS)
Sometimes, a clip submitted for the recognition
represents homogeneous region mostly or fully
populated by approximately uniform gray levels. The
latter clip obviously needs to be rejected. A single
OTR vector would not be sufficient to do so. This is
the reason we introduce another feature vector in
addition to the OTR. The other potentially problematic
situation, that the OTR intends to resolve, is a case
when a character in question represents a subset of the
template, e.g. 3 in question is matched to the
template of 8.
(a) (b) (c)
Figure 2. (a) An original character; (b) its OTR
template; (c) CBS template.
The CBS describes a shape of character image
rectangle occupied by the background. The CBS is
obtained via shrinking the character background by a
number of iterations (a natural way of implementing
that is the morphologic erosion). The shrinking
continues until a certain percentage of the background
area remains unpopulated by the object pixels. A
numbers of iterations required to get CBS depends on
the initial
X
K . Formally, CBS, denoted as
) ( X S
, is
expressed as:
A w K B X X S
X X
i
< = while ; ) (
4 mod
(2)
where ,...] 1 , 0 [ i .
X
w was set to 0.2. Eq.(2) says that
the background iteratively contracts as i increments
until the background area
X
K contracts to A w
X
portion of the rectangle A .
Both OTR and CBS representations are invariant to
a large degree to the original character thickness.
4. The Segmenter
The segmenter aims at separating the plate clip into
isolated character bounding boxes. Zero overlapping is
assumed. The algorithm is based on the adaptive
iterative thresholding and connected component
analysis.
Due to the non-homogeneous illumination and/or
presence of stamps, bolts, etc noise on the plate,
characters might occupy more than one connected
component. Most frequently, the components are
situated one above the other. Thus, we prepare an
extended bounding box of the component by
Proceedings of the 17th International Conference on Pattern Recognition (ICPR04)
1051-4651/04 $ 20.00 IEEE
spanning up to the plate bottom. Separated component
resided below might belong to noise see Figure 4 f.
To solve the problem of disconnectedness, the
character classification algorithm, see Section 5, works
on both the original and the extended connected
components, see e.g. Figure 4c and Figure 4d, and
accepts the one with higher confidence level.
Plate Character Isolation Algorithm Outline (Figure 3)
assess global threshold
g
T [9] and binarize the image
do while (right plate boundary not reached)
begin
find connected components // on the rest of the plate
if (one character) // based on width and proportion
begin
perform classification
jump to the next component
reassess global threshold
g
T //on the rest
end
else // two or more characters
reduce the
g
T by T // reduces connectivity
end
Note: Extra boundary peeling might be required to
eliminate some margins around the plate at the
preliminary stage.
Gray-scale clip 90 =
g
T
90 =
g
T
90 =
g
T
90 =
g
T
87 =
g
T
90 =
g
T
81 =
g
T
90 =
g
T
87 =
g
T 90 =
g
T
Figure 3. Sequential stages of plate segmentation as a
function of adaptive local thresholding ( 3 = T ) (plate
colors are inverted for better visibility)
(a)
(b)
(c)
(d)
(e)
(f)
Figure 4. Non-connected components belonging to same
character: (a),(b) source plates; (c),(e) connected
components found; (d) extension led to proper character
aggregation; (f) extension led to introducing extra noise
As a result of segmentation each successfully
isolated and approved characters bounding box is cut
out from the gray-scale plate clip and presented to the
classification algorithm described below.
5. The Classifier
An unknown character segmented by the algorithm,
described in Section 4, is normalized to a predefined
size. Thus, the unknown character and the templates
(OTR and CBS) are of same size and can be analyzed
element-wise.
Each of the two templates serves as a mask for
extracting the gray-scale pixels from the zones in
which the template hits the unknown character
matrix; see Figure 5. In the case of matching the
unknown character against its intra-class feature
vector, it is reasonable to anticipate that the OTR
would mask primarily the pixels belonging to the
character, and CBS mostly the background ones.
Oppositely, when the unknown character matrix is
matched to other class vectors, the OTR would hit
many of background pixels and CBS many of the
object pixels.
To estimate the similarity quantitatively we
introduce the following distance measure
k
D between
an unknown character gray level matrix G and a
template
k
C
T , where
k
C is a class label and
N k ,..., 0 = number of classes i.e. allowed characters in
license plate alphabet.
), ( ) (
CBS
CBS
OTR
OTR k
k
C
k
C
T G Var w T G Var w D + = (3)
where Var is the variance;
OTR
w and
CBS
w are weight
coefficients set to 1.0 and 0.9, respectively.
k
D is
expected to be minimal for the unknown characters
native class
k
C , i.e.
) ( min arg
k k k
D C = (4)
Plates, having high
k
C , do not resemble text-like
information [7] and are thus rejected.
(a)
(b) (c) (d)
(e) (f)
Figure 5. (a) License plate gray-scale matrix; (b)
character after segmentation; (c) characters bounding
box clipped from (a); (d) same, after size normalizati on;
(e) its masking by one of OTR templates of class 9; (f)
its masking by a corresponding CBS template
Compared to the OCR carried out on purely black-and-
white image matrix of normalized size, the gray-scale
recognition gives a boost of 10-12%.
The above-described classification procedure is
Proceedings of the 17th International Conference on Pattern Recognition (ICPR04)
1051-4651/04 $ 20.00 IEEE
fully generic, not limited to the license plate context. It
could serve recognition of any printed texts with noisy
environment.
6. The Postprocessor
The postprocessor is responsible for refining
recognition results by making use of the specific
license plate context.
A regular mode of classification extensively used at
the segmentation phase is the so-called international,
i.e. when templates of all countries of origin are mixed
within the same pool. Although the algorithm is, to a
large degree, independent on the template font used,
more precise classification results are obtained when
native country templates are applied in matching.
Thus, the postprocessor automatically determines the
country locale of the current plate. The system has
been tested on plates from Israel and Bulgaria, and its
architecture is open for further geographic extension.
The benefit of applying a proper country locale is
also due to a difference in country alphabets. Both
contain digits; the Bulgarian alphabet incorporates
several capital letters. Implementation-wise the
majority of characters recognized in the international
mode plus some grammar-inspired clues, determine the
country of origin. The additional semantic-based
refinement pass of classification picks up only the
native country templates, which allows improving the
read-rate furthermore.
(a)
(b)
(c)
(d)
Figure 6. (a),(b) Bulgarian plate characters; (c),(b)
same characters from Israeli plates
The semantic rules vary from country to country. In
Bulgaria, the plates start with a letter or two, followed
by a space, several digits and a couple of letters, see
e.g. Figure 1. The grammar model could be
schematically designated as L[L] DDDDLL, where
L and D stand for a letter and a digit, respectively.
Israeli plates strictly follow the rule: DD-DDD-DD.
On the test sample of 500 plates, 81.2% of plate
candidates have been approved at the localization
stage. The percentage of wrongly approved plates
(false positives) was 0%, as the rest to 100% plates
have been rejected. Character level read-rate versus the
misread was 85.2% / 3.4%.
7. Discussion and Conclusions
The above-described system achieves several
important goals:
- Allows a very reliable verification of a plate
candidate generated at the phase of localization
- Adaptively segments the plate image coping with
tough illumination conditions and image
distortions
- Classifies gray-scale characters of variable size
and resolution with a very reasonable accuracy.
As mentioned in Section 1, OCR packages are
incapable of delivering a magic-wand solution,
however, they could be useful when the segmentation
task is completed, and characters are properly
separated and clipped. Then, a relatively high read rate
could be anticipated. At least on most of the readable
by a human eye plates, we could expect read rate levels
of 90% and higher. The experiments are on going.
Further directions of the research lay in applying
approaches known in a context of conventional
OCR/ICR systems as multi-expert combination, or
voting [4]. Using RGB cameras with known in
advance plate background/foreground colors would
allow higher precision in character isolation. Applying
the algorithm to multiple frames of the plate video
stream would let additional gain in accuracy [1].
8. References
[1] Y.T. Cui and Q. Huang: Extracting Characters of License
Plates from Video Sequences, Machine Vision and
Applications vol 10, 1998, pp. 308-320
[2] A.K.Jain, Fundamentals of Digital Image Processing,
Prentice Hall, Englewood Cliffs, NJ, 1989
[3] W-B. Goh and K-Y. Chan, Part-Based Shape
Recognition Using Gradient Vector Field Histograms,
CAIP03, LNCS 2756, 2003, pp. 402-409,
[4] Y.S. Huang and C.Y. Suen, Combination of multiple
experts for the recognition of unconstrained handwritten
numerals, IEEE PAMI-17, 1995, pp. 90-94
[5] L. Jilin, M. Hongqing, L. Peihong, A High Performance
License Plate Recognition System Based On The Web
Technique, IEEE Conf. On Intelligent Transportation
Systems, Oakland, CA, 2001, pp.14-18
[6] S-H. Lee, Y-S. Seok, E-J. Lee Multi-National Integrated
Car-License Plate Recognition System Using Geometrical
Feature and Hybrid Pattern Vector, Conf. Circuits/Systems,
Computers and Comm., Phuket, Thailand, 2002, pp. 1256-
1259
[7] R. Lienhart and W. Effelsberg, Automatic Text
Segmentation and Text Recognition for Video Indexing,
ACM/Springer Multimedia Systems, vol. 8, 2000 pp.69-81
[8] Y. Lu, Machine Printed Character Segmentation - An
Overview, Pattern Recognition, vol.28, 1995, pp.67-80
[9] N. Otsu, A Thresholding Selection Method from Gray
Level Histograms, IEEE Trans. Syst., Man, Cybern., vol.9,
1979, pp.62-66
[10] V. Shapiro, G. Gluhchev, S. Bonchev, V. Velichkov,
License Plate Localization, Automatics and Informatics03,
Sofia, 2003, pp. 41-43
[11] M. Shridhar, J.W.V. Miller, G.Houle, L. Bijnagte,
Recognition of License Plate Images: Issues and
Perspectives, ICDAR, Bangalore, India, 1999, pp. 17-20
Proceedings of the 17th International Conference on Pattern Recognition (ICPR04)
1051-4651/04 $ 20.00 IEEE