Professional Documents
Culture Documents
Supervised Unsupervised
Basic concepts: Basic concepts:
Baysian decision rule Distance
(MPP, LR, Discri.) Agglomerative method
Parametric learning (ML, BL) k-means
Non-Parametric learning (kNN) Winner-take-all
NN (Perceptron, BP) Kohonen maps
x1 3
K-L Transform (1)
Describe vector x in terms of a set of
basis vectors bi.
d
x =å yi b i yi =b xT
i
i =1
d
Error: D x = å (yi - a i )b i
i =m +1
5
K-L Transform (3)
Use mean-square error to quantify the
error
ì d d ü
e (m )=E í å å (yi - a i )b i (y j - a j )b j ý
2 T
îi =m+1 j =m +1 þ
ì d d ü
=E í å å (yi - a i )(y j - a j )b i b j ý
T
îi =m +1 j =m +1 þ
d
{
= å E (yi - a i )
2
}
i =m +1
6
K-L Transform (4)
Find the optimal i to minimize 2
¶e 2
=- 2(E{yi }- a i ) =0
¶a i
a i =E{yi }
Therefore,
d the error is now equal to
e (m )= å E (yi - E {yi })
2
{ 2
}
i =m +1
d d
i =m +1
{ 2
}
= å E (bTi x - E {bTi x}) = å E {(bTi x - E {bTi x})(xT b i - E {xT b i })}
i =m +1
d d d
= å b E (x - E{x})(x - E{x}) b i = å b S x b i = å l i
{ }
T T T
i i
i =m +1 i =m +1 i =m +1 7
K-L Transform (5)
The optimal choice of basis vectors is the
eigenvectors of x
The expansion of a random vector in terms of the
eigenvectors of the covariance matrix is referred to
as the Karhunen-Loeve expansion, or the “K-L
expansion”
Without loss of generality, we will sort the
eigenvectors bi in terms of their eigenvalues. That
is 1 >= 2 >= … >= d. Then we refer to b1,
corresponding to 1, as the “major eigenvector”, or
“principal component”
8
Summary
Raw data covariance matrix
eigenvalue eigenvector principal
component
How to use error rate?