Professional Documents
Culture Documents
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
i, j = 1, 2, ..p
m=1
Proposed Estimator
An algorithm
A simulation study
Summary
i, j = 1, 2, ..p
m=1
Proposed Estimator
An algorithm
A simulation study
Summary
i, j = 1, 2, ..p
m=1
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
p
2
X
ai i () t ,
(1.1)
i=1
(1.2)
Proposed Estimator
An algorithm
A simulation study
Summary
p
2
X
ai i () t ,
(1.1)
i=1
(1.2)
Proposed Estimator
An algorithm
A simulation study
Summary
Well Conditioned
The estimator in (1.2) is well conditioned.
= min (SU(I + A)1 U T + t UA(I + A)1 U T )
min ()
Aii
>0
> t min
ip 1 +
for minip Aii > 0, > 0, t > 0.
Estimator (1.2) is not sparse.
Proposed Estimator
An algorithm
A simulation study
Summary
Well Conditioned
The estimator in (1.2) is well conditioned.
= min (SU(I + A)1 U T + t UA(I + A)1 U T )
min ()
Aii
>0
> t min
ip 1 +
for minip Aii > 0, > 0, t > 0.
Estimator (1.2) is not sparse.
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
(1.3)
Proposed Estimator
An algorithm
A simulation study
Summary
, = argmin
|| S||22 + k k1 +
p
X
i
ai {i () t}2 .
i=1
(2.1)
The estimator given in (2.1) shrinks the eigenvalues of sample
covariance matrix towards constant t by weight ai and a
sufficiently large value of will yields a sparse estimate of
covariance matrix.
Estimator given in (2.1) need not be positive definite.
Proposed Estimator
An algorithm
A simulation study
Summary
, = argmin
|| S||22 + k k1 +
p
X
i
ai {i () t}2 .
i=1
(2.1)
The estimator given in (2.1) shrinks the eigenvalues of sample
covariance matrix towards constant t by weight ai and a
sufficiently large value of will yields a sparse estimate of
covariance matrix.
Estimator given in (2.1) need not be positive definite.
Proposed Estimator
An algorithm
A simulation study
Summary
argmin
S,t,A,
|(,)R
1
h
|| S||2F + k k1 +
p
X
(2.2)
2
ai {i () t} .
i=1
where
r
n
log p
S,t,A,
ip 1 + Aii
1 + maxip Aii
Proposed Estimator
An algorithm
A simulation study
1
t min Aii /2
ip
1 + minip Aii
min (S)
+ min
ip 1 + maxip Aii
Summary
Proposed Estimator
An algorithm
A simulation study
1
t min Aii /2
ip
1 + minip Aii
min (S)
+ min
ip 1 + maxip Aii
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
min (D)
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
s log p
n
Proposed Estimator
An algorithm
A simulation study
Summary
argmin
p
n
o
X
2
kK k + kK k1 +
ai {i (K )t}2
,t,A,
K |(,)R
1a
i=1
(2.4)
where
,t,A,
R
1a
,t,A,
R
1a
is given by:
r
n
log p
, t > 0,
= (, ) :
n
o
1
Aii
min ()
t min
lambda/2 +
ip
1 + Aii
1 + maxip Aii
(2.5)
Proposed Estimator
An algorithm
A simulation study
Summary
(2.6)
0 kF pk
0 k. Therefore the rate of
Note that k
convergence in Frobenius norm of the correlation matrix based
estimator is the same as the estimator defined in (2.2).
Proposed Estimator
An algorithm
A simulation study
Remark:
0 k = OP log p/n . This rate
i) For s = O(log p), k
of operator norm convergence is same as that of banded
estimator proposed of Bickel and Levina (2008).
ii) Rothman (2012) proposed an estimator of covariance
matrix based on similar loss function. The choice of different
penalty function yields very different estimate. Moreover
although his estimator yields a positive definite covariance
matrix estimator but whether their algorithm gives optimal
solution to their optimization problem, is hard to justify.
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
An algorithm
Let f () be the objective function of (2.2). Then,
f () = ||
S||2F
+ k k1 +
p
X
ai {i () t}2 .
(3.1)
i=1
= f2 ().
where I is the identity matrix, C = I + UAU T and
B = S + t UAU T .
Proposed Estimator
An algorithm
A simulation study
Summary
An algorithm
n
ij = sign (BC 1 )ij max |(BC 1 )ij |
,0 .
2(1 + maxip Aii )
(3.2)
Proposed Estimator
An algorithm
A simulation study
Summary
An Algorithm
Choice of U:
Note that U is the matrix of eigenvectors of , which is unknown.
In practice, one can chose U as matrix of eigenvectors of
corresponding eigenvalue decomposition of S + I for some > 0
i.e. let S + I = U1 D1 U1T , then take U = U1 .
Choice of and :
For given value of > 0, we can find the value of satisfying:
n
< 2 (1 + min Aii )
ip
o
min (S)
+ 2 t min Aii ,
ip
1 + maxi Aii
Proposed Estimator
An algorithm
A simulation study
Computational time
We compare the computational time of the our algorithm with
Glasso and PDSCE [Rothman (2012)] algorithm.
Summary
Proposed Estimator
An algorithm
A simulation study
Computational time
We compare the computational timing of our algorithm to
some other existing algorithms graphical lasso [Friedman et
al.(2008)], PDSCE [Rothman (2011)].
The exact timing of these algorithm also depends upon the
implementation, platform etc. (we did our computations in
R 3.1 on a AMD 2.8GHz processor).
Unlike Glasso algorithm for which the computational time
depends upon how dense the true covariance matrix is, our
algorithm takes roughly same amount of time for both dense
and sparse covariance matrices.
Although the proposed method requires optimization over a
S,t,A, , our algorithm is
grid of values of (, ) R
1
computationally efficient and easily scalable to large scale
data analysis problems.
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Simulation
We generate random vectors from multivariate t-distribution
with mean vector zero and various structured covariance
matrices with degrees of freedom 5.
(i) Hub Graph: The rows/columns of 0 are partitioned into
J equally-sized disjoint groups: {V1 V2 , ..., VJ } =
{1, 2, ..., p}, each group is associated with a pivotal row k.
Let size |V1 | = s. We set 0i,j = 0j,i = for i Vk and
0i,j = 0j,i = 0 otherwise. In our experiment,
J = [p/s], k = 1, s + 1, 2s + 1, ..., and we always take
= 1/(s + 1) with J = 20.
(ii) Neighborhood Graph: We first uniformly sample
(y1 , y2 , ..., yn ) from a unit square. We then set
0i,j = 0j,i = with probability
1
( 2) exp(4kyi yj k2 ). The remaining entries of 0 are
set to be zero. We always take to be 0.245.
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
(4.1)
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Ledoit-Wolf
Glasoo
PDSCE
JPEN
50
100
200
19.5(7.083)
36.5(19.19)
21.9(3.558)
382.(23.08)
422.(54.06)
340.(48.35)
21.3(2.845)
22.9(7.590)
24.9(5.753)
19.5(7.072)
18.9(4.077)
18.3(4.663)
Ledoit-Wolf
Glasoo
PDSCE
JPEN
64.5(6.19)
92.9(11.3)
106(13.5)
470(49.8)
559(72.9)
463(59.7)
57.7(10.93)
89.4(12.65)
97.9(17.34)
50.6(4.8)
57.8(4.76)
61.9(4.93)
p=1000
50
100
200
Summary
Proposed Estimator
An algorithm
A simulation study
Ledoit-Wolf
0.47(0.004)
0.48(0.002)
0.34(0.002)
p=500
Glasoo
PDSCE
13.4(0.198)
0.48(0.003)
10.8(0.090)
0.48(0.002)
7.97(0.070)
0.34(0.001)
Ledoit-Wolf
Glasoo
PDSCE
JPEN
50
100
200
0.29(0.003)
0.27(0.002)
0.26(0.001)
18.2(0.2)
11.9(0.131)
9.21(0.179)
0.30(0.003)
0.27(0.002)
0.27(0.001)
0.28(0.003)
0.26(0.002)
0.25(0.001)
JPEN
0.47(0.004)
0.34(0.032)
0.28(0.005)
p=1000
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Simulation
The average relative error and their standard deviations are
given in table above. (Please refer the manuscript for detailed
simulations). The numbers in the bracket is the standard error
estimate of relative error.
The JPEN estimate of covariance matrix outperforms other
methods for all values of p and n and for all four types of
covariance matrices.
Among all the methods, the PDSC estimates are closer to
JPEN estimate in the terms of ARE.
The Ledoit-Wolf estimate performs good in terms of ARE but
the estimated covariance matrix is not sparse.
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Method
Logistic Regression
SVM
Naive Bayes
Graphical Lasso
Joint Penalty
p=50
21.0(0.84)
16.70(0.85)
13.3(0.75)
10.9(1.3)
9.9(0.98)
p=100
19.31(0.89)
16.76(0.97)
14.33(0.85)
9.4(0.89)
8.9(0.93)
p=200
21.5(0.85)
18.18(0.96)
14.63(0.75)
9.8(0.90)
8.2(0.81)
Proposed Estimator
An algorithm
A simulation study
Summary
A Joint penalty estimate of covariance matrix is proposed.
The estimator is both well-conditioned and sparse
simultaneously. The proposed approach allows one to take
advantage of any prior structure if known on the eigenvalues
of true covariance matrix.
The theoretical consistency of JPEN estimator is establish in
both Frobenius and Operator norm which guarantees
consistency for principal components, hence we expect that
PCA will be one of the most important applications of the
method.
The proposed algorithm is very fast, efficient and easily
scalable to large scale optimization problems.
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Proposed Estimator
An algorithm
A simulation study
Summary
Acknowledgment
I would like to express my deep gratitude to Professor Hira L. Koul
for his valuable and constructive suggestions during the planning
and development of this research work.
Proposed Estimator
An algorithm
A simulation study
Summary