1.1K views

Original Title: Cholesky Decomposition

Uploaded by Karen Kole

- A note on efficient computation of the gradient
- matrics modul
- Midterm F15 Sols
- SM1 Wks3-8
- Parallel Computing
- poisson
- CholeSky
- Jan 2009
- Simultaneous Communication and Control
- Cap1+2_ENGLISH
- Generating Normally Distributed Random Numbers in Excel
- Inverse Matrix
- 9akk105151d0125 Prioritization Module
- Matrix
- Past Year Questions on Matrices
- GE2115_lp
- Dynamic Model Analysis - Advanced Matrix Methods and Unit-Root Econometrics Representation Theorems 2nd - Mario Faliva and Maria Grazia Zoia
- A Comparison of Combined Overlap Block Fuzzy Cognitive Maps (COBFCM) and Combined Overlap Block Neutrosophic Cognitive Map (COBNCM) in finding the hidden patterns and indeterminacies in Psychological Causal Models: Case Study of ADHD
- Iit Jee Syllabus
- ece220matlab

You are on page 1of 7

It is not too hard to simultaneously simulate (to model) random variables. In Excel, for example, we can use use =NORMSINV(RAND()) to create standard random normal variables. The RAND() function is a uniform distribution bounded by {0,1}. The NORMSINV() translates the random number into the zvalue that corresponds to the probability given by a cumulative distribution. For example, =NORMSINV(5%) returns -1.645 because 5% of the area under a normal curve lies to the left of - 1.645 standard deviations. But no realistic asset or portfolio contains only one risk factor. To model several risk factors, we could simply generate multiple random variables. Put more technically, the realistic modeling scenario is a multivariate distribution function that models multiple random variables. But the problem with this approach, if we just stop there, is that correlations are not included. What we really want to do is simulate random variables but in such a way that we capture or reflect the correlations between the variables. In short, we want random but correlated variables. The typical way to incorporate the correlation structure is by way of a Cholesky decomposition (or factorization). For FRM candidates, Jorion briefly touches on the Cholesky factorization in the 4th Edition FRM Handbook (pages 99 to 100); but, if you are not familiar with matrix math, this may not be a sufficient introduction. In the EditGrid spreadsheet below, I performed a Cholesky decomposition for a simple three-asset case. This can be viewed separately or opened into a new sheet, if you would like to edit yourself. Please note: the decomposition below is not the endgame. It is a step along the way. It produces, for us to use, a matrix that can be used to produce returns that are random but correlated. The sheet below has four small sections, each step is numbered in green. The lower triangle (LU) is the result of the Cholesky Decomposition. It is the thing we can use to simulate random variables, that itself is "informed" by our covariance matrix.

The covariance matrix. This contains the implied correlation structure; in fact, a covariance matrix can itself be decomposed into a correlation matrix and a volatility vector. The covariance matrix(R) will be decomposed into a lower-triangle matrix (L) and an uppertriangle matrix (U). Note they are mirrors of each other. Both have identical diagonals; their zero elements and nonzero elements are merely "flipped"

Given that R = LU, we can solve for all of the matrix elements: a,b,c (the diagonal) and x, y, z. Note that is by definition. That's what a Cholesky decomposition is, it is the solution the produces two triangular matrices whose product is the original (covariance) matrix. (Note, if you play with the variables it is possible to produce an error: the matrix must be 'positive definite' and so not all matrices can be decomposed this way).

4. Given the solution for the matrix elements, then I calculated the product of the triangle matrix to ensure the produce does equal the original covariance matrix (i.e., does LU = R?). Note, in Excel a single array formula can be used with = MMULT(); in EditGrid, it is just a set of MMULT() formulas.

Cholesky Matrix

Cholesky algorithm Cholesky factorization Cholesky matrix indefinite matrix negative definite matrix negative semidefinte matrix

If we think of matrices as multi-dimensional generalizations of numbers, we may draw useful analogies between numbers and matrices. Not least of these is an analogy between positive numbers and positive definite matrices. Just as we can take square roots of positive numbers, so can we take "square roots" of positive definite matrices.

Positive Definite Matrices

A symmetric matrix x is said to be: positive definite if > 0 for all row vectors b 0; positive semidefinite if 0 for all row vectors b; negative definite if < 0 for all row vectors b 0; negative semidefinite if 0 for all row vectors b; indefinite if none of the above hold. These definitions may seem abstruse, but they lead to an intuitively appealing result. A symmetric matrix x is: positive definite if all its eigenvalues are real and positive;

positive semidefinite if all its eigenvalues are real and nonnegative; negative definite if all its eigenvalues are real and negative; negative semidefinite if all its eigenvalues are real and nonpositive; indefinite if none of the above hold. It is useful to think of positive definite matrices as analogous to positive numbers and positive semidefinite matrices as analogous to nonnegative numbers. The essential difference between semidefinite matrices and their definite analogues is that the former can be singular whereas the latter cannot. This follows because a matrix is singular if and only if it has a 0 eigenvalue.

Matrix "Square Roots"

Nonnegative numbers have real square roots. Negative numbers do not. An analogous result holds for matrices. Any positive semidefinite matrix h can be factored in the form for some real square matrix k, which we may think of as a matrix square root of h. The matrix k is not unique, so multiple factorizations of a given matrix h are possible. This is analogous to the fact that square roots of positive numbers are not unique either. If h is nonsingular (positive definite), k will be nonsingular. If h is singular, k will be singular.

Cholesky Factorization

A particularly easy factorization to perform is one known as the Cholesky factorization. Any positive semidefinite matrix has a factorization of the form where g is a lower triangular matrix. Solving for g is straightforward. Suppose we wish to factor the positive definite matrix

[1 ]

[2 ]

Also by inspection, Since we already . Proceeding in this manner, we obtain a matrix g in 6 steps:

[3 ]

The above example illustrates a Cholesky algorithm, which generalizes for higher dimensional matrices. Our algorithm entails two types of calculations: 1. Calculating diagonal elements (steps 1, 4, and 6) entails taking a square root.

2. Calculating off-diagonal elements , i > j, (steps 2, 3, and 5) entails dividing some number by the last-calculated diagonal element. For a positive definite matrix h, all diagonal elements will be nonzero. Solving for each entails taking the square root of a nonnegative number. We may take either the positive or negative root. Standard practice is to take only positive roots. Defined in this manner, the Cholesky matrix of a positive definite matrix is unique. The same algorithm applies for singular positive semidefinite matrices h, but the result is not generally called a Cholesky matrix. This is just an issue of terminology. When the algorithm is applied to the singular h, at least one diagonal elements equals 0. If only the last diagonal element equals 0, we can obtain g as we did in our example. If some other diagonal element equals 0, off-diagonal element will be indeterminate. We can set such indeterminate values equal to any value within an interval [a, a], for some a 0. Consider the matrix

[4 ]

[5 ]

In the fifth step, we multiply the second row of g by the third column of g' to obtain

[6 ]

We already know

, so we have

[7 ] [8 ]

which provides us with no means of determining . It is indeterminate, so we set it equal to a variable x and proceed with the algorithm. We obtain

[9 ]

For the element to be real, we can set x equal to any value in the interval [3,3]. The interval of acceptable values for indeterminate components will vary, but it will always include 0. For this reason, it is standard practice to set all indeterminate values equal to 0. With this selection, we obtain

[10 ]

We can leave g in this form, or we can delete the second column, which contains only 0's. The resulting 3 2 matrix provides a valid factorization of h since

[11 ]

If a matrix h is not positive semidefinite, our Cholesky algorithm will, at some point, attempt to take a square root of a negative number and fail. Accordingly, the Cholesky algorithm is a means of testing if a matrix is positive semidefinite.

Computational Issues

In exact arithmetic, the Cholesky algorithm will run to completion with all diagonal elements > 0 if and only if the matrixh is positive definite. It will run to completion with all diagonal elements 0 and at least one diagonal element = 0 if and only if the matrix h is singular positive semidefinite. Things are more complicated if arithmetic is performed with rounding, as is done on a computer. Off-diagonal elements are obtained by dividing by diagonal elements. If a diagonal element is close to 0, any roundoff error may be magnified in such a division. For example, if a diagonal element should be .00000001, but roundoff error causes it to be calculated as . 00000002, division by this number will yield an off-diagonal element that is half of what it should be. An algorithm is said to be unstable if roundoff error can be magnified in this way or if it can cause the algorithm to fail. The Cholesky algorithm is unstable for singular positive semidefinite matrices h. It is also unstable for positive definite matrices h that have one or more eigenvalues close to 0.

- A note on efficient computation of the gradientUploaded byMuhammad Bilal
- matrics modulUploaded byFadzilah Yaya
- Midterm F15 SolsUploaded byDylan Ler
- SM1 Wks3-8Uploaded byErin Gallagher
- Parallel ComputingUploaded byFilipe Amaro
- poissonUploaded byriteshnainani
- CholeSkyUploaded bynavri_nalhad
- Jan 2009Uploaded byPrashant Fegade
- Simultaneous Communication and ControlUploaded byAshish Baniwal
- Cap1+2_ENGLISHUploaded byCecília Perdigão
- Generating Normally Distributed Random Numbers in ExcelUploaded bymuliawan_firdaus
- Inverse MatrixUploaded byΈνκινουαν Κόγκ Αδάμου
- 9akk105151d0125 Prioritization ModuleUploaded byWitness Wii Mujoro
- MatrixUploaded byEvanProAristo
- Past Year Questions on MatricesUploaded byPoh Peng Lew
- GE2115_lpUploaded bysrinisanjai
- Dynamic Model Analysis - Advanced Matrix Methods and Unit-Root Econometrics Representation Theorems 2nd - Mario Faliva and Maria Grazia ZoiaUploaded bykantscribd
- A Comparison of Combined Overlap Block Fuzzy Cognitive Maps (COBFCM) and Combined Overlap Block Neutrosophic Cognitive Map (COBNCM) in finding the hidden patterns and indeterminacies in Psychological Causal Models: Case Study of ADHDUploaded byAnonymous 0U9j6BLllB
- Iit Jee SyllabusUploaded byAnupamSaxena
- ece220matlabUploaded byKrish Prasad
- International Journal of Mathematics and Statistics Invention (IJMSI)Uploaded byinventionjournals
- 07052705.pdfUploaded bysanket
- Linear Systems Ch4 [Compatibility Mode]Uploaded byUmair Zulfiqar
- Matlab TutorialUploaded bymuhammadnaumair
- Syllabus Ece 2007-08Uploaded byVenkatesh Desineni
- A Review of Watermarking Algorithms Fordigital ImageUploaded byVirgil Peptu
- List of Practicals - Comp Sci - XIIUploaded byNavneetRai
- 62a56Practice Questions GeneralUploaded bysidd_barm
- computer science practical file printedUploaded byHrishank Prabhakar
- 01262622Uploaded byAditya Gupta

- class XII Assignment - MatricesUploaded byCRPF School
- 1966 - Potter - Matrix Quadratic SolutionsUploaded byJames Brown
- MATEX Nth Power of a Square MatrixUploaded byjetski222
- MatricesUploaded byNik Norhafiza
- 08Chap4Uploaded byTamath
- Time varying systemsUploaded byManoj Ravi
- Exercise 5Uploaded byAbhinav Mishra
- MagicUploaded byanandakhapa
- Linear-Algebra.pdfUploaded byckvirtualize
- 4Uploaded bykaloibest
- Lecture Notes MatricesUploaded byRR886
- Matriks (matematik tingkatan 5)Uploaded byRohasniRoslan
- Singh Studentproblems Chapter1Uploaded byHelbert Agluba Paat
- matrices BEST.pdfUploaded byassefa
- Matrices.pdfUploaded byvarshithreddy
- Práctica 1_IntroRoboticsToolboxUploaded byfimmee2006
- Orthogonal MatricesUploaded byrgolfnut1
- math123Uploaded byShailendraPatel
- V2I1034Uploaded byeditor_ijarcsse
- beamerHIT112LectureSlides1.pdfUploaded byKeith Tanaka Magaka
- Ben Israel 1963Uploaded byAlexy Flemming
- A hadamarad TransformUploaded byAnup Dakre
- prove that eigen value of hermittian matrix are alays real.give example.use the above result to show that det (H-3I) cannot be zero.if h is hermittian matrix and I is unit matrixUploaded byAman Khera
- MIT18_06SCF11_Ses2.5sum.pdfUploaded bySarthi G
- LN3 Properties Eigenvalues&EigenvectorsUploaded byNikhilesh Prabhakar
- Hermitian matrix.pdfUploaded bymars
- Quiz 1---Submission or Return form (Rekod Penerimaan & Pemulangan Hasil Pelajar).docxUploaded byRen Jie
- chapter15Uploaded byHugo Libera Morales Quiros
- 1445009365Krash_Sample.pdfUploaded byJagan Rampalli
- inter maths 1AUploaded bySameer Salman