HOME

TheInfoList



OR:

In
numerical analysis Numerical analysis is the study of algorithms that use numerical approximation (as opposed to symbolic computation, symbolic manipulations) for the problems of mathematical analysis (as distinguished from discrete mathematics). It is the study of ...
and
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
, lower–upper (LU) decomposition or factorization factors a
matrix Matrix (: matrices or matrixes) or MATRIX may refer to: Science and mathematics * Matrix (mathematics), a rectangular array of numbers, symbols or expressions * Matrix (logic), part of a formula in prenex normal form * Matrix (biology), the m ...
as the product of a lower
triangular matrix In mathematics, a triangular matrix is a special kind of square matrix. A square matrix is called if all the entries ''above'' the main diagonal are zero. Similarly, a square matrix is called if all the entries ''below'' the main diagonal are z ...
and an upper triangular matrix (see
matrix multiplication In mathematics, specifically in linear algebra, matrix multiplication is a binary operation that produces a matrix (mathematics), matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the n ...
and matrix decomposition). The product sometimes includes a
permutation matrix In mathematics, particularly in matrix theory, a permutation matrix is a square binary matrix that has exactly one entry of 1 in each row and each column with all other entries 0. An permutation matrix can represent a permutation of elements. ...
as well. LU decomposition can be viewed as the matrix form of
Gaussian elimination In mathematics, Gaussian elimination, also known as row reduction, is an algorithm for solving systems of linear equations. It consists of a sequence of row-wise operations performed on the corresponding matrix of coefficients. This method can a ...
. Computers usually solve square
systems of linear equations In mathematics, a system of linear equations (or linear system) is a collection of two or more linear equations involving the same variables. For example, : \begin 3x+2y-z=1\\ 2x-2y+4z=-2\\ -x+\fracy-z=0 \end is a system of three equations in ...
using LU decomposition, and it is also a key step when inverting a matrix or computing the
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of a matrix. It is also sometimes referred to as LR decomposition (factors into left and right triangular matrices). The LU decomposition was introduced by the Polish astronomer Tadeusz Banachiewicz in 1938, who first wrote product equation LU=A=h^Tg (The last form in his alternate yet equivalent matrix notation appears as g\times h. )


Definitions

Let ''A'' be a square matrix. An LU factorization refers to expression of ''A'' into product of two factors – a lower triangular matrix ''L'' and an upper triangular matrix ''U'': A = LU. Sometimes factorization is impossible without prior reordering of ''A'' to prevent division by zero or uncontrolled growth of rounding errors hence alternative expression becomes: PAQ = LU , where in formal notation
permutation matrices In mathematics, particularly in Matrix (mathematics), matrix theory, a permutation matrix is a square binary matrix that has exactly one entry of 1 in each row and each column with all other entries 0. An permutation matrix can represent a permu ...
factors ''P'' and ''Q'' indicate permutation of rows (or columns) of ''A''. In theory ''P'' (or ''Q'') are obtained by permutations of rows (or columns) of the
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
, in practice the corresponding permutations are applied directly to rows (or columns) of ''A''. Matrix ''A'' of side n has n^2 coefficients while two triangle matrices combined contain n(n+1) coefficients, therefore n coefficients of matrices ''LU'' are not independent. Usual convention is to set ''L'' unitriangular, i.e. with all n main diagonal elements equal one. However, setting instead ''U'' matrix unitriangular reduces to the same procedure after
transpose In linear algebra, the transpose of a Matrix (mathematics), matrix is an operator which flips a matrix over its diagonal; that is, it switches the row and column indices of the matrix by producing another matrix, often denoted by (among other ...
of matrix product: B = A^T = (LU)^T =U^TL^T, (cf. properties of
transpose In linear algebra, the transpose of a Matrix (mathematics), matrix is an operator which flips a matrix over its diagonal; that is, it switches the row and column indices of the matrix by producing another matrix, often denoted by (among other ...
). Now U^T is lower triangle while L^T is upper unitriangular factor of ''B''. This demonstrates also, that operations on rows (e.g. pivoting) are equivalent to those on columns of a transposed matrix, and in general choice of row or column algorithm offers no advantage. In the lower triangular matrix all elements above the main diagonal are zero, in the upper triangular matrix, all the elements below the diagonal are zero. For example, for a 3 × 3 matrix ''A'', its LU decomposition looks like this: : \begin a_ & a_ & a_ \\ a_ & a_ & a_ \\ a_ & a_ & a_ \end = \begin \ell_ & 0 & 0 \\ \ell_ & \ell_ & 0 \\ \ell_ & \ell_ & \ell_ \end \begin u_ & u_ & u_ \\ 0 & u_ & u_ \\ 0 & 0 & u_ \end. Without a proper ordering or permutations in the matrix, the factorization may fail to materialize. For example, it is easy to verify (by expanding the
matrix multiplication In mathematics, specifically in linear algebra, matrix multiplication is a binary operation that produces a matrix (mathematics), matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the n ...
) that a_ = \ell_ u_. If a_ = 0, then at least one of \ell_ and u_ has to be zero, which implies that either ''L'' or ''U'' is
singular Singular may refer to: * Singular, the grammatical number that denotes a unit quantity, as opposed to the plural and other forms * Singular or sounder, a group of boar, see List of animal names * Singular (band), a Thai jazz pop duo *'' Singula ...
. This is impossible if ''A'' is nonsingular (invertible). In terms of operations, zeroing/elimination of remaining elements of first column of ''A'' involves division of a_, a_ with a_, impossible if it is 0. This is a procedural problem. It can be removed by simply reordering the rows of ''A'' so that the first element of the permuted matrix is nonzero. The same problem in subsequent factorization steps can be removed the same way. For numerical stability against rounding errors/division by small numbers it is important to select a_ of large absolute value (cf. pivoting).


LU Through recursion

The above example of 3\times 3 matrices demonstrates that matrix product of top row and leftmost columns of involved matrices plays special role for LU to succeed. Let us mark consecutive versions of matrices with (0),\;(1),\dots and then let us write matrix product A\equiv A^=L^U^ in such way that these rows and columns are separated from the rest. In doing so we shall use
block matrix In mathematics, a block matrix or a partitioned matrix is a matrix that is interpreted as having been broken into sections called blocks or submatrices. Intuitively, a matrix interpreted as a block matrix can be visualized as the original matrix w ...
notation, such that e.g. a\equiv a_ is an ordinary number, ^T\equiv(a_, a_)^T is a row vector and =(a_,a_) is a column vector and A' is sub-matrix of matrix A^ without top row and leftmost column. Then we can replace A^=L^U^ with a block
matrix product In mathematics, specifically in linear algebra, matrix multiplication is a binary operation that produces a matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the number of rows in the s ...
. Namely it turns out that one can multiply matrix blocks in such way as if they were ordinary numbers, i.e. row times column, except that now their components are sub-matrices, sometimes reduced to scalars or vectors. Thus u denotes a vector obtained from after multiplication of each component by a number u, ^T is an
outer product In linear algebra, the outer product of two coordinate vectors is the matrix whose entries are all products of an element in the first vector with an element in the second vector. If the two coordinate vectors have dimensions ''n'' and ''m'', the ...
of vectors , i.e. a matrix which first column is u_, next is u_ and so on for all components of and L^U^ is a product of sub-matrices of L^,\;U^ : \left( \begin a & & ^\textsf & \\ \hline & & & \\ & & A' & \\ & & & \end \right)\,=\, \left( \begin & & ^\textsf & \\ \hline & & & \\ & & L^ & \\ & & & \end \right)\; \left( \begin u & & ^\textsf & \\ \hline & & & \\ & & U^ & \\ & & & \end \right)\,= : \left( \begin u & & ^\textsf & \\ \hline & & & \\ u & & ^\textsf + L^U^ & \\ & & & \end \right) From equality of first and last matrices follow final u=a, \bf u=w, = while matrix A' becomes updated/replaced with A^\equiv L^U^=A'-^T. Now comes the crucial observation: ''nothing prevents us to treat A^ the same way as we did with A^'', and again, and again... If dimension of A is n\times n, after n-1 such steps all columns \bf v form sub-diagonal part of triangle matrix L and all pivots a combined with rows ^T form upper triangle matrix U, as required. In the above example n=3 so 2 steps suffice. The above procedure demonstrates that at no step the top diagonal pivot element a of consecutive sub-matrices can be zero. To avoid it columns or rows may be swapped so that a becomes nonzero. Such procedure involving permutation is called LUP, decomposition with pivoting. Permutation of columns corresponds to matrix product AQ^ where Q^ is a permutation matrix, i.e. the identity matrix I after the same column permutation. After all steps such LUP decomposition applies to AQ^\cdots Q^\equiv AQ=LU. Present computation scheme and similar in Cormen et al. are examples of ''recurrence algorithms''. They demonstrate two general properties of LU: (i) need for pivoting at each step and (ii) that final values of L,\;U matrices are obtained gradually, one row or column per step. Recurrence algorithms are not overly costly in terms of algebraic operations yet they suffer from practical disadvantage due to need to update and store most elements of A at each step. It will be seen that by reordering calculations it is possible to dispose with storage of intermediate values.


LU factorization with partial pivoting

It turns out that a proper permutation of rows (or columns) to select column (or row) absolute maximal pivot a_ is sufficient for numerically stable LU factorization, except for known pathological cases. It is called LU factorization with partial pivoting (LUP): PA = LU, (AQ=LU), where ''L'' and ''U'' are again lower and upper triangular matrices, and ''P'' and ''Q'' are corresponding
permutation matrices In mathematics, particularly in Matrix (mathematics), matrix theory, a permutation matrix is a square binary matrix that has exactly one entry of 1 in each row and each column with all other entries 0. An permutation matrix can represent a permu ...
, which, when left/right-multiplied to ''A'', reorder the rows/columns of ''A''. It turns out that all square matrices can be factorized in this form,. and the factorization is numerically stable in practice. This makes LUP decomposition a useful technique in practice. A variant called rook pivoting at each step involves search of maximum element the way rook moves on a chessboard, along column, row, column again and so on till reaching a pivot maximal in both its row and column. It can be proven that for large matrices of random elements its cost of operations at each step is similarly to partial pivoting proportional to the length of matrix side unlike its square for full pivoting.


LU factorization with full pivoting

An LU factorization with full pivoting involves both row and column permutations to find absolute maximum element in the whole submatrix: : PAQ = LU, where ''L'', ''U'' and ''P'' are defined as before, and ''Q'' is a permutation matrix that reorders the columns of ''A''.


Lower-diagonal-upper (LDU) decomposition

A Lower-diagonal-upper (LDU) decomposition is a decomposition of the form : A = LDU, where ''D'' is a
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagon ...
, and ''L'' and ''U'' are unitriangular matrices, meaning that all the entries on the diagonals of ''L'' and ''U'' are one.


Rectangular matrices

Above we required that ''A'' be a square matrix, but these decompositions can all be generalized to rectangular matrices as well., In that case, ''L'' and ''D'' are square matrices both of which have the same number of rows as ''A'', and ''U'' has exactly the same dimensions as ''A''. ''Upper triangular'' should be interpreted as having only zero entries below the main diagonal, which starts at the upper left corner. Similarly, the more precise term for ''U'' is that it is the
row echelon form In linear algebra, a matrix is in row echelon form if it can be obtained as the result of Gaussian elimination. Every matrix can be put in row echelon form by applying a sequence of elementary row operations. The term ''echelon'' comes from the F ...
of the matrix ''A''.


Example

We factor the following 2-by-2 matrix: : \begin 4 & 3 \\ 6 & 3 \end = \begin \ell_ & 0 \\ \ell_ & \ell_ \end \begin u_ & u_ \\ 0 & u_ \end. One way to find the LU decomposition of this simple matrix would be to simply solve the linear equations by inspection. Expanding the
matrix multiplication In mathematics, specifically in linear algebra, matrix multiplication is a binary operation that produces a matrix (mathematics), matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the n ...
gives :\begin \ell_ \cdot u_ + 0 \cdot 0 &= 4 \\ \ell_ \cdot u_ + 0 \cdot u_ &= 3 \\ \ell_ \cdot u_ + \ell_ \cdot 0 &= 6 \\ \ell_ \cdot u_ + \ell_ \cdot u_ &= 3. \end This system of equations is underdetermined. In this case any two nonzero elements of ''L'' and ''U'' matrices are parameters of the solution and can be set arbitrarily to any nonzero value. Therefore, to find the unique LU decomposition, it is necessary to put some restriction on ''L'' and ''U'' matrices. For example, we can conveniently require the lower triangular matrix ''L'' to be a unit triangular matrix, so that all the entries of its main diagonal are set to one. Then the system of equations has the following solution: :\begin \ell_ = \ell_ &= 1 \\ \ell_ &= 1.5 \\ u_ &= 4 \\ u_ &= 3 \\ u_ &= -1.5 \end Substituting these values into the LU decomposition above yields : \begin 4 & 3 \\ 6 & 3 \end = \begin 1 & 0 \\ 1.5 & 1 \end \begin 4 & 3 \\ 0 & -1.5 \end.


Existence and uniqueness


Square matrices

Any square matrix A admits ''LUP'' and ''PLU'' factorizations. If A is
invertible In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers. Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that ...
, then it admits an ''LU'' (or ''LDU'') factorization
if and only if In logic and related fields such as mathematics and philosophy, "if and only if" (often shortened as "iff") is paraphrased by the biconditional, a logical connective between statements. The biconditional is true in two cases, where either bo ...
all its leading principal minors are nonzero (for example \begin 0 & 1 \\ 1 & 0 \end does not admit an ''LU'' or ''LDU'' factorization). If A is a singular matrix of rank k , then it admits an ''LU'' factorization if the first k leading principal minors are nonzero, although the converse is not true. If a square, invertible matrix has an ''LDU'' (factorization with all diagonal entries of ''L'' and ''U'' equal to 1), then the factorization is unique. In that case, the ''LU'' factorization is also unique if we require that the diagonal of L (or U ) consists of ones. In general, any square matrix A_ could have one of the following: # a unique LU factorization (as mentioned above); # infinitely many LU factorizations if any of the first (''n''−1) columns are linearly dependent; # no LU factorization if the first (''n''−1) columns are linearly independent and at least one leading principal minor is zero. In Case 3, one can approximate an LU factorization by changing a diagonal entry a_ to a_ \pm \varepsilon to avoid a zero leading principal minor.


Symmetric positive-definite matrices

If ''A'' is a symmetric (or Hermitian, if ''A'' is complex)
positive-definite matrix In mathematics, a symmetric matrix M with real entries is positive-definite if the real number \mathbf^\mathsf M \mathbf is positive for every nonzero real column vector \mathbf, where \mathbf^\mathsf is the row vector transpose of \mathbf. Mo ...
, we can arrange matters so that ''U'' is the
conjugate transpose In mathematics, the conjugate transpose, also known as the Hermitian transpose, of an m \times n complex matrix \mathbf is an n \times m matrix obtained by transposing \mathbf and applying complex conjugation to each entry (the complex conjugate ...
of ''L''. That is, we can write ''A'' as : A = LL^*. \, This decomposition is called the Cholesky decomposition. If A is positive definite, then the Cholesky decomposition exists and is unique. Furthermore, computing the Cholesky decomposition is more efficient and numerically more stable than computing some other LU decompositions.


General matrices

For a (not necessarily invertible) matrix over any field, the exact necessary and sufficient conditions under which it has an LU factorization are known. The conditions are expressed in terms of the ranks of certain submatrices. The Gaussian elimination algorithm for obtaining LU decomposition has also been extended to this most general case.


Algorithms


Closed formula

When an LDU factorization exists and is unique, there is a closed (explicit) formula for the elements of ''L'', ''D'', and ''U'' in terms of ratios of determinants of certain submatrices of the original matrix ''A''. In particular, D_1 = A_, and for i = 2, \ldots, n, D_i is the ratio of the i-th principal submatrix to the (i - 1)-th principal submatrix. Computation of the determinants is computationally expensive, so this explicit formula is not used in practice.


Using Gaussian elimination

The following algorithm is essentially a modified form of
Gaussian elimination In mathematics, Gaussian elimination, also known as row reduction, is an algorithm for solving systems of linear equations. It consists of a sequence of row-wise operations performed on the corresponding matrix of coefficients. This method can a ...
. Computing an LU decomposition using this algorithm requires \tfrac n^3 floating-point operations, ignoring lower-order terms. Partial pivoting adds only a quadratic term; this is not the case for full pivoting.


Generalized explanation


= Notation

= Given an ''N'' × ''N'' matrix A = (a_)_, define A^ as the original, unmodified version of the matrix A. The parenthetical superscript (e.g., (0)) of the matrix A is the version of the matrix. The matrix A^ is the A matrix in which the elements below the main diagonal have already been eliminated to 0 through Gaussian elimination for the first n columns. Below is a matrix to observe to help us remember the notation (where each * represents any
real number In mathematics, a real number is a number that can be used to measure a continuous one- dimensional quantity such as a duration or temperature. Here, ''continuous'' means that pairs of values can have arbitrarily small differences. Every re ...
in the matrix): A^ = \begin * & & & \cdots & & & * \\ 0 & \ddots & & & & \\ & \ddots & * & & & \\ \vdots & & 0 & a_^ & & & \vdots \\ & & \vdots & a_^ & * \\ & & & \vdots & \vdots & \ddots \\ 0 & \cdots & 0 & a_^ & * & \cdots & * \end


= Procedure

= During this process, we gradually modify the matrix A using row operations until it becomes the matrix U in which all the elements below the main diagonal are equal to zero. During this, we will simultaneously create two separate matrices P and L, such that PA = LU. We define the final
permutation matrix In mathematics, particularly in matrix theory, a permutation matrix is a square binary matrix that has exactly one entry of 1 in each row and each column with all other entries 0. An permutation matrix can represent a permutation of elements. ...
P as the identity matrix which has all the same rows swapped in the same order as the A matrix while it transforms into the matrix U. For our matrix A^, we may start by swapping rows to provide the desired conditions for the n-th column. For example, we might swap rows to perform partial pivoting, or we might do it to set the pivot element a_ on the main diagonal to a nonzero number so that we can complete the Gaussian elimination. For our matrix A^, we want to set every element below a_^ to zero (where a_^ is the element in the n-th column of the main diagonal). We will denote each element below a_^ as a_^ (where i = n+1, \dotsc, N). To set a_^ to zero, we set row_i=row_i-(\ell_)\cdot row_n for each row i. For this operation, \ell_ := /. Once we have performed the row operations for the first N-1 columns, we have obtained an upper triangular matrix A^ which is denoted by U. We can also create the lower triangular matrix denoted as L, by directly inputting the previously calculated values of \ell_ via the formula below. :L = \begin 1 & 0 & \cdots & 0 \\ \ell_ & \ddots & \ddots & \vdots \\ \vdots & \ddots & \ddots & 0 \\ \ell_ & \cdots & \ell_ & 1 \end


Example

If we are given the matrixA = \begin 0 & 5 & \frac \\ 4 & 2 & 1 \\ 2 & 7 & 9 \\ \end,we will choose to implement partial pivoting and thus swap the first and second row so that our matrix A and the first iteration of our P matrix respectively becomeA^=\begin 4 & 2 & 1 \\ 0 & 5 & \frac \\ 2 & 7 & 9 \\ \end,\quad P^=\begin 0 & 1 & 0 \\ 1 & 0 & 0 \\ 0 & 0 & 1 \\ \end.Once we have swapped the rows, we can eliminate the elements below the main diagonal on the first column by performing \begin row_2=row_2-(\ell_)\cdot row_1 \\ row_3=row_3-(\ell_)\cdot row_1 \endsuch that,\begin \ell_= \frac=0 \\ \ell_= \frac=0.5 \endOnce these rows have been subtracted, we have derived from A^ the matrix A^= \begin 4 & 2 & 1 \\ 0 & 5 & \frac \\ 0 & 6 & 8.5 \\ \end.Because we are implementing partial pivoting, we swap the second and third rows of our derived matrix and the current version of our P matrix respectively to obtainA^=\begin 4 & 2 & 1 \\ 0 & 6 & 8.5 \\ 0 & 5 & \frac \\ \end, \quad P^=\begin 0 & 1 & 0 \\ 0 & 0 & 1 \\ 1 & 0 & 0 \\ \end.Now, we eliminate the elements below the main diagonal on the second column by performing row_3=row_3-(\ell_)\cdot row_2 such that \ell_= \frac . Because no nonzero elements exist below the main diagonal in our current iteration of A after this row subtraction, this row subtraction derives our final A matrix (denoted as U) and final P matrix:A^=A^=U=\begin 4 & 2 & 1 \\ 0 & 6 & 8.5 \\ 0 & 0 & 0.25 \\ \end, \quad P=\begin 0 & 1 & 0 \\ 0 & 0 & 1 \\ 1 & 0 & 0 \\ \end.After also switching the corresponding rows, we obtain our final L matrix:L = \begin 1 & 0 & 0 \\ \ell_ & 1 & 0 \\ \ell_ & \ell_ & 1 \\ \end = \begin 1 & 0 & 0 \\ 0.5 & 1 & 0 \\ 0 & \frac & 1 \\ \endNow these matrices have a relation such that PA=LU.


Relations when no rows are swapped

If we did not swap rows at all during this process, we can perform the row operations simultaneously for each column n by setting A^ := L^_n A^, where L^_n is the ''N'' × ''N'' identity matrix with its ''n''-th column replaced by the
transpose In linear algebra, the transpose of a Matrix (mathematics), matrix is an operator which flips a matrix over its diagonal; that is, it switches the row and column indices of the matrix by producing another matrix, often denoted by (among other ...
d vector \begin0 & \dotsm & 0 & 1 & -\ell_ & \dotsm & -\ell_ \end^\textsf. In other words, the lower triangular matrix :L^_n = \begin 1 & & & & & \\ & \ddots & & & & \\ & & 1 & & & \\ & & -\ell_ & & & \\ & & \vdots & & \ddots & \\ & & -\ell_ & & & 1 \end. Performing all the row operations for the first N-1 columns using the A^ := L^ _n A^ formula is equivalent to finding the decomposition A = L_1 L_1^ A^ = L_1 A^ = L_1 L_2 L_2^ A^ = L_1 L_2 A^ = \dotsm = L_1 \dotsm L_ A^. Denote L = L_1 \dotsm L_ so that A=LA^=LU. Now let's compute the sequence of L_1 \dotsm L_. We know that L_ has the following formula. :L_n = \begin 1 & & & & & \\ & \ddots & & & & \\ & & 1 & & & \\ & & \ell_ & & & \\ & & \vdots & & \ddots & \\ & & \ell_ & & & 1 \end If there are two lower triangular matrices with 1s in the main diagonal, and neither have a nonzero item below the main diagonal in the same column as the other, then we can include all nonzero items at their same location in the product of the two matrices. For example: \left(\begin 1 & 0 & 0 & 0 & 0 \\ 77 & 1 & 0 & 0 & 0 \\ 12 & 0 & 1 & 0 & 0 \\ 63 & 0 & 0 & 1 & 0 \\ 7 & 0 & 0 & 0 & 1 \end\right)\left(\begin 1 & 0 & 0 & 0 & 0 \\ 0 & 1 & 0 & 0 & 0 \\ 0 & 22 & 1 & 0 & 0 \\ 0 & 33 & 0 & 1 & 0 \\ 0 & 44 & 0 & 0 & 1 \end\right)=\left(\begin 1 & 0 & 0 & 0 & 0 \\ 77 & 1 & 0 & 0 & 0 \\ 12 & 22 & 1 & 0 & 0 \\ 63 & 33 & 0 & 1 & 0 \\ 7 & 44 & 0 & 0 & 1 \end\right) Finally, multiply L_ together and generate the fused matrix denoted as L (as previously mentioned). Using the matrix L, we obtain A = LU. It is clear that in order for this algorithm to work, one needs to have a_^ \neq 0 at each step (see the definition of \ell_). If this assumption fails at some point, one needs to interchange ''n''-th row with another row below it before continuing. This is why an LU decomposition in general looks like P^A = L U .


LU Banachiewicz decomposition

Although Banachiewicz (1938) ''LU'' decomposition algorithm preceded the advent of programmed electronic computers, it was ready made for direct implementation into code as index swapping, transpose and column by column multiplication remain native built capabilities of the most programming languages and are handled by compilers alone with little delay of actual execution. The peculiar matrix notation used by Banachiewicz enabled him to multiply matrices column by column, a convenient feature for mechanical calculations as he could reveal consecutive factors by sliding a ruler to next rows of matrices. For human readers however, his equations are best transformed into standard matrix notation. To obtain from a full matrix ''A'' triangle matrices ''U'' and ''L'' calculations start by copying top row and leftmost column of ''A'' respectively into corresponding positions of matrices ''U'' and ''L''. The known unit diagonal elements of ''L'' are not stored neither used throughout the whole process. Next calculations continue for the subsequent rows and columns till the bottom right corner of ''A''. The figure illustrates calculations for 3-rd row and column, assuming previous stages were already completed. Involved matrices are named above squares marking their content. Matrix products and subtractions are applied only to elements in the thick frame boxes. Green filled thin frame boxes indicate values already known, from previous stages. Blue boxes indicate places in ''U'' and ''L'' matrices for storing of results. Note that at each stage the result elements of ''L'' need to be divided by the corresponding pivot element on the main diagonal of ''U''. This applies to the leftmost column of ''L'' too. Note that after completion of 3-rd stage the involved elements of matrix ''A'' are no longer used and neither those from previous stages. This enables replacement of these elements with the result values of ''U'' and ''L'', i.e. execution of ''LU'' decomposition ''in place'', so that the whole ''A'' is replaced with ''U'' and ''L'' except for the unit diagonal of ''L''. Banachiewicz ''LU'' algorithm is well suited for partial pivoting by choosing the absolute maximum pivot from the newly calculated row of ''U'' and subsequently swapping its columns so that it lands on the main diagonal. More details can be figured out from inspection of the enclosed Fortran90 code. All partial pivoting ''LU'' algorithms cost roughly the same amount, of order \ operations, where n is number of rows or columns of ''A''.


LU Crout decomposition

Note that the decomposition obtained through this procedure is a ''Doolittle decomposition'': the main diagonal of ''L'' is composed solely of ''1''s. If one would proceed by removing elements ''above'' the main diagonal by adding multiples of the ''columns'' (instead of removing elements ''below'' the diagonal by adding multiples of the ''rows''), we would obtain a '' Crout decomposition'', where the main diagonal of ''U'' is of ''1''s. Another (equivalent) way of producing a Crout decomposition of a given matrix ''A'' is to obtain a Doolittle decomposition of the transpose of ''A''. Indeed, if A^\textsf = L_0 U_0 is the LU-decomposition obtained through the algorithm presented in this section, then by taking L = U_0^\textsf and U = L_0^\textsf, we have that A = LU is a Crout decomposition.


Randomized algorithm

It is possible to find a low rank approximation to an LU decomposition using a randomized algorithm. Given an input matrix A and a desired low rank k, the randomized LU returns permutation matrices P, Q and lower/upper trapezoidal matrices L, U of size m \times k and k \times n respectively, such that with high probability \left\, PAQ-LU \right\, _2 \le C\sigma_, where C is a constant that depends on the parameters of the algorithm and \sigma_ is the (k+1)-th singular value of the input matrix A.


Theoretical complexity

If two matrices of order ''n'' can be multiplied in time ''M''(''n''), where ''M''(''n'') ≥ ''n''''a'' for some ''a'' > 2, then an LU decomposition can be computed in time O(''M''(''n'')). This means, for example, that an O(''n''2.376) algorithm exists based on the Coppersmith–Winograd algorithm.


Sparse-matrix decomposition

Special algorithms have been developed for factorizing large
sparse matrices In numerical analysis and scientific computing, a sparse matrix or sparse array is a matrix (mathematics), matrix in which most of the elements are zero. There is no strict definition regarding the proportion of zero-value elements for a matrix ...
. These algorithms attempt to find sparse factors ''L'' and ''U''. Ideally, the cost of computation is determined by the number of nonzero entries, rather than by the size of the matrix. These algorithms use the freedom to exchange rows and columns to minimize fill-in (entries that change from an initial zero to a nonzero value during the execution of an algorithm). General treatment of orderings that minimize fill-in can be addressed using
graph theory In mathematics and computer science, graph theory is the study of ''graph (discrete mathematics), graphs'', which are mathematical structures used to model pairwise relations between objects. A graph in this context is made up of ''Vertex (graph ...
.


Applications


Solving linear equations

Given a system of linear equations in matrix form :A\mathbf x = \mathbf b, we want to solve the equation for x, given ''A'' and b. Suppose we have already obtained the LUP decomposition of ''A'' such that PA = LU, so LU \mathbf x = P \mathbf b. In this case the solution is done in two logical steps: # First, we solve the equation L \mathbf y = P \mathbf b for y. # Second, we solve the equation U \mathbf x = \mathbf y for x. In both cases we are dealing with triangular matrices (''L'' and ''U''), which can be solved directly by forward and backward substitution without using the
Gaussian elimination In mathematics, Gaussian elimination, also known as row reduction, is an algorithm for solving systems of linear equations. It consists of a sequence of row-wise operations performed on the corresponding matrix of coefficients. This method can a ...
process (however we do need this process or equivalent to compute the ''LU'' decomposition itself). The above procedure can be repeatedly applied to solve the equation multiple times for different b. In this case it is faster (and more convenient) to do an LU decomposition of the matrix ''A'' once and then solve the triangular matrices for the different b, rather than using Gaussian elimination each time. The matrices ''L'' and ''U'' could be thought to have "encoded" the Gaussian elimination process. The cost of solving a system of linear equations is approximately \frac n^3 floating-point operations if the matrix A has size n. This makes it twice as fast as algorithms based on QR decomposition, which costs about \frac n^3 floating-point operations when Householder reflections are used. For this reason, LU decomposition is usually preferred.


Inverting a matrix

When solving systems of equations, ''b'' is usually treated as a vector with a length equal to the height of matrix ''A''. In matrix inversion however, instead of vector ''b'', we have matrix ''B'', where ''B'' is an ''n''-by-''p'' matrix, so that we are trying to find a matrix ''X'' (also a ''n''-by-''p'' matrix): :AX = LUX = B. We can use the same algorithm presented earlier to solve for each column of matrix ''X''. Now suppose that ''B'' is the identity matrix of size ''n''. It would follow that the result ''X'' must be the inverse of ''A''.


Computing the determinant

Given the LUP decomposition A = P^ LU of a square matrix ''A'', the determinant of ''A'' can be computed straightforwardly as :\det(A) = \det\left(P^\right) \det(L) \det(U) = (-1)^S \left( \prod_^n l_ \right) \left( \prod_^n u_ \right) . The second equation follows from the fact that the determinant of a triangular matrix is simply the product of its diagonal entries, and that the determinant of a permutation matrix is equal to (−1)''S'' where ''S'' is the number of row exchanges in the decomposition. In the case of LU decomposition with full pivoting, \det(A) also equals the right-hand side of the above equation, if we let ''S'' be the total number of row and column exchanges. The same method readily applies to LU decomposition by setting ''P'' equal to the identity matrix.


History

The LU decomposition is related to elimination of linear systems of equations, as e.g. described by Ralston. The solution of ''N'' linear equations in ''N'' unknowns by elimination was already known to ancient Chinese. Before Gauss many mathematicians in Eurasia were performing and perfecting it yet as the method became relegated to school grade, few of them left any detailed descriptions. Thus the name Gaussian elimination is only a convenient abbreviation of a complex history. The Polish astronomer Tadeusz Banachiewicz introduced the LU decomposition in 1938. To quote: "It appears that Gauss and Doolittle applied the method f eliminationonly to symmetric equations. More recent authors, for example, Aitken, Banachiewicz, Dwyer, and Crout … have emphasized the use of the method, or variations of it, in connection with non-symmetric problems … Banachiewicz … saw the point … that the basic problem is really one of matrix factorization, or “decomposition” as he called it." Banachiewicz was the first to consider elimination in terms of matrices and in this way formulated LU decomposition, as demonstrated by his graphic illustration. His calculations follow ordinary matrix ones, yet notation deviates in that he preferred to write one factor transposed, to be able to multiply them mechanically column by column, by sliding ruler down consecutive rows of both (using
arithmometer The arithmometer () was the first digital data, digital mechanical calculator strong and reliable enough to be used daily in an office environment. This calculator could add and subtract two numbers directly and perform Multiplication algorithm, ...
). Combined with swapped order of indices his formulae in modern notation read \cdot IA'=\;\rightarrow\;A'\equiv(A, ),\;\;A=G\cdot H\;\rightarrow\;A^T=G^TH, where \;\;IA\;\rightarrow\;A^T, \equiv _1,\cdots,x_n,-1/math>, primes refer to matrices extended with the last column, and the last component of is -1. Matrix formulae to calculate rows and columns of LU factors by recursion are given in the remaining part of Banachiewicz's paper as Eq. (2.3) and (2.4) (see F90 code example). This paper by Banachiewicz contains both derivation of LU and R^TR factors of respectively non-symmetric and symmetric matrices. They are sometimes confused as later publications tend to tie his name solely with the rediscovery of Cholesky decomposition. Banachiewicz himself can be excused of inaction as already next year he suffered from persecution by occupiers, spending three month in the Sachsenhausen Concentration Camp, on release from which he carried himself from a train his collaborator and co-prisoner Antoni Wilk, who died of exhaustion a week later.


Code examples


Fortran90 code example

Module mlu Implicit None Integer, Parameter :: SP = Kind(1d0) ! set I/O real precision Private Public luban, lusolve Contains Subroutine luban (a, tol, g, h, ip, condinv, detnth) ! By Banachiewicz (1938, hereafter B38) LU decomposition method calculates such ! triangles L=G^T, and U=H that square B=A^T=G^TH=LU. Partial pivoting ! by column permutation IP(:) is modern addition. ! Within the code a, g correspond to B38 A^T and G^T, so that a=gh holds. ! ! Normal use is for square A, however for RHS l already known ! input of (A, l)^T yields (L, y^T)^T where x in L^Tx=y is solution of Ax=l. Real (SP), Intent (In) :: a (:, :)! input matrix A(m,n), n<=m Real (SP), Intent (In) :: tol ! tolerance for near zero pivot Real (SP), Intent (Out) :: g (size(a,dim=1), size(a,dim=2)) ! L(m,n) Real (SP), Intent (Out) :: h (size(a,dim=2), size(a,dim=2)) ! U(n,n) ! note U columns are permuted Real (SP), Intent (Out) :: condinv ! 1/cond(A), 0 for singular A Real (SP), Intent (Out) :: detnth ! sign*Abs(det(A))**(1/n) Integer, Intent (Out) :: ip (size(a, dim=2)) ! columns permutation ! Integer :: k, n, j, l, isig Real (SP) :: tol0, pivmax, pivmin, piv ! n = size (a, dim=2) tol0 = Max (tol, 3._SP*epsilon(tol0))! use default for tol=0 ! ! Rectangular A and G are permitted under condition: If (n > size(a, dim=1) .Or. n < 1) Stop 91 Forall (k=1:n) ip (k) = k h = 0._SP g = 0._SP isig = 1 detnth = 0._SP pivmax = Maxval (Abs (a(1, :))) pivmin = pivmax ! Do k = 1, n ! Banachiewicz (1938) Eq. (2.3) h(k, ip(k:)) = a(k, ip(k:)) - Matmul(g(k, :k-1), h(:k-1, ip(k:))) ! ! Find row pivot j = (Maxloc(Abs(h(k, ip(k:))), dim=1) + k-1) If (j /= k) Then ! Swap columns j and k isig = - isig ! Change Det(A) sign because of permutation l = ip (k) ip (k) = ip (j) ip (j) = l End If piv = Abs (h(k, ip(k))) pivmax = Max (piv, pivmax) ! Adjust condinv pivmin = Min (piv, pivmin) If (piv < tol0) Then ! singular matrix isig = 0 pivmax = 1._SP Exit Else ! Account for pivot contribution to Det(A) sign and value If (h(k, ip(k)) < 0._SP) isig = - isig detnth = detnth + Log (piv) End If ! ! Transposed Banachiewicz (1938) Eq. (2.4) g (k+1:, k) = (a(k+1:, ip(k)) - & Matmul(g(k+1:, :k-1), h(:k-1, ip(k)))) / h (k, ip(k)) g (k, k) = 1._SP End Do ! detnth = isig * Exp (detnth/n) condinv = Abs (isig) * pivmin / pivmax ! Test for square A(n,n) by uncommenting below ! Print *, ', AQ-LU, ',Maxval (Abs(a(:,ip(:))-Matmul(g, h(:,ip(:))))) End Subroutine luban Subroutine lusolve(l,u,ip,x) ! Solves Ax=a system using triangle factors LU=A Real (SP), Intent (In) :: l (:, :) ! lower triangle matrix L(n,n) Real (SP), Intent (In) :: u (:, :) ! upper triangle matrix U(n,n) Integer, Intent (In) :: ip (:) ! columns permutation IP(n) Real (SP), Intent (InOut) :: x (:, :) ! X(n,m) for m sets of input ! right hand sides replaced with output unknowns Integer :: n, m, i, j n = size(ip) m = size(x, dim=2) If (n<1.Or.m<1.Or.Any( ,n=shape(l)).Or.Any(shape(l)/=shape(u)).Or. & n/=size(x,dim=1)) Stop 91 Do i = 1, m Do j = 1, n x(j,i) = x(j,i)-dot_product(x(:j-1,i),l(j,:j-1)) End Do Do j = n, 1, -1 x(j,i) = (x(j,i)-dot_product(x(j+1:,i),u(j,ip(j+1:)))) / & u(j,ip(j)) End Do End Do End Subroutine lusolve End Module mlu


C code example

/* INPUT: A - array of pointers to rows of a square matrix having dimension N * Tol - small tolerance number to detect failure when the matrix is near degenerate * OUTPUT: Matrix A is changed, it contains a copy of both matrices L-E and U as A=(L-E)+U such that P*A=L*U. * The permutation matrix is not stored as a matrix, but in an integer vector P of size N+1 * containing column indexes where the permutation matrix has "1". The last element P S+N, * where S is the number of row exchanges needed for determinant computation, det(P)=(-1)^S */ int LUPDecompose(double **A, int N, double Tol, int *P) /* INPUT: A,P filled in LUPDecompose; b - rhs vector; N - dimension * OUTPUT: x - solution vector of A*x=b */ void LUPSolve(double **A, int *P, double *b, int N, double *x) /* INPUT: A,P filled in LUPDecompose; N - dimension * OUTPUT: IA is the inverse of the initial matrix */ void LUPInvert(double **A, int *P, int N, double **IA) /* INPUT: A,P filled in LUPDecompose; N - dimension. * OUTPUT: Function returns the determinant of the initial matrix */ double LUPDeterminant(double **A, int *P, int N)


C# code example

public class SystemOfLinearEquations


MATLAB code example

function LU = LUDecompDoolittle(A) n = length(A); LU = A; for k = 2:n for i = 1 : k-1 lamda = LU(k,i) / LU (i, i); LU(k,i) = lamda; LU(k,i+1:n) = LU(k,i+1:n) - LU(i,i+1:n) * lamda; end end end function x = SolveLinearSystem(LU, B) n = length(LU); y = zeros(size(B)); % find solution of Ly = B for i = 1:n y(i,:) = B(i,:) - LU(i,1:i)*y(1:i,:); end % find solution of Ux = y x = zeros(size(B)); for i = n:(-1):1 x(i,:) = (y(i,:) - LU(i,(i + 1):n)*x((i + 1):n,:))/LU(i, i); end end A = 4 3 3; 6 3 3; 3 4 3 LU = LUDecompDoolittle(A) B = 1 2 3; 4 5 6; 7 8 9; 10 11 12 x = SolveLinearSystem(LU, B) A * x


See also

* Block LU decomposition * Bruhat decomposition * Cholesky decomposition * Crout matrix decomposition * Incomplete LU factorization * LU Reduction * Matrix decomposition * QR decomposition


Notes


References

* . * . * . * . * . * . * . See Section 3.5. ''N'' − 1 * . * . * . * . * . * . * . *


External links

References
LU decomposition
on ''MathWorld''.
LU decomposition
on ''Math-Linux''.

at ''Holistic Numerical Methods Institute''

MATLAB reference. Computer code
LAPACK
is a collection of FORTRAN subroutines for solving dense linear algebra problems
ALGLIB
includes a partial port of the LAPACK to C++, C#, Delphi, etc.

Prof. J. Loomis,
University of Dayton The University of Dayton (UD) is a Private university, private, Catholic research university in Dayton, Ohio, United States. Founded in 1850 by the Society of Mary (Marianists), Society of Mary, it is one of three Marianist universities in the U ...

C code
Mathematics Source Library
Rust code

LU in X10
Online resources
WebApp descriptively solving systems of linear equations with LU Decomposition

Matrix Calculator with steps, including LU decomposition

LU Decomposition Tool
uni-bonn.de
LU Decomposition
by Ed Pegg, Jr.,
The Wolfram Demonstrations Project The Wolfram Demonstrations Project is an open-source collection of interactive programmes called Demonstrations. It is hosted by Wolfram Research. At its launch, it contained 1300 demonstrations but has grown to over 10,000. The site won a Pa ...
, 2007. {{Numerical linear algebra Matrix decompositions Numerical linear algebra Articles with example C code Articles with example C Sharp code de:Gaußsches Eliminationsverfahren#LR-Zerlegung