HOME

TheInfoList



OR:

In the
mathematical Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
discipline of
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as: :a_1x_1+\cdots +a_nx_n=b, linear maps such as: :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrices ...
, the Schur decomposition or Schur triangulation, named after
Issai Schur Issai Schur (10 January 1875 – 10 January 1941) was a Russian mathematician who worked in Germany for most of his life. He studied at the University of Berlin. He obtained his doctorate in 1901, became lecturer in 1903 and, after a stay at th ...
, is a
matrix decomposition In the mathematical discipline of linear algebra, a matrix decomposition or matrix factorization is a factorization of a matrix into a product of matrices. There are many different matrix decompositions; each finds use among a particular class of ...
. It allows one to write an arbitrary complex square matrix as unitarily equivalent to an upper triangular matrix whose diagonal elements are the eigenvalues of the original matrix.


Statement

The Schur decomposition reads as follows: if ''A'' is an square matrix with
complex Complex commonly refers to: * Complexity, the behaviour of a system whose components interact in multiple ways so possible interactions are difficult to describe ** Complex system, a system composed of many components which may interact with each ...
entries, then ''A'' can be expressed as(Section 2.3 and further at p. 79(Section 7.7 at p. 313 : A = Q U Q^ where ''Q'' is a
unitary matrix In linear algebra, a complex square matrix is unitary if its conjugate transpose is also its inverse, that is, if U^* U = UU^* = UU^ = I, where is the identity matrix. In physics, especially in quantum mechanics, the conjugate transpose is ...
(so that its inverse ''Q''−1 is also the
conjugate transpose In mathematics, the conjugate transpose, also known as the Hermitian transpose, of an m \times n complex matrix \boldsymbol is an n \times m matrix obtained by transposing \boldsymbol and applying complex conjugate on each entry (the complex c ...
''Q''* of ''Q''), and ''U'' is an upper triangular matrix, which is called a Schur form of ''A''. Since ''U'' is similar to ''A'', it has the same
spectrum A spectrum (plural ''spectra'' or ''spectrums'') is a condition that is not limited to a specific set of values but can vary, without gaps, across a continuum. The word was first used scientifically in optics to describe the rainbow of colors ...
, and since it is triangular, its
eigenvalue In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
s are the diagonal entries of ''U''. The Schur decomposition implies that there exists a nested sequence of ''A''-invariant subspaces , and that there exists an ordered
orthonormal basis In mathematics, particularly linear algebra, an orthonormal basis for an inner product space ''V'' with finite dimension is a basis for V whose vectors are orthonormal, that is, they are all unit vectors and orthogonal to each other. For examp ...
(for the standard
Hermitian form In mathematics, a sesquilinear form is a generalization of a bilinear form that, in turn, is a generalization of the concept of the dot product of Euclidean space. A bilinear form is linear in each of its arguments, but a sesquilinear form allow ...
of ) such that the first ''i'' basis vectors span for each ''i'' occurring in the nested sequence. Phrased somewhat differently, the first part says that a linear operator ''J'' on a complex finite-dimensional vector space stabilizes a complete flag .


Proof

A constructive proof for the Schur decomposition is as follows: every operator ''A'' on a complex finite-dimensional vector space has an eigenvalue ''λ'', corresponding to some eigenspace ''Vλ''. Let ''Vλ'' be its orthogonal complement. It is clear that, with respect to this orthogonal decomposition, ''A'' has matrix representation (one can pick here any orthonormal bases ''Z1'' and ''Z2'' spanning ''Vλ'' and ''Vλ'' respectively) :\begin Z_1 & Z_2 \end^ A \beginZ_1 & Z_2\end = \begin \lambda \, I_ & A_ \\ 0 & A_ \end: \begin V_ \\ \oplus \\ V_^ \end \rightarrow \begin V_ \\ \oplus \\ V_^ \end where ''Iλ'' is the identity operator on ''Vλ''. The above matrix would be upper-triangular except for the ''A''22 block. But exactly the same procedure can be applied to the sub-matrix ''A''22, viewed as an operator on ''Vλ'', and its submatrices. Continue this way until the resulting matrix is upper triangular. Since each conjugation increases the dimension of the upper-triangular block by at least one, this process takes at most ''n'' steps. Thus the space C''n'' will be exhausted and the procedure has yielded the desired result. The above argument can be slightly restated as follows: let ''λ'' be an eigenvalue of ''A'', corresponding to some eigenspace ''Vλ''. ''A'' induces an operator ''T'' on the quotient space C''n''/''Vλ''. This operator is precisely the ''A''22 submatrix from above. As before, ''T'' would have an eigenspace, say ''Wμ'' ⊂ C''n'' modulo ''Vλ''. Notice the preimage of ''Wμ'' under the quotient map is an invariant subspace of ''A'' that contains ''Vλ''. Continue this way until the resulting quotient space has dimension 0. Then the successive preimages of the eigenspaces found at each step form a flag that ''A'' stabilizes.


Notes

Although every square matrix has a Schur decomposition, in general this decomposition is not unique. For example, the eigenspace ''Vλ'' can have dimension > 1, in which case any orthonormal basis for ''Vλ'' would lead to the desired result. Write the triangular matrix ''U'' as ''U'' = ''D'' + ''N'', where ''D'' is diagonal and ''N'' is strictly upper triangular (and thus a
nilpotent matrix In linear algebra, a nilpotent matrix is a square matrix ''N'' such that :N^k = 0\, for some positive integer k. The smallest such k is called the index of N, sometimes the degree of N. More generally, a nilpotent transformation is a linear tr ...
). The diagonal matrix ''D'' contains the eigenvalues of ''A'' in arbitrary order (hence its Frobenius norm, squared, is the sum of the squared moduli of the eigenvalues of ''A'', while the Frobenius norm of ''A'', squared, is the sum of the squared
singular value In mathematics, in particular functional analysis, the singular values, or ''s''-numbers of a compact operator T: X \rightarrow Y acting between Hilbert spaces X and Y, are the square roots of the (necessarily non-negative) eigenvalues of the self ...
s of ''A''). The nilpotent part ''N'' is generally not unique either, but its
Frobenius norm In mathematics, a matrix norm is a vector norm in a vector space whose elements (vectors) are matrices (of given dimensions). Preliminaries Given a field K of either real or complex numbers, let K^ be the -vector space of matrices with m ro ...
is uniquely determined by ''A'' (just because the Frobenius norm of A is equal to the Frobenius norm of ''U'' = ''D'' + ''N''). It is clear that if ''A'' is a
normal matrix In mathematics, a complex square matrix is normal if it commutes with its conjugate transpose : The concept of normal matrices can be extended to normal operators on infinite dimensional normed spaces and to normal elements in C*-algebras. As ...
, then ''U'' from its Schur decomposition must be a
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagonal m ...
and the column vectors of ''Q'' are the
eigenvector In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
s of ''A''. Therefore, the Schur decomposition extends the spectral decomposition. In particular, if ''A'' is positive definite, the Schur decomposition of ''A'', its spectral decomposition, and its
singular value decomposition In linear algebra, the singular value decomposition (SVD) is a factorization of a real or complex matrix. It generalizes the eigendecomposition of a square normal matrix with an orthonormal eigenbasis to any \ m \times n\ matrix. It is re ...
coincide. A
commuting Commuting is periodically recurring travel between one's place of residence and place of work or study, where the traveler, referred to as a commuter, leaves the boundary of their home community. By extension, it can sometimes be any regul ...
family of matrices can be simultaneously triangularized, i.e. there exists a unitary matrix ''Q'' such that, for every ''Ai'' in the given family, ''Q Ai Q*'' is upper triangular. This can be readily deduced from the above proof. Take element ''A'' from and again consider an eigenspace ''VA''. Then ''VA'' is invariant under all matrices in . Therefore, all matrices in must share one common eigenvector in ''VA''. Induction then proves the claim. As a corollary, we have that every commuting family of normal matrices can be simultaneously
diagonalized In linear algebra, a square matrix A is called diagonalizable or non-defective if it is similar to a diagonal matrix, i.e., if there exists an invertible matrix P and a diagonal matrix D such that or equivalently (Such D are not unique.) F ...
. In the infinite dimensional setting, not every
bounded operator In functional analysis and operator theory, a bounded linear operator is a linear transformation L : X \to Y between topological vector spaces (TVSs) X and Y that maps bounded subsets of X to bounded subsets of Y. If X and Y are normed vector ...
on a Banach space has an invariant subspace. However, the upper-triangularization of an arbitrary square matrix does generalize to
compact operator In functional analysis, a branch of mathematics, a compact operator is a linear operator T: X \to Y, where X,Y are normed vector spaces, with the property that T maps bounded subsets of X to relatively compact subsets of Y (subsets with compact c ...
s. Every
compact operator In functional analysis, a branch of mathematics, a compact operator is a linear operator T: X \to Y, where X,Y are normed vector spaces, with the property that T maps bounded subsets of X to relatively compact subsets of Y (subsets with compact c ...
on a complex Banach space has a
nest A nest is a structure built for certain animals to hold eggs or young. Although nests are most closely associated with birds, members of all classes of vertebrates and some invertebrates construct nests. They may be composed of organic materi ...
of closed invariant subspaces.


Computation

The Schur decomposition of a given matrix is numerically computed by the
QR algorithm In numerical linear algebra, the QR algorithm or QR iteration is an eigenvalue algorithm: that is, a procedure to calculate the eigenvalues and eigenvectors of a matrix. The QR algorithm was developed in the late 1950s by John G. F. Francis and by ...
or its variants. In other words, the roots of the characteristic polynomial corresponding to the matrix are not necessarily computed ahead in order to obtain its Schur decomposition. Conversely, the
QR algorithm In numerical linear algebra, the QR algorithm or QR iteration is an eigenvalue algorithm: that is, a procedure to calculate the eigenvalues and eigenvectors of a matrix. The QR algorithm was developed in the late 1950s by John G. F. Francis and by ...
can be used to compute the roots of any given characteristic polynomial by finding the Schur decomposition of its
companion matrix In linear algebra, the Frobenius companion matrix of the monic polynomial : p(t)=c_0 + c_1 t + \cdots + c_t^ + t^n ~, is the square matrix defined as :C(p)=\begin 0 & 0 & \dots & 0 & -c_0 \\ 1 & 0 & \dots & 0 & -c_1 \\ 0 & 1 & \dots & 0 & -c_2 ...
. Similarly, the
QR algorithm In numerical linear algebra, the QR algorithm or QR iteration is an eigenvalue algorithm: that is, a procedure to calculate the eigenvalues and eigenvectors of a matrix. The QR algorithm was developed in the late 1950s by John G. F. Francis and by ...
is used to compute the eigenvalues of any given matrix, which are the diagonal entries of the upper triangular matrix of the Schur decomposition. Although the
QR algorithm In numerical linear algebra, the QR algorithm or QR iteration is an eigenvalue algorithm: that is, a procedure to calculate the eigenvalues and eigenvectors of a matrix. The QR algorithm was developed in the late 1950s by John G. F. Francis and by ...
is formally an infinite sequence of operations, convergence to machine precision is practically achieved in \mathcal(n^3) operations. See the Nonsymmetric Eigenproblems section in
LAPACK LAPACK ("Linear Algebra Package") is a standard software library for numerical linear algebra. It provides routines for solving systems of linear equations and linear least squares, eigenvalue problems, and singular value decomposition. It als ...
Users' Guide.


Applications

Lie theory In mathematics, the mathematician Sophus Lie ( ) initiated lines of study involving integration of differential equations, transformation groups, and contact of spheres that have come to be called Lie theory. For instance, the latter subject is ...
applications include: * Every invertible operator is contained in a
Borel group In the theory of algebraic groups, a Borel subgroup of an algebraic group ''G'' is a maximal Zariski closed and connected solvable algebraic subgroup. For example, in the general linear group ''GLn'' (''n x n'' invertible matrices), the subgroup ...
. * Every operator fixes a point of the
flag manifold In mathematics, a generalized flag variety (or simply flag variety) is a homogeneous space whose points are flags in a finite-dimensional vector space ''V'' over a field F. When F is the real or complex numbers, a generalized flag variety is a smo ...
.


Generalized Schur decomposition

Given square matrices ''A'' and ''B'', the generalized Schur decomposition factorizes both matrices as A=QSZ^* and B=QTZ^*, where ''Q'' and ''Z'' are unitary, and ''S'' and ''T'' are
upper triangular In mathematics, a triangular matrix is a special kind of square matrix. A square matrix is called if all the entries ''above'' the main diagonal are zero. Similarly, a square matrix is called if all the entries ''below'' the main diagonal ar ...
. The generalized Schur decomposition is also sometimes called the QZ decomposition. The generalized
eigenvalue In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
s \lambda that solve the generalized eigenvalue problem A\mathbf=\lambda B\mathbf (where x is an unknown nonzero vector) can be calculated as the ratio of the diagonal elements of ''S'' to those of ''T''. That is, using subscripts to denote matrix elements, the ''i''th generalized eigenvalue \lambda_i satisfies \lambda_i=T_/S_{ii}.


References

Matrix theory Articles containing proofs Matrix decompositions