HOME

TheInfoList



OR:

In linear algebra, eigendecomposition is the factorization of a matrix into a canonical form, whereby the matrix is represented in terms of its eigenvalues and eigenvectors. Only diagonalizable matrices can be factorized in this way. When the matrix being factorized is a normal or real symmetric matrix, the decomposition is called "spectral decomposition", derived from the spectral theorem.


Fundamental theory of matrix eigenvectors and eigenvalues

A (nonzero) vector of dimension is an eigenvector of a square matrix if it satisfies a linear equation of the form \mathbf \mathbf = \lambda \mathbf for some scalar . Then is called the eigenvalue corresponding to . Geometrically speaking, the eigenvectors of are the vectors that merely elongates or shrinks, and the amount that they elongate/shrink by is the eigenvalue. The above equation is called the eigenvalue equation or the eigenvalue problem. This yields an equation for the eigenvalues p\left(\lambda\right) = \det\left(\mathbf - \lambda\mathbf\right) = 0. We call the characteristic polynomial, and the equation, called the characteristic equation, is an th-order polynomial equation in the unknown . This equation will have distinct solutions, where . The set of solutions, that is, the eigenvalues, is called the spectrum of . If the field of scalars is algebraically closed, then we can factor as p(\lambda) = \left(\lambda - \lambda_1\right)^\left(\lambda - \lambda_2\right)^ \cdots \left(\lambda-\lambda_\right)^ = 0. The integer is termed the algebraic multiplicity of eigenvalue . The algebraic multiplicities sum to : \sum_^ = N. For each eigenvalue , we have a specific eigenvalue equation \left(\mathbf - \lambda_i \mathbf\right)\mathbf = 0. There will be linearly independent solutions to each eigenvalue equation. The linear combinations of the solutions (except the one which gives the zero vector) are the eigenvectors associated with the eigenvalue . The integer is termed the geometric multiplicity of . It is important to keep in mind that the algebraic multiplicity and geometric multiplicity may or may not be equal, but we always have . The simplest case is of course when . The total number of linearly independent eigenvectors, , can be calculated by summing the geometric multiplicities \sum_^ = N_. The eigenvectors can be indexed by eigenvalues, using a double index, with being the th eigenvector for the th eigenvalue. The eigenvectors can also be indexed using the simpler notation of a single index , with .


Eigendecomposition of a matrix

Let be a square matrix with linearly independent eigenvectors (where ). Then can be factored as \mathbf=\mathbf\mathbf\mathbf^ where is the square matrix whose th column is the eigenvector of , and is the
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagon ...
whose diagonal elements are the corresponding eigenvalues, . Note that only diagonalizable matrices can be factorized in this way. For example, the defective matrix \left \begin 1 & 1 \\ 0 & 1 \end \right/math> (which is a shear matrix) cannot be diagonalized. The eigenvectors are usually normalized, but they don't have to be. A non-normalized set of eigenvectors, can also be used as the columns of . That can be understood by noting that the magnitude of the eigenvectors in gets canceled in the decomposition by the presence of . If one of the eigenvalues has multiple linearly independent eigenvectors (that is, the geometric multiplicity of is greater than 1), then these eigenvectors for this eigenvalue can be chosen to be mutually orthogonal; however, if two eigenvectors belong to two different eigenvalues, it may be impossible for them to be orthogonal to each other (see Example below). One special case is that if is a normal matrix, then by the spectral theorem, it's always possible to diagonalize in an orthonormal basis . The decomposition can be derived from the fundamental property of eigenvectors: \begin \mathbf \mathbf &= \lambda \mathbf \\ \mathbf \mathbf &= \mathbf \mathbf \\ \mathbf &= \mathbf\mathbf\mathbf^ . \end The linearly independent eigenvectors with nonzero eigenvalues form a basis (not necessarily orthonormal) for all possible products , for , which is the same as the
image An image or picture is a visual representation. An image can be Two-dimensional space, two-dimensional, such as a drawing, painting, or photograph, or Three-dimensional space, three-dimensional, such as a carving or sculpture. Images may be di ...
(or range) of the corresponding matrix transformation, and also the column space of the matrix . The number of linearly independent eigenvectors with nonzero eigenvalues is equal to the rank of the matrix , and also the dimension of the image (or range) of the corresponding matrix transformation, as well as its column space. The linearly independent eigenvectors with an eigenvalue of zero form a basis (which can be chosen to be orthonormal) for the null space (also known as the kernel) of the matrix transformation .


Example

The 2 × 2 real matrix \mathbf = \begin 1 & 0 \\ 1 & 3 \\ \end may be decomposed into a diagonal matrix through multiplication of a non-singular matrix \mathbf = \begin a & b \\ c & d \end \in \mathbb^. Then \begin a & b \\ c & d \end^\begin 1 & 0 \\ 1 & 3 \end\begin a & b \\ c & d \end = \begin x & 0 \\ 0 & y \end, for some real diagonal matrix \left \begin x & 0 \\ 0 & y \end \right/math>. Multiplying both sides of the equation on the left by : \begin 1 & 0 \\ 1 & 3 \end \begin a & b \\ c & d \end = \begin a & b \\ c & d \end \begin x & 0 \\ 0 & y \end. The above equation can be decomposed into two simultaneous equations: \begin \begin 1 & 0\\ 1 & 3 \end \begin a \\ c \end = \begin ax \\ cx \end \\ .2ex\begin 1 & 0\\ 1 & 3 \end \begin b \\ d \end = \begin by \\ dy \end \end . Factoring out the eigenvalues and : \begin \begin 1 & 0\\ 1 & 3 \end \begin a \\ c \end = x\begin a \\ c \end \\ .2ex\begin 1 & 0\\ 1 & 3 \end \begin b \\ d \end = y\begin b \\ d \end \end Letting \mathbf = \begin a \\ c \end, \quad \mathbf = \begin b \\ d \end, this gives us two vector equations: \begin \mathbf \mathbf = x \mathbf \\ \mathbf \mathbf = y \mathbf \end And can be represented by a single vector equation involving two solutions as eigenvalues: \mathbf \mathbf = \lambda \mathbf where represents the two eigenvalues and , and represents the vectors and . Shifting to the left hand side and factoring out \left(\mathbf - \lambda \mathbf\right) \mathbf = \mathbf Since is non-singular, it is essential that is nonzero. Therefore, \det(\mathbf - \lambda \mathbf) = 0 Thus (1- \lambda)(3 - \lambda) = 0 giving us the solutions of the eigenvalues for the matrix as or , and the resulting diagonal matrix from the eigendecomposition of is thus Putting the solutions back into the above simultaneous equations \begin \begin 1 & 0 \\ 1 & 3 \end \begin a \\ c \end = 1\begin a \\ c \end \\ .2ex\begin 1 & 0 \\ 1 & 3 \end \begin b \\ d \end = 3\begin b \\ d \end \end Solving the equations, we have a = -2c \quad\text \quad b = 0, \qquad c,d \in \mathbb. Thus the matrix required for the eigendecomposition of is \mathbf = \begin -2c & 0 \\ c & d \end,\qquad c, d\in \mathbb, that is: \begin -2c & 0 \\ c & d \end^ \begin 1 & 0 \\ 1 & 3 \end \begin -2c & 0 \\ c & d \end = \begin 1 & 0 \\ 0 & 3 \end,\qquad c, d\in \mathbb


Matrix inverse via eigendecomposition

If a matrix can be eigendecomposed and if none of its eigenvalues are zero, then is invertible and its inverse is given by \mathbf^ = \mathbf\mathbf^\mathbf^ If \mathbf is a symmetric matrix, since \mathbf is formed from the eigenvectors of \mathbf, \mathbf is guaranteed to be an orthogonal matrix, therefore \mathbf^ = \mathbf^\mathrm. Furthermore, because is a
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagon ...
, its inverse is easy to calculate: \left mathbf^\right = \frac


Practical implications

When eigendecomposition is used on a matrix of measured, real
data Data ( , ) are a collection of discrete or continuous values that convey information, describing the quantity, quality, fact, statistics, other basic units of meaning, or simply sequences of symbols that may be further interpreted for ...
, the inverse may be less valid when all eigenvalues are used unmodified in the form above. This is because as eigenvalues become relatively small, their contribution to the inversion is large. Those near zero or at the "noise" of the measurement system will have undue influence and could hamper solutions (detection) using the inverse. Two mitigations have been proposed: truncating small or zero eigenvalues, and extending the lowest reliable eigenvalue to those below it. See also Tikhonov regularization as a statistically motivated but biased method for rolling off eigenvalues as they become dominated by noise. The first mitigation method is similar to a sparse sample of the original matrix, removing components that are not considered valuable. However, if the solution or detection process is near the noise level, truncating may remove components that influence the desired solution. The second mitigation extends the eigenvalue so that lower values have much less influence over inversion, but do still contribute, such that solutions near the noise will still be found. The reliable eigenvalue can be found by assuming that eigenvalues of extremely similar and low value are a good representation of measurement noise (which is assumed low for most systems). If the eigenvalues are rank-sorted by value, then the reliable eigenvalue can be found by minimization of the Laplacian of the sorted eigenvalues: \min\left, \nabla^2 \lambda_\mathrm\ where the eigenvalues are subscripted with an to denote being sorted. The position of the minimization is the lowest reliable eigenvalue. In measurement systems, the square root of this reliable eigenvalue is the average noise over the components of the system.


Functional calculus

The eigendecomposition allows for much easier computation of power series of matrices. If is given by f(x) = a_0 + a_1 x + a_2 x^2 + \cdots then we know that f\!\left(\mathbf\right) = \mathbf\,f\!\left(\mathbf\right)\mathbf^ Because is a
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagon ...
, functions of are very easy to calculate: \left \left(\mathbf\right)\right = f\left(\lambda_i\right) The off-diagonal elements of are zero; that is, is also a diagonal matrix. Therefore, calculating reduces to just calculating the function on each of the eigenvalues. A similar technique works more generally with the holomorphic functional calculus, using \mathbf^ = \mathbf \mathbf^ \mathbf^ from above. Once again, we find that \left \left(\mathbf\right)\right = f\left(\lambda_i\right)


Examples

\begin \mathbf^2 &= \left(\mathbf\mathbf\mathbf^\right) \left(\mathbf\mathbf\mathbf^\right) = \mathbf\mathbf\left(\mathbf^\mathbf\right)\mathbf\mathbf^ = \mathbf\mathbf^2\mathbf^ \\ .2ex \mathbf^n &= \mathbf\mathbf^n\mathbf^ \\ .2ex \exp \mathbf &= \mathbf \exp(\mathbf) \mathbf^ \end which are examples for the functions f(x)=x^2, \; f(x)=x^n, \; f(x)=\exp . Furthermore, \exp is the
matrix exponential In mathematics, the matrix exponential is a matrix function on square matrix, square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exp ...
.


Decomposition for spectral matrices

Spectral matrices are matrices that possess distinct eigenvalues and a complete set of eigenvectors. This characteristic allows spectral matrices to be fully diagonalizable, meaning they can be decomposed into simpler forms using eigendecomposition. This decomposition process reveals fundamental insights into the matrix's structure and behavior, particularly in fields such as quantum mechanics, signal processing, and numerical analysis.


Normal matrices

A complex-valued square matrix A is normal (meaning , \mathbf^*\mathbf=\mathbf \mathbf^*, where \mathbf^* is the conjugate transpose) if and only if it can be decomposed as \mathbf=\mathbf \mathbf\Lambda\mathbf^*, where \mathbf is a unitary matrix (meaning \mathbf^*=\mathbf^) and \mathbf\Lambda = diag(\lambda_1, \ldots,\lambda_n) is a
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagon ...
. The columns \mathbf_1,\cdots,\mathbf_n of \mathbf form an orthonormal basis and are eigenvectors of \mathbf with corresponding eigenvalues \lambda_1, \ldots,\lambda_n. For example, consider the 2 x 2 normal matrix \mathbf=\begin 1 & 2 \\ 2 & 1 \end. The eigenvalues are \lambda_1=3 and \lambda_2 = -1. The (normalized) eigenvectors corresponding to these eigenvalues are \mathbf_1=\frac\begin1\\1\end and \mathbf_2=\frac\begin-1\\1\end. The diagonalization is \mathbf=\mathbf \mathbf\Lambda\mathbf^*, where \mathbf=\begin 1/\sqrt & 1/\sqrt \\ 1/\sqrt & -1/\sqrt \end, \mathbf\Lambda =\begin 3 & 0 \\ 0 & -1\end and \mathbf^*=\mathbf^=\begin 1/\sqrt & 1/\sqrt \\ 1/\sqrt & -1/\sqrt \end. The verification is \mathbf \mathbf\Lambda\mathbf^*=\begin 1/\sqrt & 1/\sqrt \\ 1/\sqrt & -1/\sqrt \end\begin 3 & 0 \\ 0 & -1\end\begin 1/\sqrt & 1/\sqrt \\ 1/\sqrt & -1/\sqrt \end=\begin 1 & 2 \\ 2 & 1 \end=\mathbf. This example illustrates the process of diagonalizing a normal matrix \mathbf by finding its eigenvalues and eigenvectors, forming the unitary matrix \mathbf, the diagonal matrix \mathbf\Lambda, and verifying the decomposition.


Real symmetric matrices

As a special case, for every real symmetric matrix, the eigenvalues are real and the eigenvectors can be chosen real and orthonormal. Thus a real symmetric matrix can be decomposed as \mathbf=\mathbf \mathbf\mathbf^\mathsf, where is an orthogonal matrix whose columns are the real, orthonormal eigenvectors of , and is a diagonal matrix whose entries are the eigenvalues of .


Diagonalizable matrices

Diagonalizable matrices can be decomposed using eigendecomposition, provided they have a full set of linearly independent eigenvectors. They can be expressed as\mathbf=\mathbf \mathbf\mathbf^, where \mathbf is a matrix whose columns are eigenvectors of \mathbf and \mathbf is a diagonal matrix consisting of the corresponding eigenvalues of \mathbf.


Positive definite matrices

Positive definite matrices are matrices for which all eigenvalues are positive. They can be decomposed as \mathbf=\mathbf \mathbf^\mathsf using the Cholesky decomposition, where \mathbf is a lower triangular matrix.


Unitary and Hermitian matrices

Unitary matrices satisfy \mathbf\mathbf^*=\mathbf (real case) or \mathbf\mathbf^\dagger=\mathbf (complex case), where \mathbf^*denotes the conjugate transpose and \mathbf^\daggerdenotes the conjugate transpose. They diagonalize using unitary transformations. Hermitian matrices satisfy \mathbf=\mathbf^\dagger, where \mathbf^\daggerdenotes the conjugate transpose. They can be diagonalized using unitary or orthogonal matrices.


Useful facts


Useful facts regarding eigenvalues

*The product of the eigenvalues is equal to the
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of \det\left(\mathbf\right) = \prod_^ Note that each eigenvalue is raised to the power , the algebraic multiplicity. *The sum of the eigenvalues is equal to the trace of \operatorname\left(\mathbf\right) = \sum_^ Note that each eigenvalue is multiplied by , the algebraic multiplicity. *If the eigenvalues of are , and is invertible, then the eigenvalues of are simply . *If the eigenvalues of are , then the eigenvalues of are simply , for any holomorphic function .


Useful facts regarding eigenvectors

* If is Hermitian and full-rank, the basis of eigenvectors may be chosen to be mutually orthogonal. The eigenvalues are real. * The eigenvectors of are the same as the eigenvectors of . * Eigenvectors are only defined up to a multiplicative constant. That is, if then is also an eigenvector for any scalar . In particular, and (for any ''θ'') are also eigenvectors. * In the case of degenerate eigenvalues (an eigenvalue having more than one eigenvector), the eigenvectors have an additional freedom of linear transformation, that is to say, any linear (orthonormal) combination of eigenvectors sharing an eigenvalue (in the degenerate subspace) is itself an eigenvector (in the subspace).


Useful facts regarding eigendecomposition

* can be eigendecomposed if and only if the number of linearly independent eigenvectors, , equals the dimension of an eigenvector: * If the field of scalars is algebraically closed and if has no repeated roots, that is, if N_\lambda = N, then can be eigendecomposed. * The statement " can be eigendecomposed" does ''not'' imply that has an inverse as some eigenvalues may be zero, which is not invertible. * The statement " has an inverse" does ''not'' imply that can be eigendecomposed. A counterexample is \left \begin 1 & 1 \\ 0 & 1 \end \right/math>, which is an invertible defective matrix.


Useful facts regarding matrix inverse

* can be inverted if and only if all eigenvalues are nonzero: \lambda_i \ne 0 \quad \forall \,i * If ''and'' , the inverse is given by \mathbf^ = \mathbf\mathbf^\mathbf^


Numerical computations


Numerical computation of eigenvalues

Suppose that we want to compute the eigenvalues of a given matrix. If the matrix is small, we can compute them symbolically using the characteristic polynomial. However, this is often impossible for larger matrices, in which case we must use a numerical method. In practice, eigenvalues of large matrices are not computed using the characteristic polynomial. Computing the polynomial becomes expensive in itself, and exact (symbolic) roots of a high-degree polynomial can be difficult to compute and express: the Abel–Ruffini theorem implies that the roots of high-degree (5 or above) polynomials cannot in general be expressed simply using th roots. Therefore, general algorithms to find eigenvectors and eigenvalues are iterative. Iterative numerical algorithms for approximating roots of polynomials exist, such as Newton's method, but in general it is impractical to compute the characteristic polynomial and then apply these methods. One reason is that small round-off errors in the coefficients of the characteristic polynomial can lead to large errors in the eigenvalues and eigenvectors: the roots are an extremely ill-conditioned function of the coefficients. A simple and accurate iterative method is the power method: a random vector is chosen and a sequence of unit vectors is computed as \frac, \frac, \frac, \ldots This sequence will almost always converge to an eigenvector corresponding to the eigenvalue of greatest magnitude, provided that has a nonzero component of this eigenvector in the eigenvector basis (and also provided that there is only one eigenvalue of greatest magnitude). This simple algorithm is useful in some practical applications; for example,
Google Google LLC (, ) is an American multinational corporation and technology company focusing on online advertising, search engine technology, cloud computing, computer software, quantum computing, e-commerce, consumer electronics, and artificial ...
uses it to calculate the page rank of documents in their search engine. Also, the power method is the starting point for many more sophisticated algorithms. For instance, by keeping not just the last vector in the sequence, but instead looking at the span of ''all'' the vectors in the sequence, one can get a better (faster converging) approximation for the eigenvector, and this idea is the basis of Arnoldi iteration. Alternatively, the important QR algorithm is also based on a subtle transformation of a power method.


Numerical computation of eigenvectors

Once the eigenvalues are computed, the eigenvectors could be calculated by solving the equation \left(\mathbf - \lambda_i \mathbf\right)\mathbf_ = \mathbf using Gaussian elimination or any other method for solving matrix equations. However, in practical large-scale eigenvalue methods, the eigenvectors are usually computed in other ways, as a byproduct of the eigenvalue computation. In power iteration, for example, the eigenvector is actually computed before the eigenvalue (which is typically computed by the Rayleigh quotient of the eigenvector). In the QR algorithm for a Hermitian matrix (or any normal matrix), the orthonormal eigenvectors are obtained as a product of the matrices from the steps in the algorithm. (For more general matrices, the QR algorithm yields the Schur decomposition first, from which the eigenvectors can be obtained by a backsubstitution procedure.) For Hermitian matrices, the Divide-and-conquer eigenvalue algorithm is more efficient than the QR algorithm if both eigenvectors and eigenvalues are desired.


Additional topics


Generalized eigenspaces

Recall that the ''geometric'' multiplicity of an eigenvalue can be described as the dimension of the associated eigenspace, the nullspace of . The algebraic multiplicity can also be thought of as a dimension: it is the dimension of the associated generalized eigenspace (1st sense), which is the nullspace of the matrix for ''any sufficiently large ''. That is, it is the space of '' generalized eigenvectors'' (first sense), where a generalized eigenvector is any vector which ''eventually'' becomes 0 if is applied to it enough times successively. Any eigenvector is a generalized eigenvector, and so each eigenspace is contained in the associated generalized eigenspace. This provides an easy proof that the geometric multiplicity is always less than or equal to the algebraic multiplicity. This usage should not be confused with the ''generalized eigenvalue problem'' described below.


Conjugate eigenvector

A conjugate eigenvector or coneigenvector is a vector sent after transformation to a scalar multiple of its conjugate, where the scalar is called the conjugate eigenvalue or coneigenvalue of the linear transformation. The coneigenvectors and coneigenvalues represent essentially the same information and meaning as the regular eigenvectors and eigenvalues, but arise when an alternative coordinate system is used. The corresponding equation is \mathbf\mathbf = \lambda \mathbf^*. For example, in coherent electromagnetic scattering theory, the linear transformation represents the action performed by the scattering object, and the eigenvectors represent polarization states of the electromagnetic wave. In
optics Optics is the branch of physics that studies the behaviour and properties of light, including its interactions with matter and the construction of optical instruments, instruments that use or Photodetector, detect it. Optics usually describes t ...
, the coordinate system is defined from the wave's viewpoint, known as the Forward Scattering Alignment (FSA), and gives rise to a regular eigenvalue equation, whereas in
radar Radar is a system that uses radio waves to determine the distance ('' ranging''), direction ( azimuth and elevation angles), and radial velocity of objects relative to the site. It is a radiodetermination method used to detect and track ...
, the coordinate system is defined from the radar's viewpoint, known as the Back Scattering Alignment (BSA), and gives rise to a coneigenvalue equation.


Generalized eigenvalue problem

A generalized eigenvalue problem (second sense) is the problem of finding a (nonzero) vector that obeys \mathbf\mathbf = \lambda \mathbf \mathbf where and are matrices. If obeys this equation, with some , then we call the ''generalized eigenvector'' of and (in the second sense), and is called the ''generalized eigenvalue'' of and (in the second sense) which corresponds to the generalized eigenvector . The possible values of must obey the following equation \det(\mathbf - \lambda \mathbf) = 0. If linearly independent vectors can be found, such that for every , , then we define the matrices and such that P = \begin , & & , \\ \mathbf_1 & \cdots & \mathbf_n \\ , & & , \end \equiv \begin (\mathbf_1)_1 & \cdots & (\mathbf_n)_1 \\ \vdots & & \vdots \\ (\mathbf_1)_n & \cdots & (\mathbf_n)_n \end (D)_ = \begin \lambda_i, & \texti = j\\ 0, & \text \end Then the following equality holds \mathbf = \mathbf\mathbf\mathbf\mathbf^ And the proof is \mathbf\mathbf= \mathbf \begin , & & , \\ \mathbf_1 & \cdots & \mathbf_n \\ , & & , \end = \begin , & & , \\ A\mathbf_1 & \cdots & A\mathbf_n \\ , & & , \end = \begin , & & , \\ \lambda_1B\mathbf_1 & \cdots & \lambda_nB\mathbf_n \\ , & & , \end = \begin , & & , \\ B\mathbf_1 & \cdots & B\mathbf_n \\ , & & , \end \mathbf = \mathbf\mathbf\mathbf And since is invertible, we multiply the equation from the right by its inverse, finishing the proof. The set of matrices of the form , where is a complex number, is called a ''pencil''; the term '' matrix pencil'' can also refer to the pair of matrices. If is invertible, then the original problem can be written in the form \mathbf^\mathbf\mathbf = \lambda \mathbf which is a standard eigenvalue problem. However, in most situations it is preferable not to perform the inversion, but rather to solve the generalized eigenvalue problem as stated originally. This is especially important if and are Hermitian matrices, since in this case is not generally Hermitian and important properties of the solution are no longer apparent. If and are both symmetric or Hermitian, and is also a positive-definite matrix, the eigenvalues are real and eigenvectors and with distinct eigenvalues are -orthogonal (). In this case, eigenvectors can be chosen so that the matrix defined above satisfies \mathbf^* \mathbf B \mathbf = \mathbf or \mathbf\mathbf^*\mathbf B = \mathbf, and there exists a basis of generalized eigenvectors (it is not a defective problem). This case is sometimes called a ''Hermitian definite pencil'' or ''definite pencil''.


See also

* Eigenvalue perturbation * Frobenius covariant * Householder transformation * Jordan normal form * List of matrices * Matrix decomposition * Singular value decomposition * Sylvester's formula


Notes


References

* * * * {{refend


External links


Interactive program & tutorial of Spectral Decomposition
Matrix theory Matrix decompositions