In
linear algebra, a
square matrix
In mathematics, a square matrix is a matrix with the same number of rows and columns. An ''n''-by-''n'' matrix is known as a square matrix of order Any two square matrices of the same order can be added and multiplied.
Square matrices are often ...
is called diagonalizable or non-defective if it is
similar to a
diagonal matrix, i.e., if there exists an
invertible matrix and a diagonal matrix
such that or equivalently (Such
are not unique.) For a
finite-dimensional
In mathematics, the dimension of a vector space ''V'' is the cardinality (i.e., the number of vectors) of a basis of ''V'' over its base field. p. 44, §2.36 It is sometimes called Hamel dimension (after Georg Hamel) or algebraic dimension to disti ...
vector space a
linear map is called diagonalizable if there exists an
ordered basis of
consisting of
eigenvectors of
. These definitions are equivalent: if
has a
matrix
Matrix most commonly refers to:
* ''The Matrix'' (franchise), an American media franchise
** '' The Matrix'', a 1999 science-fiction action film
** "The Matrix", a fictional setting, a virtual reality environment, within ''The Matrix'' (franchi ...
representation
as above, then the column vectors of
form a basis consisting of eigenvectors of and the diagonal entries of
are the corresponding
eigenvalues of with respect to this eigenvector basis,
is represented by Diagonalization is the process of finding the above
and
Diagonalizable matrices and maps are especially easy for computations, once their eigenvalues and eigenvectors are known. One can raise a diagonal matrix
to a power by simply raising the diagonal entries to that power, and the
determinant of a diagonal matrix is simply the product of all diagonal entries; such computations generalize easily to Geometrically, a diagonalizable matrix is an
inhomogeneous dilation
In affine geometry, uniform scaling (or isotropic scaling) is a linear transformation that enlarges (increases) or shrinks (diminishes) objects by a ''scale factor'' that is the same in all directions. The result of uniform scaling is similar ...
(or ''anisotropic scaling'') — it
scales the space, as does a ''
homogeneous dilation
In mathematics, a homothety (or homothecy, or homogeneous dilation) is a transformation of an affine space determined by a point ''S'' called its ''center'' and a nonzero number ''k'' called its ''ratio'', which sends point X to a point X' by th ...
'', but by a different factor along each eigenvector axis, the factor given by the corresponding eigenvalue.
A square matrix that is not diagonalizable is called ''
defective''. It can happen that a matrix
with
real entries is defective over the real numbers, meaning that
is impossible for any invertible
and diagonal
with real entries, but it is possible with
complex entries, so that
is diagonalizable over the complex numbers. For example, this is the case for a generic
rotation matrix.
Many results for diagonalizable matrices hold only over an
algebraically closed field
In mathematics, a field is algebraically closed if every non-constant polynomial in (the univariate polynomial ring with coefficients in ) has a root in .
Examples
As an example, the field of real numbers is not algebraically closed, because ...
(such as the complex numbers). In this case, diagonalizable matrices are
dense in the space of all matrices, which means any defective matrix can be deformed into a diagonalizable matrix by a small
perturbation; and the
Jordan normal form theorem states that any matrix is uniquely the sum of a diagonalizable matrix and a
nilpotent matrix. Over an algebraically closed field, diagonalizable matrices are equivalent to
semi-simple matrices.
Definition
A square
matrix,
, with entries in a
field is called diagonalizable or nondefective if there exists an
invertible matrix (i.e. an element of the
general linear group GL
''n''(''F'')),
, such that
is a diagonal matrix. Formally,
Characterization
The fundamental fact about diagonalizable maps and matrices is expressed by the following:
* An
matrix
over a field
is diagonalizable
if and only if the sum of the
dimensions of its eigenspaces is equal to
, which is the case if and only if there exists a
basis of
consisting of eigenvectors of
. If such a basis has been found, one can form the matrix
having these
basis vectors as columns, and
will be a diagonal matrix whose diagonal entries are the eigenvalues of
. The matrix
is known as a
modal matrix for
.
* A linear map
is diagonalizable if and only if the sum of the
dimensions of its eigenspaces is equal to which is the case if and only if there exists a basis of
consisting of eigenvectors of
. With respect to such a basis,
will be represented by a diagonal matrix. The diagonal entries of this matrix are the eigenvalues of
Another characterization: A matrix or linear map is diagonalizable over the field
if and only if its
minimal polynomial is a product of distinct linear factors over (Put another way, a matrix is diagonalizable if and only if all of its
elementary divisor In algebra, the elementary divisors of a module over a principal ideal domain (PID) occur in one form of the structure theorem for finitely generated modules over a principal ideal domain.
If R is a PID and M a finitely generated R-module, then '' ...
s are linear.)
The following sufficient (but not necessary) condition is often useful.
* An
matrix
is diagonalizable over the field
if it has
distinct eigenvalues in i.e. if its
characteristic polynomial
In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The chara ...
has
distinct roots in however, the converse may be false. Consider
which has eigenvalues 1, 2, 2 (not all distinct) and is diagonalizable with diagonal form (
similar to
and
change of basis matrix :
The converse fails when
has an eigenspace of dimension higher than 1. In this example, the eigenspace of
associated with the eigenvalue 2 has dimension 2.
* A linear map
with
is diagonalizable if it has
distinct eigenvalues, i.e. if its characteristic polynomial has
distinct roots in
.
Let
be a matrix over If
is diagonalizable, then so is any power of it. Conversely, if
is invertible,
is algebraically closed, and
is diagonalizable for some
that is not an integer multiple of the characteristic of then
is diagonalizable. Proof: If
is diagonalizable, then
is annihilated by some polynomial which has no multiple root (since and is divided by the minimal polynomial of
Over the complex numbers
, almost every matrix is diagonalizable. More precisely: the set of complex
matrices that are ''not'' diagonalizable over considered as a
subset
In mathematics, Set (mathematics), set ''A'' is a subset of a set ''B'' if all Element (mathematics), elements of ''A'' are also elements of ''B''; ''B'' is then a superset of ''A''. It is possible for ''A'' and ''B'' to be equal; if they are ...
of has
Lebesgue measure
In measure theory, a branch of mathematics, the Lebesgue measure, named after French mathematician Henri Lebesgue, is the standard way of assigning a measure to subsets of ''n''-dimensional Euclidean space. For ''n'' = 1, 2, or 3, it coincides wit ...
zero. One can also say that the diagonalizable matrices form a dense subset with respect to the
Zariski topology: the non-diagonalizable matrices lie inside the
vanishing set of the
discriminant
In mathematics, the discriminant of a polynomial is a quantity that depends on the coefficients and allows deducing some properties of the roots without computing them. More precisely, it is a polynomial function of the coefficients of the origi ...
of the characteristic polynomial, which is a
hypersurface. From that follows also density in the usual (''strong'') topology given by a
norm. The same is not true over
The
Jordan–Chevalley decomposition expresses an operator as the sum of its semisimple (i.e., diagonalizable) part and its
nilpotent part. Hence, a matrix is diagonalizable if and only if its nilpotent part is zero. Put in another way, a matrix is diagonalizable if each block in its
Jordan form has no nilpotent part; i.e., each "block" is a one-by-one matrix.
Diagonalization

If a matrix
can be diagonalized, that is,
:
then:
:
Writing
as a
block matrix
In mathematics, a block matrix or a partitioned matrix is a matrix that is '' interpreted'' as having been broken into sections called blocks or submatrices. Intuitively, a matrix interpreted as a block matrix can be visualized as the original mat ...
of its column vectors
:
the above equation can be rewritten as
:
So the column vectors of
are
right eigenvector
In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
s of and the corresponding diagonal entry is the corresponding
eigenvalue. The invertibility of
also suggests that the eigenvectors are
linearly independent and form a basis of This is the necessary and sufficient condition for diagonalizability and the canonical approach of diagonalization. The
row vectors of
are the
left eigenvectors of
When a complex matrix
is a
Hermitian matrix (or more generally a
normal matrix), eigenvectors of
can be chosen to form an
orthonormal basis of and
can be chosen to be a
unitary matrix. If in addition,
is a real
symmetric matrix, then its eigenvectors can be chosen to be an orthonormal basis of
and
can be chosen to be an
orthogonal matrix.
For most practical work matrices are diagonalized numerically using computer software.
Many algorithms exist to accomplish this.
Simultaneous diagonalization
A set of matrices is said to be ''simultaneously diagonalizable'' if there exists a single invertible matrix
such that
is a diagonal matrix for every
in the set. The following theorem characterizes simultaneously diagonalizable matrices: A set of diagonalizable
matrices commutes if and only if the set is simultaneously diagonalizable.
The set of all
diagonalizable matrices (over with
is not simultaneously diagonalizable. For instance, the matrices
:
are diagonalizable but not simultaneously diagonalizable because they do not commute.
A set consists of commuting
normal matrices if and only if it is simultaneously diagonalizable by a
unitary matrix; that is, there exists a unitary matrix
such that
is diagonal for every
in the set.
In the language of
Lie theory, a set of simultaneously diagonalizable matrices generate a
toral Lie algebra
In mathematics, a toral subalgebra is a Lie subalgebra of a general linear Lie algebra all of whose elements are semisimple (or diagonalizable over an algebraically closed field). Equivalently, a Lie algebra is toral if it contains no nonzero n ...
.
Examples
Diagonalizable matrices
*
Involutions are diagonalizable over the reals (and indeed any field of characteristic not 2), with ±1 on the diagonal.
* Finite order
endomorphisms are diagonalizable over
(or any algebraically closed field where the characteristic of the field does not divide the order of the endomorphism) with
roots of unity on the diagonal. This follows since the minimal polynomial is
separable, because the roots of unity are distinct.
*
Projections are diagonalizable, with 0s and 1s on the diagonal.
* Real
symmetric matrices are diagonalizable by
orthogonal matrices
In linear algebra, an orthogonal matrix, or orthonormal matrix, is a real square matrix whose columns and rows are orthonormal vectors.
One way to express this is
Q^\mathrm Q = Q Q^\mathrm = I,
where is the transpose of and is the identity ma ...
; i.e., given a real symmetric matrix
is diagonal for some orthogonal matrix More generally, matrices are diagonalizable by
unitary matrices if and only if they are
normal. In the case of the real symmetric matrix, we see that so clearly
holds. Examples of normal matrices are real symmetric (or
skew-symmetric) matrices (e.g. covariance matrices) and
Hermitian matrices (or skew-Hermitian matrices). See
spectral theorems for generalizations to infinite-dimensional vector spaces.
Matrices that are not diagonalizable
In general, a
rotation matrix is not diagonalizable over the reals, but all
rotation matrices are diagonalizable over the complex field. Even if a matrix is not diagonalizable, it is always possible to "do the best one can", and find a matrix with the same properties consisting of eigenvalues on the leading diagonal, and either ones or zeroes on the superdiagonal – known as
Jordan normal form.
Some matrices are not diagonalizable over any field, most notably nonzero
nilpotent matrices. This happens more generally if the
algebraic and geometric multiplicities of an eigenvalue do not coincide. For instance, consider
:
This matrix is not diagonalizable: there is no matrix
such that
is a diagonal matrix. Indeed,
has one eigenvalue (namely zero) and this eigenvalue has algebraic multiplicity 2 and geometric multiplicity 1.
Some real matrices are not diagonalizable over the reals. Consider for instance the matrix
:
The matrix
does not have any real eigenvalues, so there is no real matrix
such that
is a diagonal matrix. However, we can diagonalize
if we allow complex numbers. Indeed, if we take
:
then
is diagonal. It is easy to find that
is the rotation matrix which rotates counterclockwise by angle
Note that the above examples show that the sum of diagonalizable matrices need not be diagonalizable.
How to diagonalize a matrix
Diagonalizing a matrix is the same process as finding its
eigenvalues and eigenvectors, in the case that the eigenvectors form a basis. For example, consider the matrix
:
The roots of the
characteristic polynomial
In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The chara ...
are the eigenvalues Solving the linear system
gives the eigenvectors
and while
gives that is,
for These vectors form a basis of so we can assemble them as the column vectors of a
change-of-basis matrix
to get:
We may see this equation in terms of transformations:
takes the standard basis to the eigenbasis, so we have:
so that
has the standard basis as its eigenvectors, which is the defining property of
Note that there is no preferred order of the eigenvectors in changing the order of the
eigenvectors in
just changes the order of the
eigenvalues
In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted b ...
in the diagonalized form of
Application to matrix functions
Diagonalization can be used to efficiently compute the powers of a matrix
:
and the latter is easy to calculate since it only involves the powers of a diagonal matrix. For example, for the matrix
with eigenvalues
in the example above we compute:
:
This approach can be generalized to
matrix exponential and other
matrix functions that can be defined as power series. For example, defining we have:
:
This is particularly useful in finding closed form expressions for terms of
linear recursive sequences
In mathematics (including combinatorics, linear algebra, and dynamical systems), a linear recurrence with constant coefficients (also known as a linear recurrence relation or linear difference equation) sets equal to 0 a polynomial that is linear ...
, such as the
Fibonacci numbers.
Particular application
For example, consider the following matrix:
:
Calculating the various powers of
reveals a surprising pattern:
:
The above phenomenon can be explained by diagonalizing To accomplish this, we need a basis of
consisting of eigenvectors of One such eigenvector basis is given by
:
where e
''i'' denotes the standard basis of R
''n''. The reverse change of basis is given by
:
Straightforward calculations show that
:
Thus, ''a'' and ''b'' are the eigenvalues corresponding to u and v, respectively. By linearity of matrix multiplication, we have that
:
Switching back to the standard basis, we have
:
The preceding relations, expressed in matrix form, are
:
thereby explaining the above phenomenon.
Quantum mechanical application
In
quantum mechanical and
quantum chemical computations matrix diagonalization is one of the most frequently applied numerical processes. The basic reason is that the time-independent
Schrödinger equation is an eigenvalue equation, albeit in most of the physical situations on an infinite dimensional space (a
Hilbert space
In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise natural ...
).
A very common approximation is to truncate Hilbert space to finite dimension, after which the Schrödinger equation can be formulated as an eigenvalue problem of a real symmetric, or complex Hermitian matrix. Formally this approximation is founded on the
variational principle, valid for Hamiltonians that are bounded from below.
First-order perturbation theory also leads to matrix eigenvalue problem for degenerate states.
See also
*
Defective matrix
*
Scaling (geometry)
*
Triangular matrix
*
Semisimple operator
*
Diagonalizable group In mathematics, an affine algebraic group is said to be diagonalizable if it is isomorphic to a subgroup of D''n'', the group of diagonal matrices. A diagonalizable group defined over a field ''k'' is said to ''split over'' ''k'' or ''k''-''split'' ...
*
Jordan normal form
*
Weight module – associative algebra generalization
*
Orthogonal diagonalization
Notes
References
{{Matrix classes
Matrices