In
linear algebra
Linear algebra is the branch of mathematics concerning linear equations such as
:a_1x_1+\cdots +a_nx_n=b,
linear maps such as
:(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n,
and their representations in vector spaces and through matrix (mathemat ...
, a
square matrix
In mathematics, a square matrix is a Matrix (mathematics), matrix with the same number of rows and columns. An ''n''-by-''n'' matrix is known as a square matrix of order Any two square matrices of the same order can be added and multiplied.
Squ ...
is called diagonalizable or non-defective if it is
similar to a
diagonal matrix
In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagon ...
. That is, if there exists an
invertible matrix
In linear algebra, an invertible matrix (''non-singular'', ''non-degenarate'' or ''regular'') is a square matrix that has an inverse. In other words, if some other matrix is multiplied by the invertible matrix, the result can be multiplied by a ...
and a diagonal matrix
such that . This is equivalent to (Such
are not unique.) This property exists for any linear map: for a
finite-dimensional
In mathematics, the dimension of a vector space ''V'' is the cardinality (i.e., the number of vectors) of a basis of ''V'' over its base field. p. 44, §2.36 It is sometimes called Hamel dimension (after Georg Hamel) or algebraic dimension to d ...
vector space
In mathematics and physics, a vector space (also called a linear space) is a set (mathematics), set whose elements, often called vector (mathematics and physics), ''vectors'', can be added together and multiplied ("scaled") by numbers called sc ...
a
linear map
In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that p ...
is called diagonalizable if there exists an
ordered basis
Order, ORDER or Orders may refer to:
* A socio-political or established or existing order, e.g. World order, Ancien Regime, Pax Britannica
* Categorization, the process in which ideas and objects are recognized, differentiated, and understood
* ...
of
consisting of
eigenvector
In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by ...
s of
. These definitions are equivalent: if
has a
matrix
Matrix (: matrices or matrixes) or MATRIX may refer to:
Science and mathematics
* Matrix (mathematics), a rectangular array of numbers, symbols or expressions
* Matrix (logic), part of a formula in prenex normal form
* Matrix (biology), the m ...
representation
as above, then the column vectors of
form a basis consisting of eigenvectors of and the diagonal entries of
are the corresponding
eigenvalue
In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
s of with respect to this eigenvector basis,
is represented by
Diagonalization is the process of finding the above
and and makes many subsequent computations easier. One can raise a diagonal matrix
to a power by simply raising the diagonal entries to that power. The
determinant
In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of a diagonal matrix is simply the product of all diagonal entries. Such computations generalize easily to
The geometric transformation represented by a diagonalizable matrix is an ''
inhomogeneous dilation'' (or ''anisotropic scaling''). That is, it can
scale the space by a different amount in different directions. The direction of each eigenvector is scaled by a factor given by the corresponding eigenvalue.
A square matrix that is not diagonalizable is called ''
defective''. It can happen that a matrix
with
real entries is defective over the real numbers, meaning that
is impossible for any invertible
and diagonal
with real entries, but it is possible with
complex
Complex commonly refers to:
* Complexity, the behaviour of a system whose components interact in multiple ways so possible interactions are difficult to describe
** Complex system, a system composed of many components which may interact with each ...
entries, so that
is diagonalizable over the complex numbers. For example, this is the case for a generic
rotation matrix
In linear algebra, a rotation matrix is a transformation matrix that is used to perform a rotation (mathematics), rotation in Euclidean space. For example, using the convention below, the matrix
:R = \begin
\cos \theta & -\sin \theta \\
\sin \t ...
.
Many results for diagonalizable matrices hold only over an
algebraically closed field
In mathematics, a field is algebraically closed if every non-constant polynomial in (the univariate polynomial ring with coefficients in ) has a root in . In other words, a field is algebraically closed if the fundamental theorem of algebra ...
(such as the complex numbers). In this case, diagonalizable matrices are
dense
Density (volumetric mass density or specific mass) is the ratio of a substance's mass to its volume. The symbol most often used for density is ''ρ'' (the lower case Greek letter rho), although the Latin letter ''D'' (or ''d'') can also be use ...
in the space of all matrices, which means any defective matrix can be deformed into a diagonalizable matrix by a small
perturbation; and the
Jordan–Chevalley decomposition states that any matrix is uniquely the sum of a diagonalizable matrix and a
nilpotent matrix
In linear algebra, a nilpotent matrix is a square matrix ''N'' such that
:N^k = 0\,
for some positive integer k. The smallest such k is called the index of N, sometimes the degree of N.
More generally, a nilpotent transformation is a linear trans ...
. Over an algebraically closed field, diagonalizable matrices are equivalent to
semi-simple matrices.
Definition
A square
matrix
with entries in a
field is called diagonalizable or nondefective if there exists an
invertible matrix (i.e. an element of the
general linear group
In mathematics, the general linear group of degree n is the set of n\times n invertible matrices, together with the operation of ordinary matrix multiplication. This forms a group, because the product of two invertible matrices is again inve ...
GL
''n''(''F'')),
, such that
is a diagonal matrix.
Characterization
The fundamental fact about diagonalizable maps and matrices is expressed by the following:
* An
matrix
over a field
is diagonalizable
if and only if
In logic and related fields such as mathematics and philosophy, "if and only if" (often shortened as "iff") is paraphrased by the biconditional, a logical connective between statements. The biconditional is true in two cases, where either bo ...
the sum of the
dimension
In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it. Thus, a line has a dimension of one (1D) because only one coo ...
s of its eigenspaces is equal to
, which is the case if and only if there exists a
basis of
consisting of eigenvectors of
. If such a basis has been found, one can form the matrix
having these
basis vectors as columns, and
will be a diagonal matrix whose diagonal entries are the eigenvalues of
. The matrix
is known as a
modal matrix for
.
* A linear map
is diagonalizable if and only if the sum of the
dimension
In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it. Thus, a line has a dimension of one (1D) because only one coo ...
s of its eigenspaces is equal to which is the case if and only if there exists a basis of
consisting of eigenvectors of
. With respect to such a basis,
will be represented by a diagonal matrix. The diagonal entries of this matrix are the eigenvalues of
The following sufficient (but not necessary) condition is often useful.
* An
matrix
is diagonalizable over the field
if it has
distinct eigenvalues in i.e. if its
characteristic polynomial
In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The ...
has
distinct roots in however, the converse may be false. Consider
which has eigenvalues 1, 2, 2 (not all distinct) and is diagonalizable with diagonal form (
similar to
and
change of basis matrix :
The converse fails when
has an eigenspace of dimension higher than 1. In this example, the eigenspace of
associated with the eigenvalue 2 has dimension 2.
* A linear map
with
is diagonalizable if it has
distinct eigenvalues, i.e. if its characteristic polynomial has
distinct roots in
.
Let
be a matrix over If
is diagonalizable, then so is any power of it. Conversely, if
is invertible,
is algebraically closed, and
is diagonalizable for some
that is not an integer multiple of the characteristic of then
is diagonalizable. Proof: If
is diagonalizable, then
is annihilated by some polynomial which has no multiple root (since and is divided by the minimal polynomial of
Over the complex numbers
, almost every matrix is diagonalizable. More precisely: the set of complex
matrices that are ''not'' diagonalizable over considered as a
subset
In mathematics, a Set (mathematics), set ''A'' is a subset of a set ''B'' if all Element (mathematics), elements of ''A'' are also elements of ''B''; ''B'' is then a superset of ''A''. It is possible for ''A'' and ''B'' to be equal; if they a ...
of has
Lebesgue measure
In measure theory, a branch of mathematics, the Lebesgue measure, named after French mathematician Henri Lebesgue, is the standard way of assigning a measure to subsets of higher dimensional Euclidean '-spaces. For lower dimensions or , it c ...
zero. One can also say that the diagonalizable matrices form a dense subset with respect to the
Zariski topology
In algebraic geometry and commutative algebra, the Zariski topology is a topology defined on geometric objects called varieties. It is very different from topologies that are commonly used in real or complex analysis; in particular, it is not ...
: the non-diagonalizable matrices lie inside the
vanishing set of the
discriminant
In mathematics, the discriminant of a polynomial is a quantity that depends on the coefficients and allows deducing some properties of the zero of a function, roots without computing them. More precisely, it is a polynomial function of the coef ...
of the characteristic polynomial, which is a
hypersurface
In geometry, a hypersurface is a generalization of the concepts of hyperplane, plane curve, and surface. A hypersurface is a manifold or an algebraic variety of dimension , which is embedded in an ambient space of dimension , generally a Euclidea ...
. From that follows also density in the usual (''strong'') topology given by a
norm. The same is not true over
The
Jordan–Chevalley decomposition expresses an operator as the sum of its semisimple (i.e., diagonalizable) part and its
nilpotent
In mathematics, an element x of a ring (mathematics), ring R is called nilpotent if there exists some positive integer n, called the index (or sometimes the degree), such that x^n=0.
The term, along with its sister Idempotent (ring theory), idem ...
part. Hence, a matrix is diagonalizable if and only if its nilpotent part is zero. Put in another way, a matrix is diagonalizable if each block in its
Jordan form
\begin
\lambda_1 1\hphantom\hphantom\\
\hphantom\lambda_1 1\hphantom\\
\hphantom\lambda_1\hphantom\\
\hphantom\lambda_2 1\hphantom\hphantom\\
\hphantom\hphantom\lambda_2\hphantom\\
\hphantom\lambda_3\hphantom\\
\hphantom\ddots\hphantom\\
...
has no nilpotent part; i.e., each "block" is a one-by-one matrix.
Diagonalization
Consider the two following arbitrary bases
and
. Suppose that there exists a linear transformation represented by a matrix
which is written with respect to basis E. Suppose also that there exists the following eigen-equation:
The alpha eigenvectors are written also with respect to the E basis. Since the set F is both a set of eigenvectors for matrix A and it spans some arbitrary vector space, then we say that there exists a matrix
which is a diagonal matrix that is similar to
. In other words,
is a diagonalizable matrix if the matrix is written in the basis F. We perform the change of basis calculation using the transition matrix
, which changes basis from E to F as follows:
,
where
is the transition matrix from E-basis to F-basis. The inverse can then be equated to a new transition matrix
which changes basis from F to E instead and so we have the following relationship :
Both
and
transition matrices are invertible. Thus we can manipulate the matrices in the following fashion:
The matrix
will be denoted as
, which is still in the E-basis. Similarly, the diagonal matrix is in the F-basis.

If a matrix
can be diagonalized, that is,
:
then:
:
The transition matrix S has the E-basis vectors as columns written in the basis F. Inversely, the inverse transition matrix P has F-basis vectors
written in the basis of E so that we can represent P in block matrix form in the following manner:
:
as a result we can write:
In block matrix form, we can consider the A-matrix to be a matrix of 1x1 dimensions whilst P is a 1xn dimensional matrix. The D-matrix can be written in full form with all the diagonal elements as an nxn dimensional matrix:
Performing the above matrix multiplication we end up with the following result:
Taking each component of the block matrix individually on both sides, we end up with the following:
:
So the column vectors of
are
right eigenvectors of and the corresponding diagonal entry is the corresponding
eigenvalue
In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
. The invertibility of
also suggests that the eigenvectors are
linearly independent
In the theory of vector spaces, a set of vectors is said to be if there exists no nontrivial linear combination of the vectors that equals the zero vector. If such a linear combination exists, then the vectors are said to be . These concep ...
and form a basis of This is the necessary and sufficient condition for diagonalizability and the canonical approach of diagonalization. The
row vector
In linear algebra, a column vector with elements is an m \times 1 matrix consisting of a single column of entries, for example,
\boldsymbol = \begin x_1 \\ x_2 \\ \vdots \\ x_m \end.
Similarly, a row vector is a 1 \times n matrix for some , co ...
s of
are the
left eigenvectors of
When a complex matrix
is a
Hermitian matrix
In mathematics, a Hermitian matrix (or self-adjoint matrix) is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the -th row and -th column is equal to the complex conjugate of the element in the ...
(or more generally a
normal matrix
In mathematics, a complex square matrix is normal if it commutes with its conjugate transpose :
:A \text \iff A^*A = AA^* .
The concept of normal matrices can be extended to normal operators on infinite-dimensional normed spaces and to nor ...
), eigenvectors of
can be chosen to form an
orthonormal basis
In mathematics, particularly linear algebra, an orthonormal basis for an inner product space V with finite Dimension (linear algebra), dimension is a Basis (linear algebra), basis for V whose vectors are orthonormal, that is, they are all unit vec ...
of and
can be chosen to be a
unitary matrix
In linear algebra, an invertible complex square matrix is unitary if its matrix inverse equals its conjugate transpose , that is, if
U^* U = UU^* = I,
where is the identity matrix.
In physics, especially in quantum mechanics, the conjugate ...
. If in addition,
is a real
symmetric matrix
In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally,
Because equal matrices have equal dimensions, only square matrices can be symmetric.
The entries of a symmetric matrix are symmetric with ...
, then its eigenvectors can be chosen to be an orthonormal basis of
and
can be chosen to be an
orthogonal matrix
In linear algebra, an orthogonal matrix, or orthonormal matrix, is a real square matrix whose columns and rows are orthonormal vectors.
One way to express this is
Q^\mathrm Q = Q Q^\mathrm = I,
where is the transpose of and is the identi ...
.
For most practical work matrices are diagonalized numerically using computer software.
Many algorithms exist to accomplish this.
Simultaneous diagonalization
A set of matrices is said to be ''simultaneously diagonalizable'' if there exists a single invertible matrix
such that
is a diagonal matrix for every
in the set. The following theorem characterizes simultaneously diagonalizable matrices: A set of diagonalizable
matrices commutes if and only if the set is simultaneously diagonalizable.
The set of all
diagonalizable matrices (over with
is not simultaneously diagonalizable. For instance, the matrices
:
are diagonalizable but not simultaneously diagonalizable because they do not commute.
A set consists of commuting
normal matrices
Normal(s) or The Normal(s) may refer to:
Film and television
* ''Normal'' (2003 film), starring Jessica Lange and Tom Wilkinson
* ''Normal'' (2007 film), starring Carrie-Anne Moss, Kevin Zegers, Callum Keith Rennie, and Andrew Airlie
* ''Norma ...
if and only if it is simultaneously diagonalizable by a
unitary matrix
In linear algebra, an invertible complex square matrix is unitary if its matrix inverse equals its conjugate transpose , that is, if
U^* U = UU^* = I,
where is the identity matrix.
In physics, especially in quantum mechanics, the conjugate ...
; that is, there exists a unitary matrix
such that
is diagonal for every
in the set.
In the language of
Lie theory
In mathematics, the mathematician Sophus Lie ( ) initiated lines of study involving integration of differential equations, transformation groups, and contact (mathematics), contact of spheres that have come to be called Lie theory. For instance, ...
, a set of simultaneously diagonalizable matrices generates a
toral Lie algebra.
Examples
Diagonalizable matrices
*
Involution
Involution may refer to: Mathematics
* Involution (mathematics), a function that is its own inverse
* Involution algebra, a *-algebra: a type of algebraic structure
* Involute, a construction in the differential geometry of curves
* Exponentiati ...
s are diagonalizable over the reals (and indeed any field of characteristic not 2), with ±1 on the diagonal.
* Finite order
endomorphism
In mathematics, an endomorphism is a morphism from a mathematical object to itself. An endomorphism that is also an isomorphism is an automorphism. For example, an endomorphism of a vector space is a linear map , and an endomorphism of a g ...
s are diagonalizable over
(or any algebraically closed field where the characteristic of the field does not divide the order of the endomorphism) with
roots of unity
In mathematics, a root of unity is any complex number that yields 1 when raised to some positive integer power . Roots of unity are used in many branches of mathematics, and are especially important in number theory, the theory of group char ...
on the diagonal. This follows since the minimal polynomial is
separable, because the roots of unity are distinct.
*
Projections are diagonalizable, with 0s and 1s on the diagonal.
* Real
symmetric matrices
In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally,
Because equal matrices have equal dimensions, only square matrices can be symmetric.
The entries of a symmetric matrix are symmetric with re ...
are diagonalizable by
orthogonal matrices; i.e., given a real symmetric matrix
is diagonal for some orthogonal matrix More generally, matrices are diagonalizable by
unitary matrices if and only if they are
normal. In the case of the real symmetric matrix, we see that so clearly
holds. Examples of normal matrices are real symmetric (or
skew-symmetric) matrices (e.g. covariance matrices) and
Hermitian matrices (or skew-Hermitian matrices). See
spectral theorem
In linear algebra and functional analysis, a spectral theorem is a result about when a linear operator or matrix can be diagonalized (that is, represented as a diagonal matrix in some basis). This is extremely useful because computations involvin ...
s for generalizations to infinite-dimensional vector spaces.
Matrices that are not diagonalizable
In general, a
rotation matrix
In linear algebra, a rotation matrix is a transformation matrix that is used to perform a rotation (mathematics), rotation in Euclidean space. For example, using the convention below, the matrix
:R = \begin
\cos \theta & -\sin \theta \\
\sin \t ...
is not diagonalizable over the reals, but all
rotation matrices are diagonalizable over the complex field. Even if a matrix is not diagonalizable, it is always possible to "do the best one can", and find a matrix with the same properties consisting of eigenvalues on the leading diagonal, and either ones or zeroes on the superdiagonal – known as
Jordan normal form
\begin
\lambda_1 1\hphantom\hphantom\\
\hphantom\lambda_1 1\hphantom\\
\hphantom\lambda_1\hphantom\\
\hphantom\lambda_2 1\hphantom\hphantom\\
\hphantom\hphantom\lambda_2\hphantom\\
\hphantom\lambda_3\hphantom\\
\hphantom\ddots\hphantom\\
...
.
Some matrices are not diagonalizable over any field, most notably nonzero
nilpotent matrices. This happens more generally if the
algebraic and geometric multiplicities of an eigenvalue do not coincide. For instance, consider
:
This matrix is not diagonalizable: there is no matrix
such that
is a diagonal matrix. Indeed,
has one eigenvalue (namely zero) and this eigenvalue has algebraic multiplicity 2 and geometric multiplicity 1.
Some real matrices are not diagonalizable over the reals. Consider for instance the matrix
:
The matrix
does not have any real eigenvalues, so there is no real matrix
such that
is a diagonal matrix. However, we can diagonalize
if we allow complex numbers. Indeed, if we take
:
then
is diagonal. It is easy to find that
is the rotation matrix which rotates counterclockwise by angle
Note that the above examples show that the sum of diagonalizable matrices need not be diagonalizable.
How to diagonalize a matrix
Diagonalizing a matrix is the same process as finding its
eigenvalues and eigenvectors
In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
, in the case that the eigenvectors form a basis. For example, consider the matrix
:
The roots of the
characteristic polynomial
In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The ...
are the eigenvalues Solving the linear system
gives the eigenvectors
and while
gives that is,
for These vectors form a basis of so we can assemble them as the column vectors of a
change-of-basis matrix
to get:
We may see this equation in terms of transformations:
takes the standard basis to the eigenbasis, so we have:
so that
has the standard basis as its eigenvectors, which is the defining property of
Note that there is no preferred order of the eigenvectors in changing the order of the
eigenvectors
In linear algebra, an eigenvector ( ) or characteristic vector is a Vector (mathematics and physics), vector that has its direction (geometry), direction unchanged (or reversed) by a given linear map, linear transformation. More precisely, an e ...
in
just changes the order of the
eigenvalues
In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
in the diagonalized form of
Application to matrix functions
Diagonalization can be used to efficiently compute the powers of a matrix
:
and the latter is easy to calculate since it only involves the powers of a diagonal matrix. For example, for the matrix
with eigenvalues
in the example above we compute:
:
This approach can be generalized to
matrix exponential
In mathematics, the matrix exponential is a matrix function on square matrix, square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exp ...
and other
matrix functions that can be defined as power series. For example, defining we have:
:
This is particularly useful in finding closed form expressions for terms of
linear recursive sequences, such as the
Fibonacci numbers
In mathematics, the Fibonacci sequence is a sequence in which each element is the sum of the two elements that precede it. Numbers that are part of the Fibonacci sequence are known as Fibonacci numbers, commonly denoted . Many writers begin the s ...
.
Particular application
For example, consider the following matrix:
:
Calculating the various powers of
reveals a surprising pattern:
:
The above phenomenon can be explained by diagonalizing To accomplish this, we need a basis of
consisting of eigenvectors of One such eigenvector basis is given by
:
where e
''i'' denotes the standard basis of R
''n''. The reverse change of basis is given by
:
Straightforward calculations show that
:
Thus, ''a'' and ''b'' are the eigenvalues corresponding to u and v, respectively. By linearity of matrix multiplication, we have that
:
Switching back to the standard basis, we have
:
The preceding relations, expressed in matrix form, are
:
thereby explaining the above phenomenon.
Quantum mechanical application
In
quantum mechanical
Quantum mechanics is the fundamental physical theory that describes the behavior of matter and of light; its unusual characteristics typically occur at and below the scale of atoms. Reprinted, Addison-Wesley, 1989, It is the foundation of a ...
and
quantum chemical computations matrix diagonalization is one of the most frequently applied numerical processes. The basic reason is that the time-independent
Schrödinger equation
The Schrödinger equation is a partial differential equation that governs the wave function of a non-relativistic quantum-mechanical system. Its discovery was a significant landmark in the development of quantum mechanics. It is named after E ...
is an eigenvalue equation, albeit in most of the physical situations on an infinite dimensional
Hilbert space
In mathematics, a Hilbert space is a real number, real or complex number, complex inner product space that is also a complete metric space with respect to the metric induced by the inner product. It generalizes the notion of Euclidean space. The ...
.
A very common approximation is to truncate (or project) the Hilbert space to finite dimension, after which the Schrödinger equation can be formulated as an eigenvalue problem of a real symmetric, or complex Hermitian matrix. Formally this approximation is founded on the
variational principle, valid for Hamiltonians that are bounded from below.
First-order perturbation theory also leads to matrix eigenvalue problem for degenerate states.
See also
*
Defective matrix
In linear algebra, a defective matrix is a square matrix that does not have a complete basis of eigenvectors, and is therefore not diagonalizable. In particular, an n \times n matrix is defective if and only if it does not have n linearly indepe ...
*
Scaling (geometry)
In affine geometry, uniform scaling (or isotropic scaling) is a linear transformation that enlarges (increases) or shrinks (diminishes) objects by a '' scale factor'' that is the same in all directions ( isotropically). The result of uniform s ...
*
Triangular matrix
In mathematics, a triangular matrix is a special kind of square matrix. A square matrix is called if all the entries ''above'' the main diagonal are zero. Similarly, a square matrix is called if all the entries ''below'' the main diagonal are z ...
*
Semisimple operator
*
Diagonalizable group
*
Jordan normal form
\begin
\lambda_1 1\hphantom\hphantom\\
\hphantom\lambda_1 1\hphantom\\
\hphantom\lambda_1\hphantom\\
\hphantom\lambda_2 1\hphantom\hphantom\\
\hphantom\hphantom\lambda_2\hphantom\\
\hphantom\lambda_3\hphantom\\
\hphantom\ddots\hphantom\\
...
*
Weight module – associative algebra generalization
*
Orthogonal diagonalization
Notes
References
{{Matrix classes
Matrices (mathematics)