HOME

TheInfoList



OR:

In
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
, an invertible matrix (''non-singular'', ''non-degenarate'' or ''regular'') is a square matrix that has an inverse. In other words, if some other
matrix Matrix (: matrices or matrixes) or MATRIX may refer to: Science and mathematics * Matrix (mathematics), a rectangular array of numbers, symbols or expressions * Matrix (logic), part of a formula in prenex normal form * Matrix (biology), the m ...
is multiplied by the invertible matrix, the result can be multiplied by an inverse to undo the operation. An invertible matrix multiplied by its inverse yields the
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
. Invertible matrices are the same size as their inverse.


Definition

An -by- square matrix is called invertible if there exists an -by- square matrix such that\mathbf = \mathbf = \mathbf_n ,where denotes the -by-
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
and the multiplication used is ordinary
matrix multiplication In mathematics, specifically in linear algebra, matrix multiplication is a binary operation that produces a matrix (mathematics), matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the n ...
. If this is the case, then the matrix is uniquely determined by , and is called the (multiplicative) ''inverse'' of , denoted by . Matrix inversion is the process of finding the matrix which when multiplied by the original matrix gives the identity matrix. Over a field, a square matrix that is ''not'' invertible is called singular or degenerate. A square matrix with entries in a field is singular
if and only if In logic and related fields such as mathematics and philosophy, "if and only if" (often shortened as "iff") is paraphrased by the biconditional, a logical connective between statements. The biconditional is true in two cases, where either bo ...
its
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
is zero. Singular matrices are rare in the sense that if a square matrix's entries are randomly selected from any bounded region on the
number line A number line is a graphical representation of a straight line that serves as spatial representation of numbers, usually graduated like a ruler with a particular origin point representing the number zero and evenly spaced marks in either dire ...
or
complex plane In mathematics, the complex plane is the plane (geometry), plane formed by the complex numbers, with a Cartesian coordinate system such that the horizontal -axis, called the real axis, is formed by the real numbers, and the vertical -axis, call ...
, the
probability Probability is a branch of mathematics and statistics concerning events and numerical descriptions of how likely they are to occur. The probability of an event is a number between 0 and 1; the larger the probability, the more likely an e ...
that the matrix is singular is 0, that is, it will "almost never" be singular. Non-square matrices, i.e. -by- matrices for which , do not have an inverse. However, in some cases such a matrix may have a left inverse or right inverse. If is -by- and the rank of is equal to , (), then has a left inverse, an -by- matrix such that . If has rank (), then it has a right inverse, an -by- matrix such that . While the most common case is that of matrices over the real or
complex Complex commonly refers to: * Complexity, the behaviour of a system whose components interact in multiple ways so possible interactions are difficult to describe ** Complex system, a system composed of many components which may interact with each ...
numbers, all of those definitions can be given for matrices over any
algebraic structure In mathematics, an algebraic structure or algebraic system consists of a nonempty set ''A'' (called the underlying set, carrier set or domain), a collection of operations on ''A'' (typically binary operations such as addition and multiplicatio ...
equipped with
addition Addition (usually signified by the Plus and minus signs#Plus sign, plus symbol, +) is one of the four basic Operation (mathematics), operations of arithmetic, the other three being subtraction, multiplication, and Division (mathematics), divis ...
and
multiplication Multiplication is one of the four elementary mathematical operations of arithmetic, with the other ones being addition, subtraction, and division (mathematics), division. The result of a multiplication operation is called a ''Product (mathem ...
(i.e. rings). However, in the case of a ring being
commutative In mathematics, a binary operation is commutative if changing the order of the operands does not change the result. It is a fundamental property of many binary operations, and many mathematical proofs depend on it. Perhaps most familiar as a pr ...
, the condition for a square matrix to be invertible is that its determinant is invertible in the ring, which in general is a stricter requirement than it being nonzero. For a noncommutative ring, the usual determinant is not defined. The conditions for existence of left-inverse or right-inverse are more complicated, since a notion of rank does not exist over rings. The set of invertible matrices together with the operation of
matrix multiplication In mathematics, specifically in linear algebra, matrix multiplication is a binary operation that produces a matrix (mathematics), matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the n ...
and entries from ring form a group, the
general linear group In mathematics, the general linear group of degree n is the set of n\times n invertible matrices, together with the operation of ordinary matrix multiplication. This forms a group, because the product of two invertible matrices is again inve ...
of degree , denoted .


Properties


Invertible matrix theorem

Let be a square -by- matrix over a field (e.g., the field of real numbers). The following statements are equivalent, i.e., they are either all true or all false for any given matrix: * is invertible, i.e. it has an inverse under matrix multiplication, i.e., there exists a such that . (In that statement, "invertible" can equivalently be replaced with "left-invertible" or "right-invertible" in which one-sided inverses are considered.) * The linear transformation mapping to is invertible, i.e., it has an inverse under function composition. (There, again, "invertible" can equivalently be replaced with either "left-invertible" or "right-invertible".) * The
transpose In linear algebra, the transpose of a Matrix (mathematics), matrix is an operator which flips a matrix over its diagonal; that is, it switches the row and column indices of the matrix by producing another matrix, often denoted by (among other ...
is an invertible matrix. * is row-equivalent to the -by-
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
. * is column-equivalent to the -by- identity matrix . * has
pivot position The pivot or pivot element is the element of a matrix, or an array, which is selected first by an algorithm (e.g. Gaussian elimination, simplex algorithm, etc.), to do certain calculations. In the case of matrix algorithms, a pivot entry is usu ...
s. * has full rank: . * has a trivial kernel: * The linear transformation mapping to is bijective; that is, the equation has exactly one solution for each in . (There, "bijective" can equivalently be replaced with "
injective In mathematics, an injective function (also known as injection, or one-to-one function ) is a function that maps distinct elements of its domain to distinct elements of its codomain; that is, implies (equivalently by contraposition, impl ...
" or "
surjective In mathematics, a surjective function (also known as surjection, or onto function ) is a function such that, for every element of the function's codomain, there exists one element in the function's domain such that . In other words, for a f ...
".) * The columns of form a basis of . (In this statement, "basis" can equivalently be replaced with either "linearly independent set" or "spanning set") * The rows of form a basis of . (Similarly, here, "basis" can equivalently be replaced with either "linearly independent set" or "spanning set") * The
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of is nonzero: . In general, a square matrix over a
commutative ring In mathematics, a commutative ring is a Ring (mathematics), ring in which the multiplication operation is commutative. The study of commutative rings is called commutative algebra. Complementarily, noncommutative algebra is the study of ring prope ...
is invertible if and only if its determinant is a unit (i.e. multiplicatively invertible element) of that ring. * The number 0 is not an
eigenvalue In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
of . (More generally, a number \lambda is an eigenvalue of if the matrix \mathbf-\lambda \mathbf is singular, where is the identity matrix.) * The matrix can be expressed as a finite product of elementary matrices.


Other properties

Furthermore, the following properties hold for an invertible matrix : * (\mathbf A^)^ = \mathbf A * (k \mathbf A)^ = k^ \mathbf A^ for nonzero scalar * (\mathbf)^+ = \mathbf x^+ \mathbf A^ if has orthonormal columns, where denotes the Moore–Penrose inverse and is a vector * (\mathbf A^\mathrm)^ = (\mathbf A^)^\mathrm * For any invertible -by- matrices and , (\mathbf)^ = \mathbf B^ \mathbf A^. More generally, if \mathbf A_1, \dots, \mathbf A_k are invertible -by- matrices, then (\mathbf A_1 \mathbf A_2 \cdots \mathbf A_ \mathbf A_k)^ = \mathbf A_k^ \mathbf A_^ \cdots \mathbf A_2^ \mathbf A_1^. *\det \mathbf A^ = (\det \mathbf A)^. The rows of the inverse matrix of a matrix are orthonormal to the columns of (and vice versa interchanging rows for columns). To see this, suppose that where the rows of are denoted as v_i^ and the columns of as u_j for 1 \leq i,j \leq n. Then clearly, the Euclidean inner product of any two v_i^ u_j = \delta_. This property can also be useful in constructing the inverse of a square matrix in some instances, where a set of
orthogonal In mathematics, orthogonality (mathematics), orthogonality is the generalization of the geometric notion of ''perpendicularity''. Although many authors use the two terms ''perpendicular'' and ''orthogonal'' interchangeably, the term ''perpendic ...
vectors (but not necessarily orthonormal vectors) to the columns of are known. In which case, one can apply the iterative
Gram–Schmidt process In mathematics, particularly linear algebra and numerical analysis, the Gram–Schmidt process or Gram-Schmidt algorithm is a way of finding a set of two or more vectors that are perpendicular to each other. By technical definition, it is a metho ...
to this initial set to determine the rows of the inverse . A matrix that is its own inverse (i.e., a matrix such that and consequently ) is called an involutory matrix.


In relation to its adjugate

The adjugate of a matrix can be used to find the inverse of as follows: If is an invertible matrix, then : \mathbf^ = \frac \operatorname(\mathbf).


In relation to the identity matrix

It follows from the
associativity In mathematics, the associative property is a property of some binary operations that rearranging the parentheses in an expression will not change the result. In propositional logic, associativity is a Validity (logic), valid rule of replaceme ...
of matrix multiplication that if : \mathbf = \mathbf \ for ''finite square'' matrices and , then also : \mathbf = \mathbf\


Density

Over the field of real numbers, the set of singular -by- matrices, considered as a
subset In mathematics, a Set (mathematics), set ''A'' is a subset of a set ''B'' if all Element (mathematics), elements of ''A'' are also elements of ''B''; ''B'' is then a superset of ''A''. It is possible for ''A'' and ''B'' to be equal; if they a ...
of is a
null set In mathematical analysis, a null set is a Lebesgue measurable set of real numbers that has measure zero. This can be characterized as a set that can be covered by a countable union of intervals of arbitrarily small total length. The notio ...
, that is, has
Lebesgue measure In measure theory, a branch of mathematics, the Lebesgue measure, named after French mathematician Henri Lebesgue, is the standard way of assigning a measure to subsets of higher dimensional Euclidean '-spaces. For lower dimensions or , it c ...
zero. That is true because singular matrices are the roots of the
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
function. It is a
continuous function In mathematics, a continuous function is a function such that a small variation of the argument induces a small variation of the value of the function. This implies there are no abrupt changes in value, known as '' discontinuities''. More preci ...
because it is a
polynomial In mathematics, a polynomial is a Expression (mathematics), mathematical expression consisting of indeterminate (variable), indeterminates (also called variable (mathematics), variables) and coefficients, that involves only the operations of addit ...
in the entries of the matrix. Thus in the language of
measure theory In mathematics, the concept of a measure is a generalization and formalization of geometrical measures (length, area, volume) and other common notions, such as magnitude (mathematics), magnitude, mass, and probability of events. These seemingl ...
,
almost all In mathematics, the term "almost all" means "all but a negligible quantity". More precisely, if X is a set (mathematics), set, "almost all elements of X" means "all elements of X but those in a negligible set, negligible subset of X". The meaning o ...
-by- matrices are invertible. Furthermore, the set of -by- invertible matrices is
open Open or OPEN may refer to: Music * Open (band), Australian pop/rock band * The Open (band), English indie rock band * ''Open'' (Blues Image album), 1969 * ''Open'' (Gerd Dudek, Buschi Niebergall, and Edward Vesala album), 1979 * ''Open'' (Go ...
and dense in the
topological space In mathematics, a topological space is, roughly speaking, a Geometry, geometrical space in which Closeness (mathematics), closeness is defined but cannot necessarily be measured by a numeric Distance (mathematics), distance. More specifically, a to ...
of all -by- matrices. Equivalently, the set of singular matrices is closed and nowhere dense in the space of -by- matrices. In practice, however, non-invertible matrices may be encountered. In numerical calculations, matrices that are invertible but close to a non-invertible matrix may still be problematic and are said to be ill-conditioned.


Examples

This example with rank of is a non-invertible matrix: : \mathbf = \begin 2 & 4\\ 2 & 4 \end . We can see the rank of this 2-by-2 matrix is 1, which is , so it is non-invertible. Consider the following 2-by-2 matrix: : \mathbf = \begin-1 & \tfrac \\ 1 & -1\end . The matrix \mathbf is invertible. To check this, one can compute that \det \mathbf = -\frac , which is non-zero. As an example of a non-invertible, or singular, matrix, consider: : \mathbf = \begin -1 & \tfrac \\ \tfrac & -1 \end . The determinant of \mathbf is 0, which is a necessary and sufficient condition for a matrix to be non-invertible.


Methods of matrix inversion


Gaussian elimination

Gaussian elimination In mathematics, Gaussian elimination, also known as row reduction, is an algorithm for solving systems of linear equations. It consists of a sequence of row-wise operations performed on the corresponding matrix of coefficients. This method can a ...
is a useful and easy way to compute the inverse of a matrix. To compute a matrix inverse using this method, an
augmented matrix In linear algebra, an augmented matrix (A \vert B) is a k \times (n+1) matrix obtained by appending a k-dimensional column vector B, on the right, as a further column to a k \times n-dimensional matrix A. This is usually done for the purpose of p ...
is first created with the left side being the matrix to invert and the right side being the
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
. Then, Gaussian elimination is used to convert the left side into the identity matrix, which causes the right side to become the inverse of the input matrix. For example, take the following matrix: \mathbf = \begin-1 & \tfrac \\ 1 & -1\end. The first step to compute its inverse is to create the augmented matrix \left(\!\!\begin -1 & \tfrac & 1 & 0 \\ 1 & -1 & 0 & 1 \end\!\!\right) . Call the first row of this matrix R_1 and the second row R_2. Then, add row 1 to row 2 (R_1 + R_2 \to R_2). This yields \left(\!\!\begin -1 & \tfrac & 1 & 0 \\ 0 & \tfrac & 1 & 1 \end\!\!\right). Next, subtract row 2, multiplied by 3, from row 1 (R_1 - 3\, R_2 \to R_1), which yields \left(\!\!\begin -1 & 0 & -2 & -3 \\ 0 & \tfrac & 1 & 1 \end\!\!\right). Finally, multiply row 1 by −1 (-R_1 \to R_1) and row 2 by 2 (2\, R_2 \to R_2). This yields the identity matrix on the left side and the inverse matrix on the right:\left(\!\!\begin 1 & 0 & 2 & 3 \\ 0 & 1 & 2 & 2 \end\!\!\right). Thus, \mathbf^ = \begin 2 & 3 \\ 2 & 2 \end. It works because the process of Gaussian elimination can be viewed as a sequence of applying left matrix multiplication using elementary row operations using elementary matrices (\mathbf E_n), such as \mathbf E_n \mathbf E_ \cdots \mathbf E_2 \mathbf E_1 \mathbf A = \mathbf I. Applying right-multiplication using \mathbf A^, we get \mathbf E_n \mathbf E_ \cdots \mathbf E_2 \mathbf E_1 \mathbf I = \mathbf I \mathbf A^. And the right side \mathbf I \mathbf A^ = \mathbf A^, which is the inverse we want. To obtain \mathbf E_n \mathbf E_ \cdots \mathbf E_2 \mathbf E_1 \mathbf I, we create the augumented matrix by combining with and applying
Gaussian elimination In mathematics, Gaussian elimination, also known as row reduction, is an algorithm for solving systems of linear equations. It consists of a sequence of row-wise operations performed on the corresponding matrix of coefficients. This method can a ...
. The two portions will be transformed using the same sequence of elementary row operations. When the left portion becomes , the right portion applied the same elementary row operation sequence will become .


Newton's method

A generalization of
Newton's method In numerical analysis, the Newton–Raphson method, also known simply as Newton's method, named after Isaac Newton and Joseph Raphson, is a root-finding algorithm which produces successively better approximations to the roots (or zeroes) of a ...
as used for a multiplicative inverse algorithm may be convenient if it is convenient to find a suitable starting seed: : X_ = 2X_k - X_k A X_k. Victor Pan and John Reif have done work that includes ways of generating a starting seed. Newton's method is particularly useful when dealing with families of related matrices that behave enough like the sequence manufactured for the
homotopy In topology, two continuous functions from one topological space to another are called homotopic (from and ) if one can be "continuously deformed" into the other, such a deformation being called a homotopy ( ; ) between the two functions. ...
above: sometimes a good starting point for refining an approximation for the new inverse can be the already obtained inverse of a previous matrix that nearly matches the current matrix. For example, the pair of sequences of inverse matrices used in obtaining matrix square roots by Denman–Beavers iteration. That may need more than one pass of the iteration at each new matrix, if they are not close enough together for just one to be enough. Newton's method is also useful for "touch up" corrections to the Gauss–Jordan algorithm which has been contaminated by small errors from imperfect computer arithmetic.


Cayley–Hamilton method

The Cayley–Hamilton theorem allows the inverse of to be expressed in terms of , traces and powers of : : \mathbf^ = \frac \sum_^ \mathbf^s \sum_ \prod_^ \frac \operatorname\left(\mathbf^l\right)^, where is size of , and is the trace of matrix given by the sum of the main diagonal. The sum is taken over and the sets of all k_l \geq 0 satisfying the linear
Diophantine equation ''Diophantine'' means pertaining to the ancient Greek mathematician Diophantus. A number of concepts bear this name: *Diophantine approximation In number theory, the study of Diophantine approximation deals with the approximation of real n ...
: s + \sum_^ lk_l = n - 1. The formula can be rewritten in terms of complete Bell polynomials of arguments t_l = - (l - 1)! \operatorname\left(A^l\right) as : \mathbf^ = \frac \sum_^n \mathbf^ \frac B_(t_1, t_2, \ldots, t_). That is described in more detail under Cayley–Hamilton method.


Eigendecomposition

If matrix can be eigendecomposed, and if none of its eigenvalues are zero, then is invertible and its inverse is given by : \mathbf^ = \mathbf\mathbf^\mathbf^, where is the square matrix whose th column is the
eigenvector In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by ...
q_i of , and is the
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagon ...
whose diagonal entries are the corresponding eigenvalues, that is, \Lambda_ = \lambda_i. If is symmetric, is guaranteed to be an
orthogonal matrix In linear algebra, an orthogonal matrix, or orthonormal matrix, is a real square matrix whose columns and rows are orthonormal vectors. One way to express this is Q^\mathrm Q = Q Q^\mathrm = I, where is the transpose of and is the identi ...
, therefore \mathbf^ = \mathbf^\mathrm . Furthermore, because is a diagonal matrix, its inverse is easy to calculate: : \left Lambda^\right = \frac.


Cholesky decomposition

If matrix is positive definite, then its inverse can be obtained as : \mathbf^ = \left(\mathbf^*\right)^ \mathbf^ , where is the lower triangular Cholesky decomposition of , and denotes the conjugate transpose of .


Analytic solution

Writing the transpose of the matrix of cofactors, known as an adjugate matrix, may also be an efficient way to calculate the inverse of ''small'' matrices, but the recursive method is inefficient for large matrices. To determine the inverse, we calculate a matrix of cofactors: : \mathbf^ = \mathbf^\mathrm = \begin \mathbf_ & \mathbf_ & \cdots & \mathbf_ \\ \mathbf_ & \mathbf_ & \cdots & \mathbf_ \\ \vdots & \vdots & \ddots & \vdots \\ \mathbf_ & \mathbf_ & \cdots & \mathbf_ \\ \end so that : \left(\mathbf^\right)_ = \left(\mathbf^\right)_ = \left(\mathbf_\right) where is the
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of , is the matrix of cofactors, and represents the matrix
transpose In linear algebra, the transpose of a Matrix (mathematics), matrix is an operator which flips a matrix over its diagonal; that is, it switches the row and column indices of the matrix by producing another matrix, often denoted by (among other ...
.


Inversion of 2 × 2 matrices

The ''cofactor equation'' listed above yields the following result for matrices. Inversion of these matrices can be done as follows: : \mathbf^ = \begin a & b \\ c & d \\ \end^ = \frac \begin \,\,\,d & \!\!-b \\ -c & \,a \\ \end = \frac \begin \,\,\,d & \!\!-b \\ -c & \,a \\ \end. This is possible because is the reciprocal of the determinant of the matrix in question, and the same strategy could be used for other matrix sizes. The Cayley–Hamilton method gives : \mathbf^ = \frac \left \left( \operatorname\mathbf \right) \mathbf - \mathbf \right.


Inversion of 3 × 3 matrices

A computationally efficient matrix inversion is given by : \mathbf^ = \begin a & b & c\\ d & e & f \\ g & h & i\\ \end^ = \frac \begin \, A & \, B & \,C \\ \, D & \, E & \, F \\ \, G & \, H & \, I\\ \end^\mathrm = \frac \begin \, A & \, D & \,G \\ \, B & \, E & \,H \\ \, C & \,F & \, I\\ \end (where the scalar is not to be confused with the matrix ). If the determinant is non-zero, the matrix is invertible, with the entries of the intermediary matrix on the right side above given by : \begin A &=& (ei - fh), &\quad& D &=& -(bi - ch), &\quad& G &=& (bf - ce), \\ B &=& -(di - fg), &\quad& E &=& (ai - cg), &\quad& H &=& -(af - cd), \\ C &=& (dh - eg), &\quad& F &=& -(ah - bg), &\quad& I &=& (ae - bd). \\ \end The determinant of can be computed by applying the rule of Sarrus as follows: : \det(\mathbf) = aA + bB + cC. The Cayley–Hamilton decomposition gives : \mathbf^ = \frac\left( \tfrac\left (\operatorname\mathbf)^ - \operatorname(\mathbf^)\right\mathbf - \mathbf\operatorname\mathbf + \mathbf^\right). The general inverse can be expressed concisely in terms of the
cross product In mathematics, the cross product or vector product (occasionally directed area product, to emphasize its geometric significance) is a binary operation on two vectors in a three-dimensional oriented Euclidean vector space (named here E), and ...
and
triple product In geometry and algebra, the triple product is a product of three 3- dimensional vectors, usually Euclidean vectors. The name "triple product" is used for two different products, the scalar-valued scalar triple product and, less often, the ve ...
. If a matrix \mathbf = \begin \mathbf_0 & \mathbf_1 & \mathbf_2\end (consisting of three column vectors, \mathbf_0, \mathbf_1, and \mathbf_2) is invertible, its inverse is given by : \mathbf^ = \frac\begin ^\mathrm \\ ^\mathrm \\ ^\mathrm \end. The determinant of , , is equal to the triple product of , , and —the volume of the
parallelepiped In geometry, a parallelepiped is a three-dimensional figure formed by six parallelograms (the term ''rhomboid'' is also sometimes used with this meaning). By analogy, it relates to a parallelogram just as a cube relates to a square. Three equiva ...
formed by the rows or columns: : \det(\mathbf) = \mathbf_0\cdot(\mathbf_1\times\mathbf_2). The correctness of the formula can be checked by using cross- and triple-product properties and by noting that for groups, left and right inverses always coincide. Intuitively, because of the cross products, each row of is orthogonal to the non-corresponding two columns of (causing the off-diagonal terms of \mathbf = \mathbf^\mathbf be zero). Dividing by : \det(\mathbf) = \mathbf_0\cdot(\mathbf_1\times\mathbf_2) causes the diagonal entries of to be unity. For example, the first diagonal is: : 1 = \frac \mathbf\cdot(\mathbf_1\times\mathbf_2).


Inversion of 4 × 4 matrices

With increasing dimension, expressions for the inverse of get complicated. For , the Cayley–Hamilton method leads to an expression that is still tractable: :\begin \mathbf^ = \frac\Bigl( &\tfrac\bigl( (\operatorname\mathbf)^ - 3\operatorname\mathbf\operatorname(\mathbf^) + 2\operatorname(\mathbf^)\bigr) \mathbf \\ 3mu&\ \ \ - \tfrac\mathbf\bigl((\operatorname\mathbf)^ - \operatorname(\mathbf^)\bigr) + \mathbf^\operatorname\mathbf - \mathbf^ \Bigr). \end


Blockwise inversion

Let \mathbf M = \begin \mathbf & \mathbf \\ \mathbf & \mathbf \end where , , and are matrix sub-blocks of arbitrary size and \mathbf M / \mathbf A := \mathbf D - \mathbf C \mathbf A^ \mathbf B is the Schur complement of . ( must be square, so that it can be inverted. Furthermore, and must be nonsingular.) Matrices can also be ''inverted blockwise'' by using the analytic inversion formula: The strategy is particularly advantageous if is diagonal and is a small matrix, since they are the only matrices requiring inversion. The nullity theorem says that the nullity of equals the nullity of the sub-block in the lower right of the inverse matrix, and that the nullity of equals the nullity of the sub-block in the upper right of the inverse matrix. The inversion procedure that led to Equation () performed matrix block operations that operated on and first. Instead, if and are operated on first, and provided and are nonsingular, the result is Equating the upper-left sub-matrices of Equations () and () leads to where Equation () is the Woodbury matrix identity, which is equivalent to the binomial inverse theorem. If and are both invertible, then the above two block matrix inverses can be combined to provide the simple factorization By the Weinstein–Aronszajn identity, one of the two matrices in the block-diagonal matrix is invertible exactly when the other is. This formula simplifies significantly when the upper right block matrix is the
zero matrix In mathematics, particularly linear algebra, a zero matrix or null matrix is a matrix all of whose entries are zero. It also serves as the additive identity of the additive group of m \times n matrices, and is denoted by the symbol O or 0 followe ...
. This formulation is useful when the matrices and have relatively simple inverse formulas (or pseudo inverses in the case where the blocks are not all square. In this special case, the block matrix inversion formula stated in full generality above becomes :\begin \mathbf & \mathbf \\ \mathbf & \mathbf \end^ = \begin \mathbf^ & \mathbf \\ -\mathbf^\mathbf^ & \mathbf^ \end. If the given invertible matrix is a symmetric matrix with invertible block the following block inverse formula holds where \mathbf = \mathbf - \mathbf\mathbf^\mathbf^T. This requires 2 inversions of the half-sized matrices and and only 4 multiplications of half-sized matrices, if organized properly \begin \mathbf_1 &= \mathbf\mathbf^, \\ mu\mathbf_2 &= \mathbf_1\mathbf^=\mathbf\mathbf^\mathbf^T, \\ mu\mathbf_3 &= \mathbf^\mathbf_1=\mathbf^\mathbf\mathbf^, \\ mu\mathbf_4 &= \mathbf_1^T\mathbf_3=\mathbf^\mathbf^T \mathbf^\mathbf\mathbf^, \end together with some additions, subtractions, negations and transpositions of negligible complexity. Any matrix \mathbf has an associated positive semidefinite, symmetric matrix \mathbf^T\mathbf, which is exactly invertible (and positive definite), if and only if \mathbf is invertible. By writing \mathbf^=\left(\mathbf^T\mathbf\right)^\mathbf^T matrix inversion can be reduced to inverting symmetric matrices and 2 additional matrix multiplications, because the positive definite matrix \mathbf^T\mathbf satisfies the invertibility condition for its left upper block . Those formulas together allow to construct a divide and conquer algorithm that uses blockwise inversion of associated symmetric matrices to invert a matrix with the same time complexity as the
matrix multiplication algorithm Because matrix multiplication is such a central operation in many numerical algorithms, much work has been invested in making matrix multiplication algorithms efficient. Applications of matrix multiplication in computational problems are found in m ...
that is used internally.T. H. Cormen, C. E. Leiserson, R. L. Rivest, C. Stein, ''Introduction to Algorithms'', 3rd ed., MIT Press, Cambridge, MA, 2009, §28.2. Research into matrix multiplication complexity shows that there exist matrix multiplication algorithms with a complexity of operations, while the best proven lower bound is .


By Neumann series

If a matrix has the property that : \lim_ (\mathbf I - \mathbf A)^n = 0 then is nonsingular and its inverse may be expressed by a Neumann series: : \mathbf A^ = \sum_^\infty (\mathbf I - \mathbf A)^n. Truncating the sum results in an "approximate" inverse which may be useful as a preconditioner. Note that a truncated series can be accelerated exponentially by noting that the Neumann series is a geometric sum. As such, it satisfies : \sum_^ (\mathbf I - \mathbf A)^n = \prod_^\left(\mathbf I + (\mathbf I - \mathbf A)^\right). Therefore, only matrix multiplications are needed to compute terms of the sum. More generally, if is "near" the invertible matrix in the sense that : \lim_ \left(\mathbf I - \mathbf X^ \mathbf A\right)^n = 0 \mathrm \lim_ \left(\mathbf I - \mathbf A \mathbf X^\right)^n = 0 then is nonsingular and its inverse is : \mathbf A^ = \sum_^\infty \left(\mathbf X^ (\mathbf X - \mathbf A)\right)^n \mathbf X^~. If it is also the case that has rank 1 then this simplifies to : \mathbf A^ = \mathbf X^ - \frac~.


''p''-adic approximation

If is a matrix with
integer An integer is the number zero (0), a positive natural number (1, 2, 3, ...), or the negation of a positive natural number (−1, −2, −3, ...). The negations or additive inverses of the positive natural numbers are referred to as negative in ...
or
rational Rationality is the quality of being guided by or based on reason. In this regard, a person acts rationally if they have a good reason for what they do, or a belief is rational if it is based on strong evidence. This quality can apply to an ...
entries, and we seek a solution in arbitrary-precision rationals, a -adic approximation method converges to an exact solution in , assuming standard matrix multiplication is used. The method relies on solving linear systems via Dixon's method of -adic approximation (each in ) and is available as such in software specialized in arbitrary-precision matrix operations, for example, in IML.


Reciprocal basis vectors method

Given an square matrix \mathbf = \left x^ \right, 1 \leq i,j \leq n , with rows interpreted as vectors \mathbf_ = x^ \mathbf_ ( Einstein summation assumed) where the \mathbf_ are a standard
orthonormal basis In mathematics, particularly linear algebra, an orthonormal basis for an inner product space V with finite Dimension (linear algebra), dimension is a Basis (linear algebra), basis for V whose vectors are orthonormal, that is, they are all unit vec ...
of
Euclidean space Euclidean space is the fundamental space of geometry, intended to represent physical space. Originally, in Euclid's ''Elements'', it was the three-dimensional space of Euclidean geometry, but in modern mathematics there are ''Euclidean spaces ...
\mathbb^ (\mathbf_ = \mathbf^, \mathbf_ \cdot \mathbf^ = \delta_i^j), then using
Clifford algebra In mathematics, a Clifford algebra is an algebra generated by a vector space with a quadratic form, and is a unital associative algebra with the additional structure of a distinguished subspace. As -algebras, they generalize the real number ...
(or geometric algebra) we compute the reciprocal (sometimes called dual) column vectors: :\mathbf^ = x_ \mathbf^ = (-1)^ (\mathbf_ \wedge\cdots\wedge ()_ \wedge\cdots\wedge\mathbf_) \cdot (\mathbf_ \wedge\ \mathbf_ \wedge\cdots\wedge\mathbf_)^ as the columns of the inverse matrix \mathbf^ = _ Note that, the place "()_" indicates that "\mathbf_" is removed from that place in the above expression for \mathbf^. We then have \mathbf\mathbf^ = \left \mathbf_ \cdot \mathbf^ \right= \left \delta_^ \right= \mathbf_ , where \delta_^ is the
Kronecker delta In mathematics, the Kronecker delta (named after Leopold Kronecker) is a function of two variables, usually just non-negative integers. The function is 1 if the variables are equal, and 0 otherwise: \delta_ = \begin 0 &\text i \neq j, \\ 1 &\ ...
. We also have \mathbf^\mathbf = \left left(\mathbf_\cdot\mathbf^\right)\left(\mathbf^\cdot\mathbf_\right)\right= \left mathbf_\cdot\mathbf^\right= \left delta_^\right= \mathbf_, as required. If the vectors \mathbf_ are not linearly independent, then (\mathbf_ \wedge \mathbf_ \wedge\cdots\wedge\mathbf_) = 0 and the matrix \mathbf is not invertible (has no inverse).


Derivative of the matrix inverse

Suppose that the invertible matrix A depends on a parameter ''t''. Then the derivative of the inverse of A with respect to ''t'' is given by : \frac = - \mathbf^ \frac \mathbf^. To derive the above expression for the derivative of the inverse of A, one can differentiate the definition of the matrix inverse \mathbf^\mathbf=\mathbf and then solve for the inverse of A: : \frac = \frac\mathbf + \mathbf^\frac = \frac = \mathbf. Subtracting \mathbf^\frac from both sides of the above and multiplying on the right by \mathbf^ gives the correct expression for the derivative of the inverse: : \frac = - \mathbf^ \frac \mathbf^. Similarly, if \varepsilon is a small number then : \left(\mathbf + \varepsilon\mathbf\right)^ = \mathbf^ - \varepsilon \mathbf^ \mathbf \mathbf^ + \mathcal(\varepsilon^2)\,. More generally, if : \frac = \sum_i g_i (\mathbf) \frach_i (\mathbf), then, : f (\mathbf + \varepsilon\mathbf) = f (\mathbf) + \varepsilon\sum_i g_i (\mathbf) \mathbf h_i (\mathbf) + \mathcal\left(\varepsilon^2\right). Given a positive integer n, : \begin \frac &= \sum_^n \mathbf^\frac\mathbf^,\\ \frac &= -\sum_^n \mathbf^\frac\mathbf^. \end Therefore, : \begin (\mathbf + \varepsilon \mathbf)^ &= \mathbf^ + \varepsilon \sum_^n \mathbf^\mathbf\mathbf^ + \mathcal\left(\varepsilon^2\right),\\ (\mathbf + \varepsilon \mathbf)^ &= \mathbf^ - \varepsilon \sum_^n \mathbf^\mathbf\mathbf^ + \mathcal\left(\varepsilon^2\right). \end


Generalized inverses

Some of the properties of inverse matrices are shared by generalized inverses (such as the Moore–Penrose inverse), which can be defined for any ''m''-by-''n'' matrix.


Applications

For most practical applications, it is not necessary to invert a matrix to solve a
system of linear equations In mathematics, a system of linear equations (or linear system) is a collection of two or more linear equations involving the same variable (math), variables. For example, : \begin 3x+2y-z=1\\ 2x-2y+4z=-2\\ -x+\fracy-z=0 \end is a system of th ...
; however, for a unique solution, it is necessary for the matrix involved to be invertible. Decomposition techniques like LU decomposition are much faster than inversion, and various fast algorithms for special classes of linear systems have also been developed.


Regression/least squares

Although an explicit inverse is not necessary to estimate the vector of unknowns, it is the easiest way to estimate their accuracy and is found in the diagonal of a matrix inverse (the posterior covariance matrix of the vector of unknowns). However, faster algorithms to compute only the diagonal entries of a matrix inverse are known in many cases.


Matrix inverses in real-time simulations

Matrix inversion plays a significant role in
computer graphics Computer graphics deals with generating images and art with the aid of computers. Computer graphics is a core technology in digital photography, film, video games, digital art, cell phone and computer displays, and many specialized applications. ...
, particularly in
3D graphics 3D computer graphics, sometimes called CGI, 3D-CGI or three-dimensional computer graphics, are graphics that use a three-dimensional representation of geometric data (often Cartesian) that is stored in the computer for the purposes of perfor ...
rendering and 3D simulations. Examples include screen-to-world ray casting, world-to-subspace-to-world object transformations, and physical simulations.


Matrix inverses in MIMO wireless communication

Matrix inversion also plays a significant role in the MIMO (Multiple-Input, Multiple-Output) technology in wireless communications. The MIMO system consists of ''N'' transmit and ''M'' receive antennas. Unique signals, occupying the same
frequency band Spectral bands are regions of a given spectrum, having a specific range of wavelengths or frequencies. Most often, it refers to electromagnetic bands, regions of the electromagnetic spectrum. More generally, spectral bands may also be means in ...
, are sent via ''N'' transmit antennas and are received via ''M'' receive antennas. The signal arriving at each receive antenna will be a
linear combination In mathematics, a linear combination or superposition is an Expression (mathematics), expression constructed from a Set (mathematics), set of terms by multiplying each term by a constant and adding the results (e.g. a linear combination of ''x'' a ...
of the ''N'' transmitted signals forming an ''N'' × ''M'' transmission matrix H. It is crucial for the matrix H to be invertible so that the receiver can figure out the transmitted information.


See also


References


Further reading

* * * *


External links

* * {{DEFAULTSORT:Invertible Matrix Linear algebra Matrices (mathematics) Determinants Matrix theory