HOME

TheInfoList



OR:

In
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
, a minor of a
matrix Matrix (: matrices or matrixes) or MATRIX may refer to: Science and mathematics * Matrix (mathematics), a rectangular array of numbers, symbols or expressions * Matrix (logic), part of a formula in prenex normal form * Matrix (biology), the m ...
is the
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of some smaller square matrix generated from by removing one or more of its rows and columns. Minors obtained by removing just one row and one column from square matrices (first minors) are required for calculating matrix cofactors, which are useful for computing both the determinant and inverse of square matrices. The requirement that the square matrix be smaller than the original matrix is often omitted in the definition.


Definition and illustration


First minors

If is a square matrix, then the ''minor'' of the entry in the -th row and -th column (also called the ''minor'', or a ''first minor'') is the
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of the submatrix formed by deleting the -th row and -th column. This number is often denoted . The ''cofactor'' is obtained by multiplying the minor by . To illustrate these definitions, consider the following matrix, \begin 1 & 4 & 7 \\ 3 & 0 & 5 \\ -1 & 9 & 11 \\ \end To compute the minor and the cofactor , we find the determinant of the above matrix with row 2 and column 3 removed. M_ = \det \begin 1 & 4 & \Box \\ \Box & \Box & \Box \\ -1 & 9 & \Box \\ \end= \det \begin 1 & 4 \\ -1 & 9 \\ \end = 9-(-4) = 13 So the cofactor of the entry is C_ = (-1)^(M_) = -13.


General definition

Let be an matrix and an
integer An integer is the number zero (0), a positive natural number (1, 2, 3, ...), or the negation of a positive natural number (−1, −2, −3, ...). The negations or additive inverses of the positive natural numbers are referred to as negative in ...
with , and . A ''minor'' of , also called ''minor determinant of order '' of or, if , the ''th'' ''minor determinant'' of (the word "determinant" is often omitted, and the word "degree" is sometimes used instead of "order") is the determinant of a matrix obtained from by deleting rows and columns. Sometimes the term is used to refer to the matrix obtained from as above (by deleting rows and columns), but this matrix should be referred to as a ''(square) submatrix'' of , leaving the term "minor" to refer to the determinant of this matrix. For a matrix as above, there are a total of \cdot minors of size . The ''minor of order zero'' is often defined to be 1. For a square matrix, the ''zeroth minor'' is just the determinant of the matrix.Elementary Matrix Algebra (Third edition), Franz E. Hohn, The Macmillan Company, 1973, Let \begin I &= 1 \le i_1 < i_2 < \cdots < i_k \le m, \\ pt J &= 1 \le j_1 < j_2 < \cdots < j_k \le n, \end be ordered sequences (in natural order, as it is always assumed when talking about minors unless otherwise stated) of indexes. The minor \det \bigl( (\mathbf A_)_ \bigr) corresponding to these choices of indexes is denoted \det_ A or \det \mathbf A_ or mathbf A or M_ or M_ or M_ (where the denotes the sequence of indexes , etc.), depending on the source. Also, there are two types of denotations in use in literature: by the minor associated to ordered sequences of indexes and , some authors mean the determinant of the matrix that is formed as above, by taking the elements of the original matrix from the rows whose indexes are in and columns whose indexes are in , whereas some other authors mean by a minor associated to and the determinant of the matrix formed from the original matrix by deleting the rows in and columns in ; which notation is used should always be checked. In this article, we use the inclusive definition of choosing the elements from rows of and columns of . The exceptional case is the case of the first minor or the -minor described above; in that case, the exclusive meaning M_ = \det \bigl( \left( \mathbf A_ \right)_ \bigr) is standard everywhere in the literature and is used in this article also.


Complement

The complement of a minor of a square matrix, , is formed by the determinant of the matrix from which all the rows () and columns () associated with have been removed. The complement of the first minor of an element is merely that element.


Applications of minors and cofactors


Cofactor expansion of the determinant

The cofactors feature prominently in Laplace's formula for the expansion of determinants, which is a method of computing larger determinants in terms of smaller ones. Given an matrix , the determinant of , denoted , can be written as the sum of the cofactors of any row or column of the matrix multiplied by the entries that generated them. In other words, defining C_ = (-1)^ M_ then the cofactor expansion along the -th column gives: \begin \det(\mathbf A) &= a_C_ + a_C_ + a_C_ + \cdots + a_C_ \\ pt &= \sum_^ a_ C_ \\ pt &= \sum_^ a_(-1)^ M_ \end The cofactor expansion along the -th row gives: \begin \det(\mathbf A) &= a_C_ + a_C_ + a_C_ + \cdots + a_C_ \\ pt &= \sum_^ a_ C_ \\ pt &= \sum_^ a_ (-1)^ M_ \end


Inverse of a matrix

One can write down the inverse of an
invertible matrix In linear algebra, an invertible matrix (''non-singular'', ''non-degenarate'' or ''regular'') is a square matrix that has an inverse. In other words, if some other matrix is multiplied by the invertible matrix, the result can be multiplied by a ...
by computing its cofactors by using
Cramer's rule In linear algebra, Cramer's rule is an explicit formula for the solution of a system of linear equations with as many equations as unknowns, valid whenever the system has a unique solution. It expresses the solution in terms of the determinants of ...
, as follows. The matrix formed by all of the cofactors of a square matrix is called the cofactor matrix (also called the matrix of cofactors or, sometimes, ''comatrix''): \mathbf C = \begin C_ & C_ & \cdots & C_ \\ C_ & C_ & \cdots & C_ \\ \vdots & \vdots & \ddots & \vdots \\ C_ & C_ & \cdots & C_ \end Then the inverse of is the transpose of the cofactor matrix times the reciprocal of the determinant of : \mathbf A^ = \frac \mathbf C^\mathsf. The transpose of the cofactor matrix is called the adjugate matrix (also called the ''classical adjoint'') of . The above formula can be generalized as follows: Let \begin I &= 1 \le i_1 < i_2 < \ldots < i_k \le n, \\ pt J &= 1 \le j_1 < j_2 < \ldots < j_k \le n, \end be ordered sequences (in natural order) of indexes (here is an matrix). Then mathbf A^ = \pm\frac, where denote the ordered sequences of indices (the indices are in natural order of magnitude, as above) complementary to , so that every index appears exactly once in either or , but not in both (similarly for the and ) and denotes the determinant of the submatrix of formed by choosing the rows of the index set and columns of index set . Also, mathbf A = \det \bigl( (A_)_ \bigr). A simple proof can be given using wedge product. Indeed, \bigl \mathbf A^ \bigr (e_1\wedge\ldots \wedge e_n) = \pm(\mathbf A^e_)\wedge \ldots \wedge(\mathbf A^e_)\wedge e_\wedge\ldots \wedge e_, where e_1, \ldots, e_n are the basis vectors. Acting by on both sides, one gets \begin &\ \bigl mathbf A^ \bigr \det \mathbf A (e_1\wedge\ldots \wedge e_n) \\ pt =&\ \pm (e_)\wedge \ldots \wedge(e_)\wedge (\mathbf A e_)\wedge\ldots \wedge (\mathbf A e_) \\ pt =&\ \pm mathbf A(e_1\wedge\ldots \wedge e_n). \end The sign can be worked out to be (-1)^\wedge \!\!\left( \sum_^ i_s - \sum_^ j_s \right), so the sign is determined by the sums of elements in and .


Other applications

Given an matrix with real entries (or entries from any other field) and rank , then there exists at least one non-zero minor, while all larger minors are zero. We will use the following notation for minors: if is an matrix, is a
subset In mathematics, a Set (mathematics), set ''A'' is a subset of a set ''B'' if all Element (mathematics), elements of ''A'' are also elements of ''B''; ''B'' is then a superset of ''A''. It is possible for ''A'' and ''B'' to be equal; if they a ...
of with elements, and is a subset of with elements, then we write for the minor of that corresponds to the rows with index in and the columns with index in . * If , then is called a ''principal minor''. * If the matrix that corresponds to a principal minor is a square upper-left submatrix of the larger matrix (i.e., it consists of matrix elements in rows and columns from 1 to , also known as a leading principal submatrix), then the principal minor is called a ''leading principal minor (of order )'' or ''corner (principal) minor (of order )''. For an square matrix, there are leading principal minors. * A ''basic minor'' of a matrix is the determinant of a square submatrix that is of maximal size with nonzero determinant. * For Hermitian matrices, the leading principal minors can be used to test for positive definiteness and the principal minors can be used to test for positive semidefiniteness. See Sylvester's criterion for more details. Both the formula for ordinary
matrix multiplication In mathematics, specifically in linear algebra, matrix multiplication is a binary operation that produces a matrix (mathematics), matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the n ...
and the Cauchy–Binet formula for the determinant of the product of two matrices are special cases of the following general statement about the minors of a product of two matrices. Suppose that is an matrix, is an matrix, is a
subset In mathematics, a Set (mathematics), set ''A'' is a subset of a set ''B'' if all Element (mathematics), elements of ''A'' are also elements of ''B''; ''B'' is then a superset of ''A''. It is possible for ''A'' and ''B'' to be equal; if they a ...
of with elements and is a subset of with elements. Then mathbf = \sum_ mathbf mathbf\, where the sum extends over all subsets of with elements. This formula is a straightforward extension of the Cauchy–Binet formula.


Multilinear algebra approach

A more systematic, algebraic treatment of minors is given in
multilinear algebra Multilinear algebra is the study of Function (mathematics), functions with multiple vector space, vector-valued Argument of a function, arguments, with the functions being Linear map, linear maps with respect to each argument. It involves concept ...
, using the wedge product: the -minors of a matrix are the entries in the -th exterior power map. If the columns of a matrix are wedged together at a time, the minors appear as the components of the resulting -vectors. For example, the 2 × 2 minors of the matrix \begin 1 & 4 \\ 3 & \!\!-1 \\ 2 & 1 \\ \end are −13 (from the first two rows), −7 (from the first and last row), and 5 (from the last two rows). Now consider the wedge product (\mathbf_1 + 3\mathbf_2 + 2\mathbf_3) \wedge (4\mathbf_1 - \mathbf_2 + \mathbf_3) where the two expressions correspond to the two columns of our matrix. Using the properties of the wedge product, namely that it is bilinear and alternating, \mathbf_i \wedge \mathbf_i = 0, and antisymmetric, \mathbf_i\wedge \mathbf_j = - \mathbf_j\wedge \mathbf_i, we can simplify this expression to -13 \mathbf_1\wedge \mathbf_2 -7 \mathbf_1\wedge \mathbf_3 +5 \mathbf_2\wedge \mathbf_3 where the coefficients agree with the minors computed earlier.


A remark about different notation

In some books, instead of ''cofactor'' the term ''adjunct'' is used. Felix Gantmacher, ''Theory of matrices'' (1st ed., original language is Russian), Moscow: State Publishing House of technical and theoretical literature, 1953, p.491, Moreover, it is denoted as and defined in the same way as cofactor: \mathbf_ = (-1)^ \mathbf_ Using this notation the inverse matrix is written this way: \mathbf^ = \frac\begin A_ & A_ & \cdots & A_ \\ A_ & A_ & \cdots & A_ \\ \vdots & \vdots & \ddots & \vdots \\ A_ & A_ & \cdots & A_ \end Keep in mind that ''adjunct'' is not adjugate or adjoint. In modern terminology, the "adjoint" of a matrix most often refers to the corresponding adjoint operator.


See also

* Submatrix * Compound matrix


References


External links


MIT Linear Algebra Lecture on Cofactors
at Google Video, from MIT OpenCourseWare


Springer Encyclopedia of Mathematics entry for ''Minor''
{{linear algebra Matrix theory Determinants