HOME

TheInfoList



OR:

In
matrix calculus In mathematics, matrix calculus is a specialized notation for doing multivariable calculus, especially over spaces of matrix (mathematics), matrices. It collects the various partial derivatives of a single Function (mathematics), function with ...
, Jacobi's formula expresses the
derivative In mathematics, the derivative is a fundamental tool that quantifies the sensitivity to change of a function's output with respect to its input. The derivative of a function of a single variable at a chosen input value, when it exists, is t ...
of the
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of a matrix ''A'' in terms of the adjugate of ''A'' and the derivative of ''A''., Part Three, Section 8.3 If is a differentiable map from the real numbers to matrices, then : \frac \det A(t) = \operatorname \left (\operatorname(A(t)) \, \frac\right ) = \left(\det A(t) \right) \cdot \operatorname \left (A(t)^ \cdot \, \frac\right ) where is the trace of the matrix and \operatorname(X) is its adjugate matrix. (The latter equality only holds if ''A''(''t'') is
invertible In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers. Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that ...
.) As a special case, : = \operatorname(A)_. Equivalently, if stands for the differential of , the general formula is : d \det (A) = \operatorname (\operatorname(A) \, dA) = \det (A) \operatorname \left (A^ d A\right ) The formula is named after the mathematician Carl Jacobi.


Derivation


Via matrix computation

Theorem. (Jacobi's formula) For any differentiable map ''A'' from the real numbers to ''n'' × ''n'' matrices, : d \det (A) = \operatorname (\operatorname(A) \, dA). ''Proof.'' Laplace's formula for the determinant of a matrix ''A'' can be stated as :\det(A) = \sum_j A_ \operatorname^ (A)_. Notice that the summation is performed over some arbitrary row ''i'' of the matrix. The determinant of ''A'' can be considered to be a function of the elements of ''A'': :\det(A) = F\,(A_, A_, \ldots , A_, A_, \ldots , A_) so that, by the
chain rule In calculus, the chain rule is a formula that expresses the derivative of the Function composition, composition of two differentiable functions and in terms of the derivatives of and . More precisely, if h=f\circ g is the function such that h ...
, its differential is :d \det(A) = \sum_i \sum_j \,dA_. This summation is performed over all ''n''×''n'' elements of the matrix. To find ∂''F''/∂''A''''ij'' consider that on the right hand side of Laplace's formula, the index ''i'' can be chosen at will. (In order to optimize calculations: Any other choice would eventually yield the same result, but it could be much harder). In particular, it can be chosen to match the first index of ∂ / ∂''A''''ij'': : = = \sum_k Thus, by the product rule, : = \sum_k \operatorname^(A)_ + \sum_k A_ . Now, if an element of a matrix ''A''''ij'' and a cofactor adjT(''A'')''ik'' of element ''A''''ik'' lie on the same row (or column), then the cofactor will not be a function of ''Aij'', because the cofactor of ''A''''ik'' is expressed in terms of elements not in its own row (nor column). Thus, : = 0, so : = \sum_k \operatorname^(A)_ . All the elements of ''A'' are independent of each other, i.e. : = \delta_, where ''δ'' is the
Kronecker delta In mathematics, the Kronecker delta (named after Leopold Kronecker) is a function of two variables, usually just non-negative integers. The function is 1 if the variables are equal, and 0 otherwise: \delta_ = \begin 0 &\text i \neq j, \\ 1 &\ ...
, so : = \sum_k \operatorname^(A)_ \delta_ = \operatorname^(A)_. Therefore, :d(\det(A)) = \sum_i \sum_j \operatorname^(A)_ \,d A_ = \sum_j \sum_i \operatorname(A)_ \,d A_ = \sum_j (\operatorname(A) \,d A)_ = \operatorname(\operatorname(A) \,dA).\ \square


Via chain rule

Lemma 1. \det'(I)=\mathrm, where \det' is the differential of \det. This equation means that the differential of \det, evaluated at the identity matrix, is equal to the trace. The differential \det'(I) is a linear operator that maps an ''n'' × ''n'' matrix to a real number. ''Proof.'' Using the definition of a
directional derivative In multivariable calculus, the directional derivative measures the rate at which a function changes in a particular direction at a given point. The directional derivative of a multivariable differentiable (scalar) function along a given vect ...
together with one of its basic properties for differentiable functions, we have :\det'(I)(T)=\nabla_T \det(I)=\lim_\frac \det(I+\varepsilon T) is a polynomial in \varepsilon of order ''n''. It is closely related to the
characteristic polynomial In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The ...
of T. The constant term in that polynomial (the term with \varepsilon = 0) is 1, while the linear term in \varepsilon is \mathrm\ T. Lemma 2. For an invertible matrix ''A'', we have: \det'(A)(T)=\det A \; \mathrm(A^T). ''Proof.'' Consider the following function of ''X'': :\det X = \det (A A^ X) = \det (A) \ \det(A^ X) We calculate the differential of \det X and evaluate it at X = A using Lemma 1, the equation above, and the chain rule: :\det'(A)(T) = \det A \ \det'(I) (A^ T) = \det A \ \mathrm(A^ T) Theorem. (Jacobi's formula) \frac \det A = \mathrm\left(\mathrm\ A\frac\right) ''Proof.'' If A is invertible, by Lemma 2, with T = dA/dt :\frac \det A = \det A \; \mathrm \left(A^ \frac\right) = \mathrm \left( \mathrm\ A \; \frac \right) using the equation relating the adjugate of A to A^. Now, the formula holds for all matrices, since the set of invertible linear matrices is dense in the space of matrices.


Via diagonalization

Both sides of the Jacobi formula are polynomials in the matrix coefficients of and . It is therefore sufficient to verify the polynomial identity on the dense subset where the eigenvalues of are distinct and nonzero. If factors differentiably as A=BC, then : \mathrm(A^A')= \mathrm((BC)^(BC)')= \mathrm(B^B')+ \mathrm(C^C'). In particular, if is invertible, then I=L^L and : 0=\mathrm(I^I')= \mathrm(L(L^)')+ \mathrm(L^L'). Since has distinct eigenvalues, there exists a differentiable complex invertible matrix such that A = L^DL and is diagonal. Then : \mathrm(A^A')= \mathrm(L(L^)')+ \mathrm(D^D')+ \mathrm(L^L')= \mathrm(D^D'). Let \lambda_i, i=1,\ldots,n be the eigenvalues of . Then : \left(\ln\det A\right)' = \left(\sum_^\ln \lambda_i \right)' = \sum_^n \lambda_i'/\lambda_i = \mathrm(D^D')= \mathrm(A^A'), which is the Jacobi formula for matrices with distinct nonzero eigenvalues.


Corollary

The following is a useful relation connecting the trace to the determinant of the associated
matrix exponential In mathematics, the matrix exponential is a matrix function on square matrix, square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exp ...
: This statement is clear for diagonal matrices, and a proof of the general claim follows. For any
invertible matrix In linear algebra, an invertible matrix (''non-singular'', ''non-degenarate'' or ''regular'') is a square matrix that has an inverse. In other words, if some other matrix is multiplied by the invertible matrix, the result can be multiplied by a ...
A(t), in the previous section "Via Chain Rule", we showed that :\frac \det A(t) = \det A(t) \; \operatorname \left(A(t)^ \, \frac A(t)\right) Considering A(t) = \exp(tB) in this equation yields: : \frac \det e^ =\operatorname(B) \det e^ The desired result follows as the solution to this ordinary differential equation.


Applications

Several forms of the formula underlie the Faddeev–LeVerrier algorithm for computing the
characteristic polynomial In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The ...
, and explicit applications of the Cayley–Hamilton theorem. For example, starting from the following equation, which was proved above: :\frac \det A(t) = \det A(t) \ \operatorname \left(A(t)^ \, \frac A(t)\right) and using A(t) = t I - B, we get: :\frac \det (tI-B) = \det (tI-B) \operatorname tI-B)^= \operatorname operatorname (tI-B)/math> where adj denotes the adjugate matrix.


Remarks


References

* * {{DEFAULTSORT:Jacobi's Formula Determinants Matrix theory Articles containing proofs