In
matrix calculus
In mathematics, matrix calculus is a specialized notation for doing multivariable calculus, especially over spaces of matrices. It collects the various partial derivatives of a single function with respect to many variables, and/or of a mu ...
, Jacobi's formula expresses the
derivative
In mathematics, the derivative of a function of a real variable measures the sensitivity to change of the function value (output value) with respect to a change in its argument (input value). Derivatives are a fundamental tool of calculus. ...
of the
determinant
In mathematics, the determinant is a scalar value that is a function of the entries of a square matrix. It characterizes some properties of the matrix and the linear map represented by the matrix. In particular, the determinant is nonzero if ...
of a matrix ''A'' in terms of the
adjugate
In linear algebra, the adjugate or classical adjoint of a square matrix is the transpose of its cofactor matrix and is denoted by . It is also occasionally known as adjunct matrix, or "adjoint", though the latter today normally refers to a differe ...
of ''A'' and the derivative of ''A''.
[, Part Three, Section 8.3]
If is a differentiable map from the real numbers to matrices, then
:
where is the
trace of the matrix . (The latter equality only holds if ''A''(''t'') is
invertible
In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers.
Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that ...
.)
As a special case,
:
Equivalently, if stands for the
differential of , the general formula is
:
The formula is named after the mathematician
Carl Gustav Jacob Jacobi
Carl Gustav Jacob Jacobi (; ; 10 December 1804 – 18 February 1851) was a German mathematician who made fundamental contributions to elliptic functions, dynamics, differential equations, determinants, and number theory. His name is occasio ...
.
Derivation
Via Matrix Computation
We first prove a preliminary lemma:
Lemma. Let ''A'' and ''B'' be a pair of square matrices of the same dimension ''n''. Then
:
''Proof.'' The product ''AB'' of the pair of matrices has components
:
Replacing the matrix ''A'' by its
transpose
In linear algebra, the transpose of a matrix is an operator which flips a matrix over its diagonal;
that is, it switches the row and column indices of the matrix by producing another matrix, often denoted by (among other notations).
The tr ...
''A''
T is equivalent to permuting the indices of its components:
:
The result follows by taking the trace of both sides:
:
Theorem. (Jacobi's formula) For any differentiable map ''A'' from the real numbers to ''n'' × ''n'' matrices,
:
''Proof.''
Laplace's formula for the determinant of a matrix ''A'' can be stated as
:
Notice that the summation is performed over some arbitrary row ''i'' of the matrix.
The determinant of ''A'' can be considered to be a function of the elements of ''A'':
:
so that, by the
chain rule
In calculus, the chain rule is a formula that expresses the derivative of the Function composition, composition of two differentiable functions and in terms of the derivatives of and . More precisely, if h=f\circ g is the function such that h(x) ...
, its differential is
:
This summation is performed over all ''n''×''n'' elements of the matrix.
To find ∂''F''/∂''A''
''ij'' consider that on the right hand side of Laplace's formula, the index ''i'' can be chosen at will. (In order to optimize calculations: Any other choice would eventually yield the same result, but it could be much harder). In particular, it can be chosen to match the first index of ∂ / ∂''A''
''ij'':
:
Thus, by the product rule,
:
Now, if an element of a matrix ''A''
''ij'' and a
cofactor adj
T(''A'')
''ik'' of element ''A''
''ik'' lie on the same row (or column), then the cofactor will not be a function of ''A
ij'', because the cofactor of ''A''
''ik'' is expressed in terms of elements not in its own row (nor column). Thus,
:
so
:
All the elements of ''A'' are independent of each other, i.e.
:
where ''δ'' is the
Kronecker delta
In mathematics, the Kronecker delta (named after Leopold Kronecker) is a function of two variables, usually just non-negative integers. The function is 1 if the variables are equal, and 0 otherwise:
\delta_ = \begin
0 &\text i \neq j, \\
1 ...
, so
:
Therefore,
:
and applying the Lemma yields
:
Via Chain Rule
Lemma 1.
, where
is the differential of
.
This equation means that the differential of
, evaluated at the identity matrix, is equal to the trace. The differential
is a linear operator that maps an ''n'' × ''n'' matrix to a real number.
''Proof.'' Using the definition of a
directional derivative
In mathematics, the directional derivative of a multivariable differentiable (scalar) function along a given vector v at a given point x intuitively represents the instantaneous rate of change of the function, moving through x with a velocity ...
together with one of its basic properties for differentiable functions, we have
:
is a polynomial in
of order ''n''. It is closely related to the
characteristic polynomial
In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The ...
of
. The constant term (
) is 1, while the linear term in
is
.
Lemma 2. For an invertible matrix ''A'', we have:
.
''Proof.'' Consider the following function of ''X'':
:
We calculate the differential of
and evaluate it at
using Lemma 1, the equation above, and the chain rule:
:
Theorem. (Jacobi's formula)
''Proof.'' If
is invertible, by Lemma 2, with
:
using the equation relating the
adjugate
In linear algebra, the adjugate or classical adjoint of a square matrix is the transpose of its cofactor matrix and is denoted by . It is also occasionally known as adjunct matrix, or "adjoint", though the latter today normally refers to a differe ...
of
to
. Now, the formula holds for all matrices, since the set of invertible linear matrices is dense in the space of matrices.
Corollary
The following is a useful relation connecting the
trace to the determinant of the associated
matrix exponential
In mathematics, the matrix exponential is a matrix function on square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exponential giv ...
:
This statement is clear for diagonal matrices, and a proof of the general claim follows.
For any
invertible matrix
In linear algebra, an -by- square matrix is called invertible (also nonsingular or nondegenerate), if there exists an -by- square matrix such that
:\mathbf = \mathbf = \mathbf_n \
where denotes the -by- identity matrix and the multiplicati ...
, in the previous section
"Via Chain Rule", we showed that
:
Considering
in this equation yields:
:
The desired result follows as the solution to this ordinary differential equation.
Applications
Several forms of the formula underlie the
Faddeev–LeVerrier algorithm for computing the
characteristic polynomial
In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The ...
, and explicit applications of the
Cayley–Hamilton theorem
In linear algebra, the Cayley–Hamilton theorem (named after the mathematicians Arthur Cayley and William Rowan Hamilton) states that every square matrix over a commutative ring (such as the real or complex numbers or the integers) satisfies ...
. For example, starting from the following equation, which was proved above:
:
and using
, we get:
: