HOME

TheInfoList



OR:

:''This article describes a theorem concerning the dual of a
Hilbert space In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise natural ...
. For the theorems relating
linear functionals In mathematics, a linear form (also known as a linear functional, a one-form, or a covector) is a linear map from a vector space to its field of scalars (often, the real numbers or the complex numbers). If is a vector space over a field , the ...
to
measures Measure may refer to: * Measurement, the assignment of a number to a characteristic of an object or event Law * Ballot measure, proposed legislation in the United States * Church of England Measure, legislation of the Church of England * Measu ...
, see
Riesz–Markov–Kakutani representation theorem In mathematics, the Riesz–Markov–Kakutani representation theorem relates linear functionals on spaces of continuous functions on a locally compact space to measures in measure theory. The theorem is named for who introduced it for continuo ...
.'' The Riesz representation theorem, sometimes called the Riesz–Fréchet representation theorem after
Frigyes Riesz Frigyes Riesz ( hu, Riesz Frigyes, , sometimes spelled as Frederic; 22 January 1880 – 28 February 1956) was a HungarianEberhard Zeidler: Nonlinear Functional Analysis and Its Applications: Linear monotone operators. Springer, 199/ref> mathema ...
and
Maurice René Fréchet Maurice may refer to: People *Saint Maurice (died 287), Roman legionary and Christian martyr *Maurice (emperor) or Flavius Mauricius Tiberius Augustus (539–602), Byzantine emperor *Maurice (bishop of London) (died 1107), Lord Chancellor and Lo ...
, establishes an important connection between a
Hilbert space In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise natural ...
and its
continuous dual space In mathematics, any vector space ''V'' has a corresponding dual vector space (or just dual space for short) consisting of all linear forms on ''V'', together with the vector space structure of pointwise addition and scalar multiplication by const ...
. If the underlying
field Field may refer to: Expanses of open ground * Field (agriculture), an area of land used for agricultural purposes * Airfield, an aerodrome that lacks the infrastructure of an airport * Battlefield * Lawn, an area of mowed grass * Meadow, a grass ...
is the
real number In mathematics, a real number is a number that can be used to measure a ''continuous'' one-dimensional quantity such as a distance, duration or temperature. Here, ''continuous'' means that values can have arbitrarily small variations. Every ...
s, the two are isometrically isomorphic; if the underlying field is the
complex number In mathematics, a complex number is an element of a number system that extends the real numbers with a specific element denoted , called the imaginary unit and satisfying the equation i^= -1; every complex number can be expressed in the fo ...
s, the two are isometrically anti-isomorphic. The (anti-)
isomorphism In mathematics, an isomorphism is a structure-preserving mapping between two structures of the same type that can be reversed by an inverse mapping. Two mathematical structures are isomorphic if an isomorphism exists between them. The word i ...
is a particular
natural isomorphism In category theory, a branch of mathematics, a natural transformation provides a way of transforming one functor into another while respecting the internal structure (i.e., the composition of morphisms) of the categories involved. Hence, a natur ...
.


Preliminaries and notation

Let H be a
Hilbert space In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise natural ...
over a field \mathbb, where \mathbb is either the real numbers \R or the complex numbers \Complex. If \mathbb = \Complex (resp. if \mathbb = \R) then H is called a (resp. a ). Every real Hilbert space can be extended to be a dense subset of a unique (up to
bijective In mathematics, a bijection, also known as a bijective function, one-to-one correspondence, or invertible function, is a function between the elements of two sets, where each element of one set is paired with exactly one element of the other ...
isometry In mathematics, an isometry (or congruence, or congruent transformation) is a distance-preserving transformation between metric spaces, usually assumed to be bijective. The word isometry is derived from the Ancient Greek: ἴσος ''isos'' me ...
) complex Hilbert space, called its
complexification In mathematics, the complexification of a vector space over the field of real numbers (a "real vector space") yields a vector space over the complex number field, obtained by formally extending the scaling of vectors by real numbers to include ...
, which is why Hilbert spaces are often automatically assumed to be complex. Real and complex Hilbert spaces have in common many, but by no means all, properties and results/theorems. This article is intended for both mathematicians and physicists and will describe the theorem for both. In both mathematics and physics, if a Hilbert space is assumed to be real (that is, if \mathbb = \R) then this will usually be made clear. Often in mathematics, and especially in physics, unless indicated otherwise, "Hilbert space" is usually automatically assumed to mean "complex Hilbert space." Depending on the author, in mathematics, "Hilbert space" usually means either (1) a complex Hilbert space, or (2) a real complex Hilbert space.


Linear and antilinear maps

By definition, an (also called a ) f : H \to Y is a map between vector spaces that is : f(x + y) = f(x) + f(y) \quad \text x, y \in H, and (also called or ): f(c x) = \overline f(x) \quad \text x \in H \text c \in \mathbb. In contrast, a map f : H \to Y is
linear Linearity is the property of a mathematical relationship ('' function'') that can be graphically represented as a straight line. Linearity is closely related to '' proportionality''. Examples in physics include rectilinear motion, the linear ...
if it is additive and : f(c x) = c f(x) \quad \text x \in H \quad \text c \in \mathbb. Every constant 0 map is always both linear and antilinear. If \mathbb = \R then the definitions of linear maps and antilinear maps are completely identical. A linear map from a Hilbert space into a Banach space (or more generally, from any Banach space into any
topological vector space In mathematics, a topological vector space (also called a linear topological space and commonly abbreviated TVS or t.v.s.) is one of the basic structures investigated in functional analysis. A topological vector space is a vector space that is als ...
) is
continuous Continuity or continuous may refer to: Mathematics * Continuity (mathematics), the opposing concept to discreteness; common examples include ** Continuous probability distribution or random variable in probability and statistics ** Continuous ...
if and only if it is bounded; the same is true of antilinear maps. The inverse of any antilinear (resp. linear) bijection is again an antilinear (resp. linear) bijection. The composition of two linear maps is a map. Continuous dual and anti-dual spaces A on H is a function H \to \mathbb whose
codomain In mathematics, the codomain or set of destination of a function is the set into which all of the output of the function is constrained to fall. It is the set in the notation . The term range is sometimes ambiguously used to refer to either th ...
is the underlying scalar field \mathbb. Denote by H^* (resp. by \overline^*) the set of all continuous linear (resp. continuous antilinear) functionals on H, which is called the (resp. the ) of H. If \mathbb = \R then linear functionals on H are the same as antilinear functionals and consequently, the same is true for such continuous maps: that is, H^* = \overline^*. One-to-one correspondence between linear and antilinear functionals Given any functional f ~:~ H \to \mathbb, the is the functional \begin \overline : \,& H && \to \,&& \mathbb \\ & h && \mapsto\,&& \overline. \\ \end This assignment is most useful when \mathbb = \Complex because if \mathbb = \R then f = \overline and the assignment f \mapsto \overline reduces down to the identity map. The assignment f \mapsto \overline defines an antilinear
bijective In mathematics, a bijection, also known as a bijective function, one-to-one correspondence, or invertible function, is a function between the elements of two sets, where each element of one set is paired with exactly one element of the other ...
correspondence from the set of :all functionals (resp. all linear functionals, all continuous linear functionals H^*) on H, onto the set of :all functionals (resp. all linear functionals, all continuous linear functionals \overline^*) on H.


Mathematics vs. physics notations and definitions of inner product

The
Hilbert space In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise natural ...
H has an associated
inner product In mathematics, an inner product space (or, rarely, a Hausdorff pre-Hilbert space) is a real vector space or a complex vector space with an operation called an inner product. The inner product of two vectors in the space is a scalar, often ...
H \times H \to \mathbb valued in H's underlying scalar field \mathbb that is linear in one coordinate and antilinear in the other (as described in detail below). If H is a complex Hilbert space (meaning, if \mathbb = \Complex), which is very often the case, then which coordinate is antilinear and which is linear becomes a important technicality. However, if \mathbb = \R then the inner product is a
symmetric Symmetry (from grc, συμμετρία "agreement in dimensions, due proportion, arrangement") in everyday language refers to a sense of harmonious and beautiful proportion and balance. In mathematics, "symmetry" has a more precise definiti ...
map that is simultaneously linear in each coordinate (that is, bilinear) and antilinear in each coordinate. Consequently, the question of which coordinate is linear and which is antilinear is irrelevant for
real Real may refer to: Currencies * Brazilian real (R$) * Central American Republic real * Mexican real * Portuguese real * Spanish real * Spanish colonial real Music Albums * ''Real'' (L'Arc-en-Ciel album) (2000) * ''Real'' (Bright album) (2010) ...
Hilbert spaces. Notation for the inner product In mathematics, the inner product on a Hilbert space H is often denoted by \left\langle \cdot, \cdot \right\rangle or \left\langle \cdot, \cdot \right\rangle_H while in
physics Physics is the natural science that studies matter, its fundamental constituents, its motion and behavior through space and time, and the related entities of energy and force. "Physical science is that department of knowledge which r ...
, the bra–ket notation \left\langle \cdot \mid \cdot \right\rangle or \left\langle \cdot \mid \cdot \right\rangle_H is typically used instead. In this article, these two notations will be related by the equality: \left\langle x, y \right\rangle := \left\langle y \mid x \right\rangle \quad \text x, y \in H. Competing definitions of the inner product The maps \left\langle \cdot, \cdot \right\rangle and \left\langle \cdot \mid \cdot \right\rangle are assumed to have the following two properties:
  1. The map \left\langle \cdot, \cdot \right\rangle is in its coordinate; equivalently, the map \left\langle \cdot \mid \cdot \right\rangle is linear in its coordinate. Explicitly, this means that for every fixed y \in H, the map that is denoted by \left\langle \,y\mid \cdot\, \right\rangle = \left\langle \,\cdot, y\, \right\rangle : H \to \mathbb and defined by h \mapsto \left\langle \,y\mid h\, \right\rangle = \left\langle \,h, y\, \right\rangle \quad \text h \in H is a linear functional on H. * In fact, this linear functional is continuous, so \left\langle \,y\mid\cdot\, \right\rangle = \left\langle \,\cdot, y\, \right\rangle \in H^*.
  2. The map \left\langle \cdot, \cdot \right\rangle is
    linear Linearity is the property of a mathematical relationship ('' function'') that can be graphically represented as a straight line. Linearity is closely related to '' proportionality''. Examples in physics include rectilinear motion, the linear ...
    in its coordinate; equivalently, the map \left\langle \cdot \mid \cdot \right\rangle is linear in its coordinate. Explicitly, this means that for every fixed y \in H, the map that is denoted by \left\langle \,\cdot\mid y\, \right\rangle = \left\langle \,y, \cdot\, \right\rangle : H \to \mathbb and defined by h \mapsto \left\langle \,h\mid y\, \right\rangle = \left\langle \,y, h\, \right\rangle \quad \text h \in H is an antilinear functional on H. * In fact, this antilinear functional is continuous, so \left\langle \,\cdot\mid y\, \right\rangle = \left\langle \,y, \cdot\, \right\rangle \in \overline^*.
In mathematics, the prevailing convention (i.e. the definition of an inner product) is that the inner product is coordinate and antilinear in the other coordinate. In
physics Physics is the natural science that studies matter, its fundamental constituents, its motion and behavior through space and time, and the related entities of energy and force. "Physical science is that department of knowledge which r ...
, the convention/definition is unfortunately the , meaning that the inner product is coordinate and antilinear in the other coordinate. This article will not choose one definition over the other. Instead, the assumptions made above make it so that the mathematics notation \left\langle \cdot, \cdot \right\rangle satisfies the mathematical convention/definition for the inner product (that is, linear in the first coordinate and antilinear in the other), while the physics bra–ket notation \left\langle \cdot , \cdot \right\rangle satisfies the physics convention/definition for the inner product (that is, linear in the second coordinate and antilinear in the other). Consequently, the above two assumptions makes the notation used in each field consistent with that field's convention/definition for which coordinate is linear and which is antilinear.


Canonical norm and inner product on the dual space and anti-dual space

If x = y then \langle \,x\mid x\, \rangle = \langle \,x, x\, \rangle is a non-negative real number and the map \, x\, := \sqrt = \sqrt defines a canonical norm on H that makes H into a normed space. As with all normed spaces, the (continuous) dual space H^* carries a canonical norm, called the , that is defined by \, f\, _ ~:=~ \sup_ , f(x), \quad \text f \in H^*. The canonical norm on the (continuous)
anti-dual space In mathematics, a function f : V \to W between two complex vector spaces is said to be antilinear or conjugate-linear if \begin f(x + y) &= f(x) + f(y) && \qquad \text \\ f(s x) &= \overline f(x) && \qquad \text \\ \end hold for all vectors x, y ...
\overline^*, denoted by \, f\, _, is defined by using this same equation: \, f\, _ ~:=~ \sup_ , f(x), \quad \text f \in \overline^*. This canonical norm on H^* satisfies the
parallelogram law In mathematics, the simplest form of the parallelogram law (also called the parallelogram identity) belongs to elementary geometry. It states that the sum of the squares of the lengths of the four sides of a parallelogram equals the sum of the s ...
, which means that the polarization identity can be used to define a which this article will denote by the notations \left\langle f, g \right\rangle_ := \left\langle g \mid f \right\rangle_, where this inner product turns H^* into a Hilbert space. There are now two ways of defining a norm on H^*: the norm induced by this inner product (that is, the norm defined by f \mapsto \sqrt) and the usual
dual norm In functional analysis, the dual norm is a measure of size for a continuous linear function defined on a normed vector space. Definition Let X be a normed vector space with norm \, \cdot\, and let X^* denote its continuous dual space. The dual ...
(defined as the supremum over the closed unit ball). These norms are the same; explicitly, this means that the following holds for every f \in H^*: \sup_ , f(x), = \, f\, _ ~=~ \sqrt ~=~ \sqrt. As will be described later, the Riesz representation theorem can be used to give an equivalent definition of the canonical norm and the canonical inner product on H^*. The same equations that were used above can also be used to define a norm and inner product on H's
anti-dual space In mathematics, a function f : V \to W between two complex vector spaces is said to be antilinear or conjugate-linear if \begin f(x + y) &= f(x) + f(y) && \qquad \text \\ f(s x) &= \overline f(x) && \qquad \text \\ \end hold for all vectors x, y ...
\overline^*. Canonical isometry between the dual and antidual The
complex conjugate In mathematics, the complex conjugate of a complex number is the number with an equal real part and an imaginary part equal in magnitude but opposite in sign. That is, (if a and b are real, then) the complex conjugate of a + bi is equal to a - ...
\overline of a functional f, which was defined above, satisfies \, f\, _ ~=~ \left\, \overline\right\, _ \quad \text \quad \left\, \overline\right\, _ ~=~ \, g\, _ for every f \in H^* and every g \in \overline^*. This says exactly that the canonical antilinear bijection defined by \begin \operatorname :\;&& H^* &&\;\to \;& \overline^* \\ .3ex && f &&\;\mapsto\;& \overline \\ \end as well as its inverse \operatorname^ ~:~ \overline^* \to H^* are antilinear
isometries In mathematics, an isometry (or congruence, or congruent transformation) is a distance-preserving transformation between metric spaces, usually assumed to be bijective. The word isometry is derived from the Ancient Greek: ἴσος ''isos'' mea ...
and consequently also
homeomorphism In the mathematical field of topology, a homeomorphism, topological isomorphism, or bicontinuous function is a bijective and continuous function between topological spaces that has a continuous inverse function. Homeomorphisms are the isomor ...
s. The inner products on the dual space H^* and the anti-dual space \overline^*, denoted respectively by \langle \,\cdot\,, \,\cdot\, \rangle_ and \langle \,\cdot\,, \,\cdot\, \rangle_, are related by \langle \,\overline\, , \,\overline\, \rangle_ = \overline = \langle \,g\, , \,f\, \rangle_ \qquad \text f, g \in H^* and \langle \,\overline\, , \,\overline\, \rangle_ = \overline = \langle \,g\, , \,f\, \rangle_ \qquad \text f, g \in \overline^*. If \mathbb = \R then H^* = \overline^* and this canonical map \operatorname : H^* \to \overline^* reduces down to the identity map.


Riesz representation theorem

Two vectors x and y are if \langle x, y \rangle = 0, which happens if and only if \, y\, \leq \, y + s x\, for all scalars s. The
orthogonal complement In the mathematical fields of linear algebra and functional analysis, the orthogonal complement of a subspace ''W'' of a vector space ''V'' equipped with a bilinear form ''B'' is the set ''W''⊥ of all vectors in ''V'' that are orthogonal to every ...
of a subset C \subseteq H is C^ := \, which is always a closed vector subspace of H. The
Hilbert projection theorem In mathematics, the Hilbert projection theorem is a famous result of convex analysis that says that for every vector x in a Hilbert space H and every nonempty closed convex C \subseteq H, there exists a unique vector m \in C for which \, c - x\, ...
guarantees that for any nonempty closed convex subset C of a
Hilbert space In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise natural ...
there exists a unique vector m \in C such that \, m\, = \inf_ \, c\, ; that is, m \in C is the (unique)
global minimum point In mathematical analysis, the maxima and minima (the respective plurals of maximum and minimum) of a function, known collectively as extrema (the plural of extremum), are the largest and smallest value of the function, either within a given ra ...
of the function C \to c\, .


Statement

Historically, the theorem is often attributed simultaneously to Riesz and Maurice René Fréchet, Fréchet in 1907 (see references). Let \mathbb denote the underlying scalar field of H. Fix y \in H. Define \Lambda : H \to \mathbb by \Lambda(z) := \langle \,y\, , \,z\, \rangle, which is a linear functional on H since z is in the linear argument. By the Cauchy–Schwarz inequality, , \Lambda(z), = , \langle \,y\, , \,z\, \rangle, \leq \, y\, \, z\, which shows that \Lambda is bounded (equivalently, Continuous linear functional, continuous) and that \, \Lambda\, \leq \, y\, . It remains to show that \, y\, \leq \, \Lambda\, . By using y in place of z, it follows that \, y\, ^2 = \langle \,y\, , \,y\, \rangle = \Lambda y = , \Lambda(y), \leq \, \Lambda\, \, y\, (the equality \Lambda y = , \Lambda(y), holds because \Lambda y = \, y\, ^2 \geq 0 is real and non-negative). Thus that \, \Lambda\, = \, y\, . \blacksquare The proof above did not use the fact that H is Complete metric space, complete, which shows that the formula for the norm \, \langle \,y\, , \,\cdot\, \rangle\, _ = \, y\, _H holds more generally for all
inner product space In mathematics, an inner product space (or, rarely, a Hausdorff pre-Hilbert space) is a real vector space or a complex vector space with an operation called an inner product. The inner product of two vectors in the space is a scalar, often ...
s. Suppose f, g \in H are such that \varphi(z) = \langle \,f\, , \,z\, \rangle and \varphi(z) = \langle \,g\, , \,z\, \rangle for all z \in H. Then \langle \,f - g\, , \,z\, \rangle = \langle \,f\, , \,z\, \rangle - \langle \,g\, , \,z\, \rangle = \varphi(z) - \varphi(z) = 0 \quad \text z \in H which shows that \Lambda := \langle \,f - g\, , \,\cdot\, \rangle is the constant 0 linear functional. Consequently 0 = \, \langle \,f - g\, , \,\cdot\, \rangle\, = \, f - g\, , which implies that f - g = 0. \blacksquare Let K := \ker \varphi := \. If K = H (or equivalently, if \varphi = 0) then taking f_ := 0 completes the proof so assume that K \neq H and \varphi \neq 0. The continuity of \varphi implies that K is a closed subspace of H (because K = \varphi^(\) and \ is a closed subset of \mathbb). Let K^ := \ denote the
orthogonal complement In the mathematical fields of linear algebra and functional analysis, the orthogonal complement of a subspace ''W'' of a vector space ''V'' equipped with a bilinear form ''B'' is the set ''W''⊥ of all vectors in ''V'' that are orthogonal to every ...
of K in H. Because K is closed and H is a Hilbert space,Showing that there is a non-zero vector v in K^ relies on the continuity of \phi and the
Cauchy completeness In mathematical analysis, a metric space is called complete (or a Cauchy space) if every Cauchy sequence of points in has a limit that is also in . Intuitively, a space is complete if there are no "points missing" from it (inside or at the boun ...
of H. This is the only place in the proof in which these properties are used.
H can be written as the direct sum H = K \oplus K^Technically, H = K \oplus K^ means that the addition map K \times K^ \to H defined by (k, p) \mapsto k + p is a surjective
linear isomorphism In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that pre ...
and
homeomorphism In the mathematical field of topology, a homeomorphism, topological isomorphism, or bicontinuous function is a bijective and continuous function between topological spaces that has a continuous inverse function. Homeomorphisms are the isomor ...
. See the article on
complemented subspace In the branch of mathematics called functional analysis, a complemented subspace of a topological vector space X, is a vector subspace M for which there exists some other vector subspace N of X, called its (topological) complement in X, such that ...
s for more details.
(a proof of this is given in the article on the
Hilbert projection theorem In mathematics, the Hilbert projection theorem is a famous result of convex analysis that says that for every vector x in a Hilbert space H and every nonempty closed convex C \subseteq H, there exists a unique vector m \in C for which \, c - x\, ...
). Because K \neq H, there exists some non-zero p \in K^. For any h \in H, \varphi \varphi h) p - (\varphi p) h ~=~ \varphi \varphi h) p- \varphi \varphi p) h ~=~ (\varphi h) \varphi p - (\varphi p) \varphi h = 0, which shows that (\varphi h) p - (\varphi p) h ~\in~ \ker \varphi = K, where now p \in K^ implies 0 = \langle \,p\, , \,(\varphi h) p - (\varphi p) h\, \rangle ~=~ \langle \,p\, , \,(\varphi h) p \, \rangle - \langle \,p\, , \,(\varphi p) h\, \rangle ~=~ (\varphi h) \langle \,p\, , \,p \, \rangle - (\varphi p) \langle \,p\, , \,h\, \rangle. Solving for \varphi h shows that \varphi h = \frac = \left\langle \,\frac p\, \Bigg, \,h\, \right\rangle \quad \text h \in H, which proves that the vector f_ := \frac p satisfies \varphi h = \langle \,f_\, , \,h\, \rangle \text h \in H. Applying the norm formula that was proved above with y := f_ shows that \, \varphi\, _ = \left\, \left\langle \,f_\, , \,\cdot\, \right\rangle\right\, _ = \left\, f_\right\, _H. Also, the vector u := \frac has norm \, u\, = 1 and satisfies f_ := \overline u. \blacksquare It can now be deduced that K^ is 1-dimensional when \varphi \neq 0. Let q \in K^ be any non-zero vector. Replacing p with q in the proof above shows that the vector g := \frac q satisfies \varphi(h) = \langle \,g\, , \,h\, \rangle for every h \in H. The uniqueness of the (non-zero) vector f_ representing \varphi implies that f_ = g, which in turn implies that \overline \neq 0 and q = \frac f_. Thus every vector in K^ is a scalar multiple of f_. \blacksquare The formulas for the inner products follow from the polarization identity.


Observations

If \varphi \in H^* then \varphi \left(f_\right) = \left\langle f_, f_ \right\rangle = \left\, f_\right\, ^2 = \, \varphi\, ^2. So in particular, \varphi \left(f_\right) \geq 0 is always real and furthermore, \varphi \left(f_\right) = 0 if and only if f_ = 0 if and only if \varphi = 0. Linear functionals as affine hyperplanes A non-trivial continuous linear functional \varphi is often interpreted geometrically by identifying it with the affine hyperplane A := \varphi^(1) (the kernel \ker\varphi = \varphi^(0) is also often visualized alongside A := \varphi^(1) although knowing A is enough to reconstruct \ker \varphi because if A = \varnothing then \ker \varphi = H and otherwise \ker \varphi = A - A). In particular, the norm of \varphi should somehow be interpretable as the "norm of the hyperplane A". When \varphi \neq 0 then the Riesz representation theorem provides such an interpretation of \, \varphi\, in terms of the affine hyperplane A := \varphi^(1) as follows: using the notation from the theorem's statement, from \, \varphi\, ^2 \neq 0 it follows that C := \varphi^\left(\, \varphi\, ^2\right) = \, \varphi\, ^2 \varphi^(1) = \, \varphi\, ^2 A and so \, \varphi\, = \left\, f_\right\, = \inf_ \, c\, implies \, \varphi\, = \inf_ \, \varphi\, ^2 \, a\, and thus \, \varphi\, = \frac. This can also be seen by applying the
Hilbert projection theorem In mathematics, the Hilbert projection theorem is a famous result of convex analysis that says that for every vector x in a Hilbert space H and every nonempty closed convex C \subseteq H, there exists a unique vector m \in C for which \, c - x\, ...
to A and concluding that the global minimum point of the map A \to [0, \infty) defined by a \mapsto \, a\, is \frac \in A. The formulas \frac = \sup_ \frac provide the promised interpretation of the linear functional's norm \, \varphi\, entirely in terms of its associated affine hyperplane A = \varphi^(1) (because with this formula, knowing only the A is enough to describe the norm of its associated linear ). Defining \frac := 0, the infimum formula \, \varphi\, = \frac will also hold when \varphi = 0. When the supremum is taken in \R (as is typically assumed), then the supremum of the empty set is \sup \varnothing = - \infty but if the supremum is taken the non-negative reals image/range of the norm \, \,\cdot\,\, when \dim H > 0) then this supremum is instead \sup \varnothing = 0, in which case the supremum formula \, \varphi\, = \sup_ \frac will also hold when \varphi = 0 (although the atypical equality \sup \varnothing = 0 is usually unexpected and so risks causing confusion).


Constructions of the representing vector

Using the notation from the theorem above, several ways of constructing f_ from \varphi \in H^* are now described. If \varphi = 0 then f_ := 0; in other words, f_0 = 0. This special case of \varphi = 0 is henceforth assumed to be known, which is why some of the constructions given below start by assuming \varphi \neq 0. Orthogonal complement of kernel If \varphi \neq 0 then for any 0 \neq u \in (\ker\varphi)^, f_ := \frac. If u \in (\ker\varphi)^ is a unit vector (meaning \, u\, = 1) then f_ := \overline u (this is true even if \varphi = 0 because in this case f_ = \overline u = \overline u = 0). If u is a unit vector satisfying the above condition then the same is true of -u, which is also a unit vector in (\ker\varphi)^. However, \overline (-u) = \overline u = f_\varphi so both these vectors result in the same f_. Orthogonal projection onto kernel If x \in H is such that \varphi(x) \neq 0 and if x_K is the
orthogonal projection In linear algebra and functional analysis, a projection is a linear transformation P from a vector space to itself (an endomorphism) such that P\circ P=P. That is, whenever P is applied twice to any vector, it gives the same result as if it wer ...
of x onto \ker\varphi then f_ = \frac \left(x - x_K\right). Orthonormal basis Given an
orthonormal basis In mathematics, particularly linear algebra, an orthonormal basis for an inner product space ''V'' with finite dimension is a basis for V whose vectors are orthonormal, that is, they are all unit vectors and orthogonal to each other. For examp ...
\left\_ of H and a continuous linear functional \varphi \in H^*, the vector f_ \in H can be constructed uniquely by f_\varphi = \sum_ \overline e_i where all but at most countably many \varphi\left(e_i\right) will be equal to 0 and where the value of f_ does not actually depend on choice of orthonormal basis (that is, using any other orthonormal basis for H will result in the same vector). If y \in H is written as y = \sum_ a_i e_i then \varphi(y) = \sum_ \varphi\left(e_i\right) a_i = \langle f_ , y \rangle and \left\, f_\right\, ^2 = \varphi\left(f_\right) = \sum_ \varphi\left(e_i\right) \overline = \sum_ \left, \varphi\left(e_i\right)\^2 = \, \varphi\, ^2. If the orthonormal basis \left\_ = \left\_^ is a sequence then this becomes f_\varphi = \overline e_1 + \overline e_2 + \cdots and if y \in H is written as y = \sum_ a_i e_i = a_1 e_1 + a_2 e_2 + \cdots then \varphi(y) = \varphi\left(e_1\right) a_1 + \varphi\left(e_2\right) a_2 + \cdots = \langle f_ , y \rangle.


Example in finite dimensions using matrix transformations

Consider the special case of H = \Complex^n (where n > 0 is an integer) with the standard inner product \langle z \mid w \rangle := \overline^ \vec \qquad \text \; w, z \in H where w \text z are represented as column matrices \vec := \beginw_1 \\ \vdots \\ w_n\end and \vec := \beginz_1 \\ \vdots \\ z_n\end with respect to the standard orthonormal basis e_1, \ldots, e_n on H (here, e_i is 1 at its ith coordinate and 0 everywhere else; as usual, H^* will now be associated with the
dual basis In linear algebra, given a vector space ''V'' with a basis ''B'' of vectors indexed by an index set ''I'' (the cardinality of ''I'' is the dimension of ''V''), the dual set of ''B'' is a set ''B''∗ of vectors in the dual space ''V''∗ with the ...
) and where \overline^ := \left overline, \ldots, \overline\right/math> denotes the
conjugate transpose In mathematics, the conjugate transpose, also known as the Hermitian transpose, of an m \times n complex matrix \boldsymbol is an n \times m matrix obtained by transposing \boldsymbol and applying complex conjugate on each entry (the complex c ...
of \vec. Let \varphi \in H^* be any linear functional and let \varphi_1, \ldots, \varphi_n \in \Complex be the unique scalars such that \varphi\left(w_1, \ldots, w_n\right) = \varphi_1 w_1 + \cdots + \varphi_n w_n \qquad \text \; w := \left(w_1, \ldots, w_n\right) \in H, where it can be shown that \varphi_i = \varphi\left(e_i\right) for all i = 1, \ldots, n. Then the Riesz representation of \varphi is the vector f_ ~:=~ \overline e_1 + \cdots + \overline e_n ~=~ \left(\overline, \ldots, \overline\right) \in H. To see why, identify every vector w = \left(w_1, \ldots, w_n\right) in H with the column matrix \vec := \beginw_1 \\ \vdots \\ w_n\end so that f_ is identified with \vec := \begin\overline \\ \vdots \\ \overline\end = \begin\overline \\ \vdots \\ \overline\end. As usual, also identify the linear functional \varphi with its
transformation matrix In linear algebra, linear transformations can be represented by matrices. If T is a linear transformation mapping \mathbb^n to \mathbb^m and \mathbf x is a column vector with n entries, then T( \mathbf x ) = A \mathbf x for some m \times n matrix ...
, which is the
row matrix In linear algebra, a column vector with m elements is an m \times 1 matrix consisting of a single column of m entries, for example, \boldsymbol = \begin x_1 \\ x_2 \\ \vdots \\ x_m \end. Similarly, a row vector is a 1 \times n matrix for some n, c ...
\vec := \left varphi_1, \ldots, \varphi_n\right/math> so that \vec := \overline^ and the function \varphi is the assignment \vec \mapsto \vec \, \vec, where the right hand side is
matrix multiplication In mathematics, particularly in linear algebra, matrix multiplication is a binary operation that produces a matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the number of rows in the s ...
. Then for all w = \left(w_1, \ldots, w_n\right) \in H, \varphi(w) = \varphi_1 w_1 + \cdots + \varphi_n w_n = \left varphi_1, \ldots, \varphi_n\right \beginw_1 \\ \vdots \\ w_n\end = \overline^ \vec = \overline^ \vec = \left\langle \,\,f_\, \mid \,w\, \right\rangle, which shows that f_ satisfies the defining condition of the Riesz representation of \varphi. The bijective antilinear isometry \Phi : H \to H^* defined in the corollary to the Riesz representation theorem is the assignment that sends z = \left(z_1, \ldots, z_n\right) \in H to the linear functional \Phi(z) \in H^* on H defined by w = \left(w_1, \ldots, w_n\right) ~\mapsto~ \langle \,z\, \mid \,w\,\rangle = \overline w_1 + \cdots + \overline w_n, where under the identification of vectors in H with column matrices and vector in H^* with row matrices, \Phi is just the assignment \vec = \beginz_1 \\ \vdots \\ z_n\end ~\mapsto~ \overline^ = \left overline, \ldots, \overline\right As described in the corollary, \Phi's inverse \Phi^ : H^* \to H is the antilinear isometry \varphi \mapsto f_, which was just shown above to be: \varphi ~\mapsto~ f_ ~:=~ \left(\overline, \ldots, \overline\right); where in terms of matrices, \Phi^ is the assignment \vec = \left varphi_1, \ldots, \varphi_n\right~\mapsto~ \overline^ = \begin\overline \\ \vdots \\ \overline\end. Thus in terms of matrices, each of \Phi : H \to H^* and \Phi^ : H^* \to H is just the operation of conjugate transposition \vec \mapsto \overline^ (although between different spaces of matrices: if H is identified with the space of all column (respectively, row) matrices then H^* is identified with the space of all row (respectively, column) matrices). This example used the standard inner product, which is the map \langle z \mid w \rangle := \overline^ \vec, but if a different inner product is used, such as \langle z \mid w \rangle_M := \overline^ \, M \, \vec \, where M is any
Hermitian {{Short description, none Numerous things are named after the French mathematician Charles Hermite (1822–1901): Hermite * Cubic Hermite spline, a type of third-degree spline * Gauss–Hermite quadrature, an extension of Gaussian quadrature m ...
positive-definite matrix In mathematics, a symmetric matrix M with real entries is positive-definite if the real number z^\textsfMz is positive for every nonzero real column vector z, where z^\textsf is the transpose of More generally, a Hermitian matrix (that is, a ...
, or if a different orthonormal basis is used then the transformation matrices, and thus also the above formulas, will be different.


Relationship with the associated real Hilbert space

Assume that H is a complex Hilbert space with inner product \langle \,\cdot\mid\cdot\, \rangle. When the Hilbert space H is reinterpreted as a real Hilbert space then it will be denoted by H_, where the (real) inner-product on H_ is the real part of H's inner product; that is: \langle x, y \rangle_ := \operatorname \langle x, y \rangle. The norm on H_ induced by \langle \,\cdot\,, \,\cdot\, \rangle_ is equal to the original norm on H and the continuous dual space of H_ is the set of all -valued bounded \R-linear functionals on H_ (see the article about the polarization identity for additional details about this relationship). Let \psi_ := \operatorname \psi and \psi_ := \operatorname \psi denote the real and imaginary parts of a linear functional \psi, so that \psi = \operatorname \psi + i \operatorname \psi = \psi_ + i \psi_. The formula expressing a linear functional in terms of its real part is \psi(h) = \psi_(h) - i \psi_ (i h) \quad \text h \in H, where \psi_(h) = - i \psi_ (i h) for all h \in H. It follows that \ker\psi_ = \psi^(i \R), and that \psi = 0 if and only if \psi_ = 0. It can also be shown that \, \psi\, = \left\, \psi_\right\, = \left\, \psi_i\right\, where \left\, \psi_\right\, := \sup_ \left, \psi_(h)\ and \left\, \psi_i\right\, := \sup_ \left, \psi_i(h)\ are the usual
operator norm In mathematics, the operator norm measures the "size" of certain linear operators by assigning each a real number called its . Formally, it is a norm defined on the space of bounded linear operators between two given normed vector spaces. Introd ...
s. In particular, a linear functional \psi is bounded if and only if its real part \psi_ is bounded. Representing a functional and its real part The Riesz representation of a continuous linear function \varphi on a complex Hilbert space is equal to the Riesz representation of its real part \operatorname \varphi on its associated real Hilbert space. Explicitly, let \varphi \in H^* and as above, let f_\varphi \in H be the Riesz representation of \varphi obtained in (H, \langle, \cdot, \cdot \rangle), so it is the unique vector that satisfies \varphi(x) = \left\langle f_ \mid x \right\rangle for all x \in H. The real part of \varphi is a continuous real linear functional on H_ and so the Riesz representation theorem may be applied to \varphi_ := \operatorname \varphi and the associated real Hilbert space \left(H_, \langle, \cdot, \cdot \rangle_\right) to produce its Riesz representation, which will be denoted by f_. That is, f_ is the unique vector in H_ that satisfies \varphi_(x) = \left\langle f_ \mid x \right\rangle_ for all x \in H. The conclusion is f_ = f_. This follows from the main theorem because \ker\varphi_ = \varphi^(i \R) and if x \in H then \left\langle f_\varphi \mid x \right\rangle_ = \operatorname \left\langle f_\varphi \mid x \right\rangle = \operatorname \varphi(x) = \varphi_(x) and consequently, if m \in \ker\varphi_ then \left\langle f_\mid m \right\rangle_ = 0, which shows that f_ \in (\ker\varphi_)^. Moreover, \varphi(f_\varphi) = \, \varphi\, ^2 being a real number implies that \varphi_ (f_\varphi) = \operatorname \varphi(f_\varphi) = \, \varphi\, ^2. In other words, in the theorem and constructions above, if H is replaced with its real Hilbert space counterpart H_ and if \varphi is replaced with \operatorname \varphi then f_ = f_. This means that vector f_ obtained by using \left(H_, \langle, \cdot, \cdot \rangle_\right) and the real linear functional \operatorname \varphi is the equal to the vector obtained by using the origin complex Hilbert space \left(H, \left\langle, \cdot, \cdot \right\rangle\right) and original complex linear functional \varphi (with identical norm values as well). Furthermore, if \varphi \neq 0 then f_ is perpendicular to \ker\varphi_ with respect to \langle \cdot, \cdot \rangle_ where the kernel of \varphi is be a ''proper'' subspace of the kernel of its real part \varphi_. Assume now that \varphi \neq 0. Then f_ \not\in \ker\varphi_ because \varphi_\left(f_\right) = \varphi\left(f_\right) = \, \varphi\, ^2 \neq 0 and \ker\varphi is a proper subset of \ker\varphi_. The vector subspace \ker \varphi has real codimension 1 in \ker\varphi_, while \ker\varphi_ has codimension 1 in H_, and \left\langle f_, \ker\varphi_ \right\rangle_ = 0. That is, f_ is perpendicular to \ker\varphi_ with respect to \langle \cdot, \cdot \rangle_.


Canonical injections into the dual and anti-dual

Induced linear map into anti-dual The map defined by placing y into the coordinate of the inner product and letting the variable h \in H vary over the coordinate results in an functional: \langle \,\cdot \mid y\, \rangle = \langle \,y, \cdot\, \rangle : H \to \mathbb \quad \text \quad h \mapsto \langle \,h \mid y\, \rangle = \langle \,y, h\, \rangle. This map is an element of \overline^*, which is the continuous
anti-dual space In mathematics, a function f : V \to W between two complex vector spaces is said to be antilinear or conjugate-linear if \begin f(x + y) &= f(x) + f(y) && \qquad \text \\ f(s x) &= \overline f(x) && \qquad \text \\ \end hold for all vectors x, y ...
of H. The \overline^* is the operator \begin \operatorname_H^ :\;&& H &&\;\to \;& \overline^* \\ .3ex && y &&\;\mapsto\;& \langle \,\cdot \mid y\, \rangle = \langle \,y, \cdot\, \rangle \\ .3ex\end which is also an injective
isometry In mathematics, an isometry (or congruence, or congruent transformation) is a distance-preserving transformation between metric spaces, usually assumed to be bijective. The word isometry is derived from the Ancient Greek: ἴσος ''isos'' me ...
. The Fundamental theorem of Hilbert spaces, which is related to Riesz representation theorem, states that this map is surjective (and thus
bijective In mathematics, a bijection, also known as a bijective function, one-to-one correspondence, or invertible function, is a function between the elements of two sets, where each element of one set is paired with exactly one element of the other ...
). Consequently, every antilinear functional on H can be written (uniquely) in this form. If \operatorname : H^* \to \overline^* is the canonical
linear Linearity is the property of a mathematical relationship ('' function'') that can be graphically represented as a straight line. Linearity is closely related to '' proportionality''. Examples in physics include rectilinear motion, the linear ...
bijective In mathematics, a bijection, also known as a bijective function, one-to-one correspondence, or invertible function, is a function between the elements of two sets, where each element of one set is paired with exactly one element of the other ...
isometry In mathematics, an isometry (or congruence, or congruent transformation) is a distance-preserving transformation between metric spaces, usually assumed to be bijective. The word isometry is derived from the Ancient Greek: ἴσος ''isos'' me ...
f \mapsto \overline that was defined above, then the following equality holds: \operatorname ~\circ~ \operatorname_H^ ~=~ \operatorname_H^.


Extending the bra–ket notation to bras and kets

Let \left(H, \langle\cdot, \cdot \rangle_H\right) be a Hilbert space and as before, let \langle y\, , \,x \rangle_H := \langle x, y \rangle_H. Let \begin \Phi :\;&& H &&\;\to \;& H^* \\ .3ex && g &&\;\mapsto\;& \left\langle \,g\mid \cdot\, \right\rangle_H = \left\langle \,\cdot, g\, \right\rangle_H \\ \end which is a bijective antilinear isometry that satisfies (\Phi h) g = \langle h\mid g \rangle_H = \langle g, h \rangle_H \quad \text g, h \in H. Bras Given a vector h \in H, let \langle h\, , denote the continuous linear functional \Phi h; that is, \langle h\, , ~:=~ \Phi h so that this functional \langle h\, , is defined by g \mapsto \left\langle \,h\mid g\, \right\rangle_H. This map was denoted by \left\langle h \mid \cdot\, \right\rangle earlier in this article. The assignment h \mapsto \langle h , is just the isometric antilinear isomorphism \Phi ~:~ H \to H^*, which is why ~\langle c g + h\, , ~=~ \overline \langle g\mid ~+~ \langle h\, , ~ holds for all g, h \in H and all scalars c. The result of plugging some given g \in H into the functional \langle h\, , is the scalar \langle h\, , \,g \rangle_H = \langle g, h \rangle_H, which may be denoted by \langle h \mid g \rangle.The usual notation for plugging an element g into a linear map F is F(g) and sometimes Fg. Replacing F with \langle h\mid :=~ \Phi h produces \langle h\mid(g) or \langle h \mid g, which is unsightly (despite being consistent with the usual notation used with functions). Consequently, the symbol \,\rangle\, is appended to the end, so that the notation \langle h\mid g \rangle is used instead to denote this value (\Phi h) g. Bra of a linear functional Given a continuous linear functional \psi \in H^*, let \langle \psi\mid denote the vector \Phi^ \psi \in H; that is, \langle \psi\mid ~:=~ \Phi^ \psi. The assignment \psi \mapsto \langle \psi\mid is just the isometric antilinear isomorphism \Phi^ ~:~ H^* \to H, which is why ~\langle c \psi + \phi\mid ~=~ \overline \langle \psi\mid ~+~ \langle \phi\mid~ holds for all \phi, \psi \in H^* and all scalars c. The defining condition of the vector \langle \psi , \in H is the technically correct but unsightly equality \left\langle \, \langle \psi\mid \, \mid g \right\rangle_H ~=~ \psi g \quad \text g \in H, which is why the notation \left\langle \psi \mid g \right\rangle is used in place of \left\langle \, \langle \psi\mid \, \mid g \right\rangle_H = \left\langle g, \, \langle \psi\mid \right\rangle_H. With this notation, the defining condition becomes \left\langle \psi\mid g \right\rangle ~=~ \psi g \quad \text g \in H. Kets For any given vector g \in H, the notation , \,g \rangle is used to denote g; that is, \mid g \rangle : = g. The assignment g \mapsto , \,g \rangle is just the identity map \operatorname_H : H \to H, which is why ~\mid c g + h \rangle ~=~ c \mid g \rangle ~+~ \mid h \rangle~ holds for all g, h \in H and all scalars c. The notation \langle h\mid g \rangle and \langle \psi\mid g \rangle is used in place of \left\langle h\mid \, \mid g \rangle \, \right\rangle_H ~=~ \left\langle \mid g \rangle, h \right\rangle_H and \left\langle \psi\mid \, \mid g \rangle \, \right\rangle_H ~=~ \left\langle g, \, \langle \psi\mid \right\rangle_H, respectively. As expected, ~\langle \psi\mid g \rangle = \psi g~ and ~\langle h\mid g \rangle~ really is just the scalar ~\langle h\mid g \rangle_H ~=~ \langle g, h \rangle_H.


Adjoints and transposes

Let A : H \to Z be a
continuous linear operator In functional analysis and related areas of mathematics, a continuous linear operator or continuous linear mapping is a continuous linear transformation between topological vector spaces. An operator between two normed spaces is a bounded linear op ...
between
Hilbert spaces In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise naturally ...
\left(H, \langle \cdot, \cdot \rangle_H\right) and \left(Z, \langle \cdot, \cdot \rangle_Z \right). As before, let \langle y \mid x \rangle_H := \langle x, y \rangle_H and \langle y \mid x \rangle_Z := \langle x, y \rangle_Z. Denote by \begin \Phi_H :\;&& H &&\;\to \;& H^* \\ .3ex && g &&\;\mapsto\;& \langle \,g \mid \cdot\, \rangle_H \\ \end \quad \text \quad \begin \Phi_Z :\;&& Z &&\;\to \;& Z^* \\ .3ex && y &&\;\mapsto\;& \langle \,y \mid \cdot\, \rangle_Z \\ \end the usual bijective antilinear isometries that satisfy: \left(\Phi_H g\right) h = \langle g\mid h \rangle_H \quad \text g, h \in H \qquad \text \qquad \left(\Phi_Z y\right) z = \langle y \mid z \rangle_Z \quad \text y, z \in Z.


Definition of the adjoint

For every z \in Z, the scalar-valued map \langle z\mid A (\cdot) \rangle_Z on H defined by h \mapsto \langle z\mid A h \rangle_Z = \langle A h, z \rangle_Z is a continuous linear functional on H and so by the Riesz representation theorem, there exists a unique vector in H, denoted by A^* z, such that \langle z \mid A (\cdot) \rangle_Z = \left\langle A^* z \mid \cdot\, \right\rangle_H, or equivalently, such that \langle z \mid A h \rangle_Z = \left\langle A^* z \mid h \right\rangle_H \quad \text h \in H. The assignment z \mapsto A^* z thus induces a function A^* : Z \to H called the of A : H \to Z whose defining condition is \langle z \mid A h \rangle_Z = \left\langle A^* z\mid h \right\rangle_H \quad \text h \in H \text z \in Z. The adjoint A^* : Z \to H is necessarily a
continuous Continuity or continuous may refer to: Mathematics * Continuity (mathematics), the opposing concept to discreteness; common examples include ** Continuous probability distribution or random variable in probability and statistics ** Continuous ...
(equivalently, a bounded) linear operator. If H is finite dimensional with the standard inner product and if M is the
transformation matrix In linear algebra, linear transformations can be represented by matrices. If T is a linear transformation mapping \mathbb^n to \mathbb^m and \mathbf x is a column vector with n entries, then T( \mathbf x ) = A \mathbf x for some m \times n matrix ...
of A with respect to the standard orthonormal basis then M's
conjugate transpose In mathematics, the conjugate transpose, also known as the Hermitian transpose, of an m \times n complex matrix \boldsymbol is an n \times m matrix obtained by transposing \boldsymbol and applying complex conjugate on each entry (the complex c ...
\overline is the transformation matrix of the adjoint A^*.


Adjoints are transposes

It is also possible to define the or of A : H \to Z, which is the map ^A : Z^* \to H^* defined by sending a continuous linear functionals \psi \in Z^* to ^A(\psi) := \psi \circ A, where the
composition Composition or Compositions may refer to: Arts and literature *Composition (dance), practice and teaching of choreography *Composition (language), in literature and rhetoric, producing a work in spoken tradition and written discourse, to include v ...
\psi \circ A is always a continuous linear functional on H and it satisfies \, A\, = \left\, ^t A\right\, (this is true more generally, when H and Z are merely normed spaces). So for example, if z \in Z then ^A sends the continuous linear functional \langle z \mid \cdot \rangle_Z \in Z^* (defined on Z by g \mapsto \langle z \mid g \rangle_Z) to the continuous linear functional \langle z \mid A(\cdot) \rangle_Z \in H^* (defined on H by h \mapsto \langle z \mid A(h) \rangle_Z); using bra-ket notation, this can be written as ^A \langle z \mid ~=~ \langle z \mid A where the juxtaposition of \langle z \mid with A on the right hand side denotes function composition: H \xrightarrow Z \xrightarrow \Complex. The adjoint A^* : Z \to H is actually just to the transpose ^A : Z^* \to H^* when the Riesz representation theorem is used to identify Z with Z^* and H with H^*. Explicitly, the relationship between the adjoint and transpose is: which can be rewritten as: A^* ~=~ \Phi_H^ ~\circ~ ^A ~\circ~ \Phi_Z \quad \text \quad ^A ~=~ \Phi_H ~\circ~ A^* ~\circ~ \Phi_Z^. Alternatively, the value of the left and right hand sides of () at any given z \in Z can be rewritten in terms of the inner products as: \left(^A ~\circ~ \Phi_Z\right) z = \langle z \mid A (\cdot) \rangle_Z \quad \text \quad\left(\Phi_H ~\circ~ A^*\right) z = \langle A^* z\mid\cdot\, \rangle_H so that ^A ~\circ~ \Phi_Z ~=~ \Phi_H ~\circ~ A^* holds if and only if \langle z \mid A (\cdot) \rangle_Z = \langle A^* z\mid\cdot\, \rangle_H holds; but the equality on the right holds by definition of A^* z. The defining condition of A^* z can also be written \langle z \mid A ~=~ \langle A^*z \mid if bra-ket notation is used.


Descriptions of self-adjoint, normal, and unitary operators

Assume Z = H and let \Phi := \Phi_H = \Phi_Z. Let A : H \to H be a continuous (that is, bounded) linear operator. Whether or not A : H \to H is
self-adjoint In mathematics, and more specifically in abstract algebra, an element ''x'' of a *-algebra is self-adjoint if x^*=x. A self-adjoint element is also Hermitian, though the reverse doesn't necessarily hold. A collection ''C'' of elements of a st ...
,
normal Normal(s) or The Normal(s) may refer to: Film and television * ''Normal'' (2003 film), starring Jessica Lange and Tom Wilkinson * ''Normal'' (2007 film), starring Carrie-Anne Moss, Kevin Zegers, Callum Keith Rennie, and Andrew Airlie * ''Norma ...
, or
unitary Unitary may refer to: Mathematics * Unitary divisor * Unitary element * Unitary group * Unitary matrix * Unitary morphism * Unitary operator * Unitary transformation * Unitary representation * Unitarity (physics) * ''E''-unitary inverse semigrou ...
depends entirely on whether or not A satisfies certain defining conditions related to its adjoint, which was shown by () to essentially be just the transpose ^t A : H^* \to H^*. Because the transpose of A is a map between continuous linear functionals, these defining conditions can consequently be re-expressed entirely in terms of linear functionals, as the remainder of subsection will now describe in detail. The linear functionals that are involved are the simplest possible continuous linear functionals on H that can be defined entirely in terms of A, the inner product \langle \,\cdot\mid\cdot\, \rangle on H, and some given vector h \in H. Specifically, these are \left\langle A h\mid\cdot\, \right\rangle and \langle h\mid A (\cdot) \rangle where \left\langle A h\mid\cdot\, \right\rangle = \Phi (A h) = (\Phi \circ A) h \quad \text \quad \langle h\mid A (\cdot) \rangle = \left(^A \circ \Phi\right) h. Self-adjoint operators A continuous linear operator A : H \to H is called
self-adjoint In mathematics, and more specifically in abstract algebra, an element ''x'' of a *-algebra is self-adjoint if x^*=x. A self-adjoint element is also Hermitian, though the reverse doesn't necessarily hold. A collection ''C'' of elements of a st ...
it is equal to its own adjoint; that is, if A = A^*. Using (), this happens if and only if: \Phi \circ A = ^t A \circ \Phi where this equality can be rewritten in the following two equivalent forms: A = \Phi^ \circ ^t A \circ \Phi \quad \text \quad ^A = \Phi \circ A \circ \Phi^. Unraveling notation and definitions produces the following characterization of self-adjoint operators in terms of the aforementioned continuous linear functionals: A is self-adjoint if and only if for all z \in H, the linear functional \langle z\mid A (\cdot) \rangle is equal to the linear functional \langle A z\mid\cdot\, \rangle; that is, if and only if where if bra-ket notation is used, this is \langle z \mid A ~=~ \langle A z \mid \quad \text z \in H. Normal operators A continuous linear operator A : H \to H is called
normal Normal(s) or The Normal(s) may refer to: Film and television * ''Normal'' (2003 film), starring Jessica Lange and Tom Wilkinson * ''Normal'' (2007 film), starring Carrie-Anne Moss, Kevin Zegers, Callum Keith Rennie, and Andrew Airlie * ''Norma ...
if A A^* = A^* A, which happens if and only if for all z, h \in H, \left\langle A A^* z\mid h \right\rangle = \left\langle A^* A z\mid h \right\rangle. Using () and unraveling notation and definitions produces the following characterization of normal operators in terms of inner products of continuous linear functionals: A is a normal operator if and only if where the left hand side is also equal to \overline_H = \langle A z \mid A h \rangle_H. The left hand side of this characterization involves ''only'' linear functionals of the form \langle A h \mid\cdot\, \rangle while the right hand side involves ''only'' linear functions of the form \langle h \mid A(\cdot) \rangle (defined as above). So in plain English, characterization () says that an operator is ''normal'' when the inner product of any two linear functions of the first form is equal to the inner product of their second form (using the same vectors z, h \in H for both forms). In other words, if it happens to be the case (and when A is injective, it is) that the assignment of linear functionals \langle A h \mid\cdot\, \rangle ~\mapsto~ \langle h , A(\cdot) \rangle is well-defined (or alternatively, if \langle h , A(\cdot) \rangle ~\mapsto~ \langle A h \mid\cdot\, \rangle is well-defined) where h ranges over H, then A is a normal operator if and only if this assignment preserves the inner product on H^*. The fact that every self-adjoint bounded linear operator is normal follows readily by direct substitution of A^* = A into either side of A^* A = A A^*. This same fact also follows immediately from the direct substitution of the equalities () into either side of (). Alternatively, for a complex Hilbert space, the continuous linear operator A is a normal operator if and only if \, Az\, = \left\, A^* z\right\, for every z \in H, which happens if and only if \, Az\, _H = \, \langle z\, , \,A(\cdot) \rangle\, _ \quad \text z \in H. Unitary operators An invertible bounded linear operator A : H \to H is said to be
unitary Unitary may refer to: Mathematics * Unitary divisor * Unitary element * Unitary group * Unitary matrix * Unitary morphism * Unitary operator * Unitary transformation * Unitary representation * Unitarity (physics) * ''E''-unitary inverse semigrou ...
if its inverse is its adjoint: A^ = A^*. By using (), this is seen to be equivalent to \Phi \circ A^ = ^A \circ \Phi. Unraveling notation and definitions, it follows that A is unitary if and only if \langle A^ z\mid\cdot\, \rangle = \langle z\mid A (\cdot) \rangle \quad \text z \in H. The fact that a bounded invertible linear operator A : H \to H is unitary if and only if A^* A = \operatorname_H (or equivalently, ^t A \circ \Phi \circ A = \Phi) produces another (well-known) characterization: an invertible bounded linear map A is unitary if and only if \langle A z\mid A (\cdot)\, \rangle = \langle z\mid\cdot\, \rangle \quad \text z \in H. Because A : H \to H is invertible (and so in particular a bijection), this is also true of the transpose ^t A : H^* \to H^*. This fact also allows the vector z \in H in the above characterizations to be replaced with A z or A^ z, thereby producing many more equalities. Similarly, \,\cdot\, can be replaced with A(\cdot) or A^(\cdot).


See also

* * *


Citations


Notes

Proofs


Bibliography

* * * P. Halmos ''Measure Theory'', D. van Nostrand and Co., 1950. * P. Halmos, ''A Hilbert Space Problem Book'', Springer, New York 1982 ''(problem 3 contains version for vector spaces with coordinate systems)''. * * * * Walter Rudin, ''Real and Complex Analysis'', McGraw-Hill, 1966, . * {{Hilbert space Articles containing proofs Duality theories Integral representations Theorems in functional analysis