HOME

TheInfoList



OR:

In
mathematics Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
, a self-adjoint operator on a complex vector space ''V'' with inner product \langle\cdot,\cdot\rangle is a linear map ''A'' (from ''V'' to itself) that is its own adjoint. If ''V'' is
finite-dimensional In mathematics, the dimension of a vector space ''V'' is the cardinality (i.e., the number of vectors) of a basis of ''V'' over its base field. p. 44, §2.36 It is sometimes called Hamel dimension (after Georg Hamel) or algebraic dimension to disti ...
with a given orthonormal basis, this is equivalent to the condition that the
matrix Matrix most commonly refers to: * ''The Matrix'' (franchise), an American media franchise ** ''The Matrix'', a 1999 science-fiction action film ** "The Matrix", a fictional setting, a virtual reality environment, within ''The Matrix'' (franchis ...
of ''A'' is a Hermitian matrix, i.e., equal to its conjugate transpose ''A''. By the finite-dimensional spectral theorem, ''V'' has an orthonormal basis such that the matrix of ''A'' relative to this basis is a diagonal matrix with entries in the real numbers. This article deals with applying generalizations of this concept to operators on
Hilbert space In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise natural ...
s of arbitrary dimension. Self-adjoint operators are used in functional analysis and quantum mechanics. In quantum mechanics their importance lies in the Dirac–von Neumann formulation of quantum mechanics, in which physical observables such as
position Position often refers to: * Position (geometry), the spatial location (rather than orientation) of an entity * Position, a job or occupation Position may also refer to: Games and recreation * Position (poker), location relative to the dealer * ...
,
momentum In Newtonian mechanics, momentum (more specifically linear momentum or translational momentum) is the product of the mass and velocity of an object. It is a vector quantity, possessing a magnitude and a direction. If is an object's mass an ...
, angular momentum and
spin Spin or spinning most often refers to: * Spinning (textiles), the creation of yarn or thread by twisting fibers together, traditionally by hand spinning * Spin, the rotation of an object around a central axis * Spin (propaganda), an intentionally b ...
are represented by self-adjoint operators on a Hilbert space. Of particular significance is the Hamiltonian operator \hat defined by :\hat \psi = -\frac \nabla^2 \psi + V \psi, which as an observable corresponds to the total energy of a particle of mass ''m'' in a real potential field ''V''.
Differential operator In mathematics, a differential operator is an operator defined as a function of the differentiation operator. It is helpful, as a matter of notation first, to consider differentiation as an abstract operation that accepts a function and return ...
s are an important class of unbounded operators. The structure of self-adjoint operators on infinite-dimensional Hilbert spaces essentially resembles the finite-dimensional case. That is to say, operators are self-adjoint if and only if they are unitarily equivalent to real-valued
multiplication operator In operator theory, a multiplication operator is an operator defined on some vector space of functions and whose value at a function is given by multiplication by a fixed function . That is, T_f\varphi(x) = f(x) \varphi (x) \quad for all in th ...
s. With suitable modifications, this result can be extended to possibly unbounded operators on infinite-dimensional spaces. Since an everywhere-defined self-adjoint operator is necessarily bounded, one needs to be more attentive to the domain issue in the unbounded case. This is explained below in more detail.


Definitions

Let H be a
Hilbert space In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise natural ...
and A an unbounded (i.e. not necessarily bounded) operator with a dense
domain Domain may refer to: Mathematics *Domain of a function, the set of input values for which the (total) function is defined **Domain of definition of a partial function **Natural domain of a partial function **Domain of holomorphy of a function * Do ...
\operatornameA \subseteq H. This condition holds automatically when H is
finite-dimensional In mathematics, the dimension of a vector space ''V'' is the cardinality (i.e., the number of vectors) of a basis of ''V'' over its base field. p. 44, §2.36 It is sometimes called Hamel dimension (after Georg Hamel) or algebraic dimension to disti ...
since \operatornameA = H for every
linear operator In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that pre ...
on a finite-dimensional space. The graph of an (arbitrary) operator A is the set G(A) = \. An operator B is said to extend A if G(A) \subseteq G(B). This is written as A \subseteq B. Let the inner product \langle \cdot, \cdot\rangle be conjugate linear on the ''second'' argument. The adjoint operator A^* acts on the subspace \operatorname A^* \subseteq H consisting of the elements y such that : \langle Ax,y \rangle = \langle x,A^*y \rangle, \quad \forall x \in \operatorname A. The
densely defined In mathematics – specifically, in operator theory – a densely defined operator or partially defined operator is a type of partially defined function. In a topological sense, it is a linear operator that is defined "almost everywhere". ...
operator A is called symmetric (or Hermitian) if A \subseteq A^*, i.e., if \operatorname A \subseteq \operatorname A^* and Ax =A^*x for all x \in \operatorname A. Equivalently, A is symmetric if and only if : \langle Ax , y \rangle = \lang x , Ay \rangle, \quad \forall x,y\in \operatornameA. Since \operatorname A^* \supseteq \operatorname A is dense in H, symmetric operators are always closable (i.e. the closure of G(A) is the graph of an operator). If A^* is a closed extension of A, the smallest closed extension A^ of A must be contained in A^*. Hence, :A \subseteq A^ \subseteq A^* for symmetric operators and :A = A^ \subseteq A^* for closed symmetric operators. The densely defined operator A is called self-adjoint if A = A^*, that is, if and only if A is symmetric and \operatornameA = \operatornameA^*. Equivalently, a closed symmetric operator A is self-adjoint if and only if A^* is symmetric. If A is self-adjoint, then \left\langle x, A x \right\rangle is real for all x \in H, i.e., :\langle x, Ax\rangle = \overline=\overline \in \mathbb, \quad \forall x \in H. A symmetric operator A is said to be essentially self-adjoint if the closure of A is self-adjoint. Equivalently, A is essentially self-adjoint if it has a ''unique'' self-adjoint extension. In practical terms, having an essentially self-adjoint operator is almost as good as having a self-adjoint operator, since we merely need to take the closure to obtain a self-adjoint operator. In physics, the term Hermitian refers to symmetric as well as self-adjoint operators alike. The subtle difference between the two is generally overlooked.


Bounded self-adjoint operators

Let H be a
Hilbert space In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise natural ...
and A:\operatorname(A) \to H a symmetric operator. According to
Hellinger–Toeplitz theorem In functional analysis, a branch of mathematics, the Hellinger–Toeplitz theorem states that an everywhere-defined symmetric operator on a Hilbert space with inner product \langle \cdot , \cdot \rangle is bounded. By definition, an operator ' ...
, if \operatorname(A)=H then A is necessarily bounded. A bounded operator A : H \to H is self-adjoint if :\langle Ax, y\rangle = \langle x, Ay\rangle, \quad \forall x,y\in H. Every bounded operator T:H\to H can be written in the complex form T = A + i B where A:H\to H and B:H\to H are bounded self-adjoint operators. Alternatively, every positive bounded linear operator A:H \to H is self-adjoint if the Hilbert space H is ''complex''.


Properties

A bounded self-adjoint operator A : H \to H defined on \operatorname\left( A \right) = H has the following properties: * A : H \to \operatorname A \subseteq H is invertible if the
image An image is a visual representation of something. It can be two-dimensional, three-dimensional, or somehow otherwise feed into the visual system to convey information. An image can be an artifact, such as a photograph or other two-dimensiona ...
of A is dense in H. * \left\, A \right\, = \sup \left\ * The eigenvalues of A are real and the corresponding eigenvectors are orthogonal. * If \lambda is an eigenvalue of A then , \lambda , \leq \, A \, , where , \lambda , = \, A \, if \, x\, =1 and A is a compact self-adjoint operator.


Spectrum of self-adjoint operators

Let A:\operatorname(A) \to H be an unbounded operator. The resolvent set (or regular set) of A is defined as :\rho(A) = \left\. If A is bounded, the definition reduces to A - \lambda I being bijective on H. The spectrum of A is defined as the complement :\sigma(A) = \Complex \setminus \rho(A). In finite dimensions, \sigma(A)\subseteq \mathbb consists exclusively of (complex) eigenvalues. The spectrum of a self-adjoint operator is always real (i.e. \sigma(A)\subseteq \mathbb), though non-self-adjoint operators with real spectrum exist as well. For bounded ( normal) operators, however, the spectrum is real ''if and only if'' the operator is self-adjoint. This implies, for example, that a non-self-adjoint operator with real spectrum is necessarily unbounded. As a preliminary, define S=\, \textstyle m=\inf_ \langle Ax,x \rangle and \textstyle M=\sup_ \langle Ax,x \rangle with m,M \in \mathbb \cup \. Then, for every \lambda \in \Complex and every x \in \operatornameA, : \Vert (A - \lambda) x\Vert \geq d(\lambda)\cdot \Vert x\Vert, where \textstyle d(\lambda) = \inf_ , r - \lambda, . Indeed, let x \in \operatornameA \setminus \. By the
Cauchy–Schwarz inequality The Cauchy–Schwarz inequality (also called Cauchy–Bunyakovsky–Schwarz inequality) is considered one of the most important and widely used inequalities in mathematics. The inequality for sums was published by . The corresponding inequality fo ...
, : \Vert (A - \lambda) x\Vert \geq \frac =\left, \left\langle A\frac,\frac\right\rangle - \lambda\ \cdot \Vert x\Vert \geq d(\lambda)\cdot \Vert x\Vert. If \lambda \notin ,M then d(\lambda) > 0, and A - \lambda I is called ''bounded below''.


Spectral theorem

In the physics literature, the spectral theorem is often stated by saying that a self-adjoint operator has an orthonormal basis of eigenvectors. Physicists are well aware, however, of the phenomenon of "continuous spectrum"; thus, when they speak of an "orthonormal basis" they mean either an orthonormal basis in the classic sense ''or'' some continuous analog thereof. In the case of the
momentum operator In quantum mechanics, the momentum operator is the operator (physics), operator associated with the momentum (physics), linear momentum. The momentum operator is, in the position representation, an example of a differential operator. For the case o ...
P = -i\frac, for example, physicists would say that the eigenvectors are the functions f_p(x) := e^, which are clearly not in the Hilbert space L^2(\mathbb). (Physicists would say that the eigenvectors are "non-normalizable.") Physicists would then go on to say that these "generalized eigenvectors" form an "orthonormal basis in the continuous sense" for L^2(\mathbb), after replacing the usual Kronecker delta \delta_ by a
Dirac delta function In mathematics, the Dirac delta distribution ( distribution), also known as the unit impulse, is a generalized function or distribution over the real numbers, whose value is zero everywhere except at zero, and whose integral over the entire ...
\delta\left(p - p'\right). Although these statements may seem disconcerting to mathematicians, they can be made rigorous by use of the Fourier transform, which allows a general L^2 function to be expressed as a "superposition" (i.e., integral) of the functions e^, even though these functions are not in L^2. The Fourier transform "diagonalizes" the momentum operator; that is, it converts it into the operator of multiplication by p, where p is the variable of the Fourier transform. The spectral theorem in general can be expressed similarly as the possibility of "diagonalizing" an operator by showing it is unitarily equivalent to a multiplication operator. Other versions of the spectral theorem are similarly intended to capture the idea that a self-adjoint operator can have "eigenvectors" that are not actually in the Hilbert space in question.


Multiplication operator form of the spectral theorem

Firstly, let (X, \Sigma, \mu) be a σ-finite measure space and h : X \to \mathbb a
measurable function In mathematics and in particular measure theory, a measurable function is a function between the underlying sets of two measurable spaces that preserves the structure of the spaces: the preimage of any measurable set is measurable. This is in di ...
on X. Then the operator T_h : \operatornameT_h \to L^2(X,\mu), defined by :T_h \psi(x) = h(x)\psi(x), \quad \forall \psi \in \operatornameT_h, where :\operatornameT_h := \left\, is called a
multiplication operator In operator theory, a multiplication operator is an operator defined on some vector space of functions and whose value at a function is given by multiplication by a fixed function . That is, T_f\varphi(x) = f(x) \varphi (x) \quad for all in th ...
. Any multiplication operator is a self-adjoint operator. Secondly, two operators A and B with dense domains \operatornameA \subseteq H_1 and \operatornameB \subseteq H_2 in Hilbert spaces H_1 and H_2, respectively, are unitarily equivalent if and only if there is a unitary transformation U: H_1 \to H_2 such that: * U\operatornameA = \operatornameB, * U A U^ \xi = B \xi, \quad \forall \xi \in \operatornameB. If unitarily equivalent A and B are bounded, then \, A\, _=\, B\, _; if A is self-adjoint, then so is B. The spectral theorem holds for both bounded and unbounded self-adjoint operators. Proof of the latter follows by reduction to the spectral theorem for unitary operators. We might note that if T is multiplication by h, then the spectrum of T is just the
essential range In mathematics, particularly measure theory, the essential range, or the set of essential values, of a function is intuitively the 'non-negligible' range of the function: It does not change between two functions that are equal almost everywhere. O ...
of h. More complete versions of the spectral theorem exist as well that involve direct integrals and carry with it the notion of "generalized eigenvectors".


Functional calculus

One application of the spectral theorem is to define a functional calculus. That is, if f is a function on the real line and T is a self-adjoint operator, we wish to define the operator f(T). The spectral theorem shows that if T is represented as the operator of multiplication by h, then f(T) is the operator of multiplication by the composition f \circ h. One example from quantum mechanics is the case where T is the Hamiltonian operator \hat. If \hat has a true orthonormal basis of eigenvectors e_j with eigenvalues \lambda_j, then f(\hat) := e^ can be defined as the unique bounded operator with eigenvalues f(\lambda_j) := e^ such that: :f(\hat) e_j = f(\lambda_j)e_j. The goal of functional calculus is to extend this idea to the case where T has continuous spectrum (i.e. where T has no normalizable eigenvectors). It has been customary to introduce the following notation :\operatorname(\lambda) = \mathbf_ (T) where \mathbf_ is the
indicator function In mathematics, an indicator function or a characteristic function of a subset of a set is a function that maps elements of the subset to one, and all other elements to zero. That is, if is a subset of some set , one has \mathbf_(x)=1 if x\i ...
of the interval (-\infty, \lambda]. The family of projection operators E(λ) is called Borel functional calculus#Resolution of the identity, resolution of the identity for ''T''. Moreover, the following Stieltjes integral representation for ''T'' can be proved: : T = \int_^ \lambda d \operatorname(\lambda).


Formulation in the physics literature

In quantum mechanics, Dirac notation is used as combined expression for both the spectral theorem and the Borel functional calculus. That is, if ''H'' is self-adjoint and ''f'' is a Borel function, :f(H) = \int dE \left, \Psi_E \rangle f(E) \langle \Psi_E \ with :H \left, \Psi_E\right\rangle = E \left, \Psi_E\right\rangle where the integral runs over the whole spectrum of ''H''. The notation suggests that ''H'' is diagonalized by the eigenvectors Ψ''E''. Such a notation is purely
formal Formal, formality, informal or informality imply the complying with, or not complying with, some set of requirements (forms, in Ancient Greek). They may refer to: Dress code and events * Formal wear, attire for formal events * Semi-formal attire ...
. The resolution of the identity (sometimes called projection-valued measures) formally resembles the rank-1 projections \left, \Psi_E\right\rangle \left\langle\Psi_E\. In the Dirac notation, (projective) measurements are described via
eigenvalues In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted b ...
and
eigenstates In quantum physics, a quantum state is a mathematical entity that provides a probability distribution for the outcomes of each possible measurement on a system. Knowledge of the quantum state together with the rules for the system's evolution i ...
, both purely formal objects. As one would expect, this does not survive passage to the resolution of the identity. In the latter formulation, measurements are described using the
spectral measure In mathematics, the spectral theory of ordinary differential equations is the part of spectral theory concerned with the determination of the spectrum and eigenfunction expansion associated with a linear ordinary differential equation. In his disse ...
of , \Psi \rangle, if the system is prepared in , \Psi \rangle prior to the measurement. Alternatively, if one would like to preserve the notion of eigenstates and make it rigorous, rather than merely formal, one can replace the state space by a suitable rigged Hilbert space. If , the theorem is referred to as resolution of unity: :I = \int dE \left, \Psi_E\right\rangle \left\langle\Psi_E\ In the case H_\text = H - i\Gamma is the sum of an Hermitian ''H'' and a skew-Hermitian (see skew-Hermitian matrix) operator -i\Gamma, one defines the biorthogonal basis set :H^*_\text \left, \Psi_E^*\right\rangle = E^* \left, \Psi_E^*\right\rangle and write the spectral theorem as: :f\left(H_\text\right) = \int dE \left, \Psi_E\right\rangle f(E) \left\langle\Psi_E^*\ (See Feshbach–Fano partitioning method for the context where such operators appear in
scattering theory In mathematics and physics, scattering theory is a framework for studying and understanding the scattering of waves and particles. Wave scattering corresponds to the collision and scattering of a wave with some material object, for instance sunli ...
).


Formulation for symmetric operators

The spectral theorem applies only to self-adjoint operators, and not in general to symmetric operators. Nevertheless, we can at this point give a simple example of a symmetric (specifically, an essentially self-adjoint) operator that has an orthonormal basis of eigenvectors. Consider the complex Hilbert space ''L''2 ,1and the
differential operator In mathematics, a differential operator is an operator defined as a function of the differentiation operator. It is helpful, as a matter of notation first, to consider differentiation as an abstract operation that accepts a function and return ...
: A = -\frac with \mathrm(A) consisting of all complex-valued infinitely differentiable functions ''f'' on
, 1 The comma is a punctuation mark that appears in several variants in different languages. It has the same shape as an apostrophe or single closing quotation mark () in many typefaces, but it differs from them in being placed on the baseline (t ...
satisfying the boundary conditions :f(0) = f(1) = 0. Then integration by parts of the inner product shows that ''A'' is symmetric.The reader is invited to perform integration by parts twice and verify that the given boundary conditions for \operatorname(A) ensure that the boundary terms in the integration by parts vanish. The eigenfunctions of ''A'' are the sinusoids : f_n(x) = \sin(n \pi x) \qquad n= 1, 2, \ldots with the real eigenvalues ''n''2π2; the well-known orthogonality of the sine functions follows as a consequence of ''A'' being symmetric. The operator ''A'' can be seen to have a
compact Compact as used in politics may refer broadly to a pact or treaty; in more specific cases it may refer to: * Interstate compact * Blood compact, an ancient ritual of the Philippines * Compact government, a type of colonial rule utilized in British ...
inverse, meaning that the corresponding differential equation ''Af'' = ''g'' is solved by some integral (and therefore compact) operator ''G''. The compact symmetric operator ''G'' then has a countable family of eigenvectors which are complete in . The same can then be said for ''A''.


Pure point spectrum

A self-adjoint operator ''A'' on ''H'' has pure point spectrum if and only if ''H'' has an orthonormal basis ''i'' ∈ I consisting of eigenvectors for ''A''. Example. The Hamiltonian for the harmonic oscillator has a quadratic potential ''V'', that is :-\Delta + , x, ^2. This Hamiltonian has pure point spectrum; this is typical for bound state Hamiltonians in quantum mechanics. As was pointed out in a previous example, a sufficient condition that an unbounded symmetric operator has eigenvectors which form a Hilbert space basis is that it has a compact inverse.


Symmetric vs self-adjoint operators

Although the distinction between a symmetric operator and a (essentially) self-adjoint operator is subtle, it is important since self-adjointness is the hypothesis in the spectral theorem. Here we discuss some concrete examples of the distinction.


Boundary conditions

In the case where the Hilbert space is a space of functions on a bounded domain, these distinctions have to do with a familiar issue in quantum physics: One cannot define an operator—such as the momentum or Hamiltonian operator—on a bounded domain without specifying ''boundary conditions''. In mathematical terms, choosing the boundary conditions amounts to choosing an appropriate domain for the operator. Consider, for example, the Hilbert space L^2(
, 1 The comma is a punctuation mark that appears in several variants in different languages. It has the same shape as an apostrophe or single closing quotation mark () in many typefaces, but it differs from them in being placed on the baseline (t ...
(the space of square-integrable functions on the interval ,1. Let us define a momentum operator ''A'' on this space by the usual formula, setting Planck's constant equal to 1: : Af = -i\frac. We must now specify a domain for ''A'', which amounts to choosing boundary conditions. If we choose : \operatorname(A) = \left\, then ''A'' is not symmetric (because the boundary terms in the integration by parts do not vanish). If we choose : \operatorname(A) = \left\, then using integration by parts, one can easily verify that ''A'' is symmetric. This operator is not essentially self-adjoint, however, basically because we have specified too many boundary conditions on the domain of ''A'', which makes the domain of the adjoint too big (see also the
example Example may refer to: * '' exempli gratia'' (e.g.), usually read out in English as "for example" * .example, reserved as a domain name that may not be installed as a top-level domain of the Internet ** example.com, example.net, example.org, ex ...
below). Specifically, with the above choice of domain for ''A'', the domain of the closure A^ of ''A'' is :\operatorname\left(A^\right) = \left\, whereas the domain of the adjoint A^* of ''A'' is :\operatorname\left(A^*\right) = \left\. That is to say, the domain of the closure has the same boundary conditions as the domain of ''A'' itself, just a less stringent smoothness assumption. Meanwhile, since there are "too many" boundary conditions on ''A'', there are "too few" (actually, none at all in this case) for A^*. If we compute \langle g, Af\rangle for f \in \operatorname(A) using integration by parts, then since f vanishes at both ends of the interval, no boundary conditions on g are needed to cancel out the boundary terms in the integration by parts. Thus, any sufficiently smooth function g is in the domain of A^*, with A^*g = -i\,dg/dx. Since the domain of the closure and the domain of the adjoint do not agree, ''A'' is not essentially self-adjoint. After all, a general result says that the domain of the adjoint of A^\mathrm is the same as the domain of the adjoint of ''A''. Thus, in this case, the domain of the adjoint of A^\mathrm is bigger than the domain of A^\mathrm itself, showing that A^\mathrm is not self-adjoint, which by definition means that ''A'' is not essentially self-adjoint. The problem with the preceding example is that we imposed too many boundary conditions on the domain of ''A''. A better choice of domain would be to use periodic boundary conditions: :\operatorname(A) = \. With this domain, ''A'' is essentially self-adjoint. In this case, we can understand the implications of the domain issues for the spectral theorem. If we use the first choice of domain (with no boundary conditions), all functions f_\beta(x) = e^ for \beta \in \mathbb C are eigenvectors, with eigenvalues -i \beta, and so the spectrum is the whole complex plane. If we use the second choice of domain (with Dirichlet boundary conditions), ''A'' has no eigenvectors at all. If we use the third choice of domain (with periodic boundary conditions), we can find an orthonormal basis of eigenvectors for ''A'', the functions f_n(x) := e^. Thus, in this case finding a domain such that ''A'' is self-adjoint is a compromise: the domain has to be small enough so that ''A'' is symmetric, but large enough so that D(A^*)=D(A).


Schrödinger operators with singular potentials

A more subtle example of the distinction between symmetric and (essentially) self-adjoint operators comes from Schrödinger operators in quantum mechanics. If the potential energy is singular—particularly if the potential is unbounded below—the associated Schrödinger operator may fail to be essentially self-adjoint. In one dimension, for example, the operator :\hat := \frac - X^4 is not essentially self-adjoint on the space of smooth, rapidly decaying functions. In this case, the failure of essential self-adjointness reflects a pathology in the underlying classical system: A classical particle with a -x^4 potential escapes to infinity in finite time. This operator does not have a ''unique'' self-adjoint, but it does admit self-adjoint extensions obtained by specifying "boundary conditions at infinity". (Since \hat is a real operator, it commutes with complex conjugation. Thus, the deficiency indices are automatically equal, which is the condition for having a self-adjoint extension.) In this case, if we initially define \hat on the space of smooth, rapidly decaying functions, the adjoint will b