HOME

TheInfoList



OR:

In
quantum information theory Quantum information is the information of the state of a quantum system. It is the basic entity of study in quantum information theory, and can be manipulated using quantum information processing techniques. Quantum information refers to both t ...
, quantum relative entropy is a measure of distinguishability between two
quantum states In quantum physics, a quantum state is a mathematical entity that embodies the knowledge of a quantum system. Quantum mechanics specifies the construction, evolution, and measurement of a quantum state. The result is a prediction for the system re ...
. It is the quantum mechanical analog of
relative entropy Relative may refer to: General use *Kinship and family, the principle binding the most basic social units of society. If two people are connected by circumstances of birth, they are said to be ''relatives''. Philosophy *Relativism, the concept t ...
.


Motivation

For simplicity, it will be assumed that all objects in the article are finite-dimensional. We first discuss the classical case. Suppose the probabilities of a finite sequence of events is given by the probability distribution ''P'' = , but somehow we mistakenly assumed it to be ''Q'' = . For instance, we can mistake an unfair coin for a fair one. According to this erroneous assumption, our uncertainty about the ''j''-th event, or equivalently, the amount of information provided after observing the ''j''-th event, is :\; - \log q_j. The (assumed) average uncertainty of all possible events is then :\; - \sum_j p_j \log q_j. On the other hand, the
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
of the probability distribution ''p'', defined by :\; - \sum_j p_j \log p_j, is the real amount of uncertainty before observation. Therefore the difference between these two quantities :\; - \sum_j p_j \log q_j - \left(- \sum_j p_j \log p_j\right) = \sum_j p_j \log p_j - \sum_j p_j \log q_j is a measure of the distinguishability of the two probability distributions ''p'' and ''q''. This is precisely the classical relative entropy, or
Kullback–Leibler divergence In mathematical statistics, the Kullback–Leibler (KL) divergence (also called relative entropy and I-divergence), denoted D_\text(P \parallel Q), is a type of statistical distance: a measure of how much a model probability distribution is diff ...
: :D_(P\, Q) = \sum_j p_j \log \frac \!. Note #In the definitions above, the convention that 0·log 0 = 0 is assumed, since \lim_ x \log(x) = 0. Intuitively, one would expect that an event of zero probability to contribute nothing towards entropy. #The relative entropy is not a
metric Metric or metrical may refer to: Measuring * Metric system, an internationally adopted decimal system of measurement * An adjective indicating relation to measurement in general, or a noun describing a specific type of measurement Mathematics ...
. For example, it is not symmetric. The uncertainty discrepancy in mistaking a fair coin to be unfair is not the same as the opposite situation.


Definition

As with many other objects in quantum information theory, quantum relative entropy is defined by extending the classical definition from probability distributions to
density matrices In quantum mechanics, a density matrix (or density operator) is a matrix used in calculating the probabilities of the outcomes of measurements performed on physical systems. It is a generalization of the state vectors or wavefunctions: while th ...
. Let ''ρ'' be a density matrix. The
von Neumann entropy In physics, the von Neumann entropy, named after John von Neumann, is a measure of the statistical uncertainty within a description of a quantum system. It extends the concept of Gibbs entropy from classical statistical mechanics to quantum statis ...
of ''ρ'', which is the quantum mechanical analog of the Shannon entropy, is given by :S(\rho) = - \operatorname \rho \log \rho. For two density matrices ''ρ'' and ''σ'', the quantum relative entropy of ''ρ'' with respect to ''σ'' is defined by : S(\rho \, \sigma) = - \operatorname \rho \log \sigma - S(\rho) = \operatorname \rho \log \rho - \operatorname \rho \log \sigma = \operatorname\rho (\log \rho - \log \sigma). We see that, when the states are classically related, i.e. ''ρσ'' = ''σρ'', the definition coincides with the classical case, in the sense that if \rho = S D_1 S^ and \sigma = S D_2 S^ with D_1 = \text(\lambda_1, \ldots, \lambda_n) and D_2 = \text(\mu_1, \ldots, \mu_n) (because \rho and \sigma commute, they are
simultaneously diagonalizable In linear algebra, a square matrix A is called diagonalizable or non-defective if it is similar to a diagonal matrix. That is, if there exists an invertible matrix P and a diagonal matrix D such that . This is equivalent to (Such D are not ...
), then S(\rho \, \sigma) = \sum_^ \lambda_j \ln\left(\frac\right) is just the ordinary Kullback-Leibler divergence of the probability vector (\lambda_1, \ldots, \lambda_n) with respect to the probability vector (\mu_1, \ldots, \mu_n).


Non-finite (divergent) relative entropy

In general, the ''support'' of a matrix ''M'' is the orthogonal complement of its
kernel Kernel may refer to: Computing * Kernel (operating system), the central component of most operating systems * Kernel (image processing), a matrix used for image convolution * Compute kernel, in GPGPU programming * Kernel method, in machine learnin ...
, i.e. \text(M) = \text(M)^\perp . When considering the quantum relative entropy, we assume the convention that −''s'' · log 0 = ∞ for any ''s'' > 0. This leads to the definition that :S(\rho \, \sigma) = \infty when :\text(\rho) \cap \text(\sigma) \neq \. This can be interpreted in the following way. Informally, the quantum relative entropy is a measure of our ability to distinguish two quantum states where larger values indicate states that are more different. Being orthogonal represents the most different quantum states can be. This is reflected by non-finite quantum relative entropy for orthogonal quantum states. Following the argument given in the Motivation section, if we erroneously assume the state \rho has support in \text(\sigma), this is an error impossible to recover from. However, one should be careful not to conclude that the divergence of the quantum relative entropy S(\rho\, \sigma) implies that the states \rho and \sigma are orthogonal or even very different by other measures. Specifically, S(\rho\, \sigma) can diverge when \rho and \sigma differ by a ''vanishingly small amount'' as measured by some norm. For example, let \sigma have the diagonal representation \sigma=\sum_\lambda_n, f_n\rangle\langle f_n, with \lambda_n>0 for n=0,1,2,\ldots and \lambda_n=0 for n=-1,-2,\ldots where \ is an orthonormal set. The kernel of \sigma is the space spanned by the set \. Next let \rho=\sigma+\epsilon, f_\rangle\langle f_, - \epsilon, f_1\rangle\langle f_1, for a small positive number \epsilon. As \rho has support (namely the state , f_\rangle) in the kernel of \sigma, S(\rho\, \sigma) is divergent even though the trace norm of the difference (\rho-\sigma) is 2\epsilon . This means that difference between \rho and \sigma as measured by the trace norm is vanishingly small as \epsilon\to 0 even though S(\rho\, \sigma) is divergent (i.e. infinite). This property of the quantum relative entropy represents a serious shortcoming if not treated with care.


Non-negativity of relative entropy


Corresponding classical statement

For the classical
Kullback–Leibler divergence In mathematical statistics, the Kullback–Leibler (KL) divergence (also called relative entropy and I-divergence), denoted D_\text(P \parallel Q), is a type of statistical distance: a measure of how much a model probability distribution is diff ...
, it can be shown that :D_(P\, Q) = \sum_j p_j \log \frac \geq 0, and the equality holds if and only if ''P'' = ''Q''. Colloquially, this means that the uncertainty calculated using erroneous assumptions is always greater than the real amount of uncertainty. To show the inequality, we rewrite :D_(P\, Q) = \sum_j p_j \log \frac = \sum_j (- \log \frac)(p_j). Notice that log is a
concave function In mathematics, a concave function is one for which the function value at any convex combination of elements in the domain is greater than or equal to that convex combination of those domain elements. Equivalently, a concave function is any funct ...
. Therefore -log is
convex Convex or convexity may refer to: Science and technology * Convex lens, in optics Mathematics * Convex set, containing the whole line segment that joins points ** Convex polygon, a polygon which encloses a convex set of points ** Convex polytop ...
. Applying
Jensen's inequality In mathematics, Jensen's inequality, named after the Danish mathematician Johan Jensen, relates the value of a convex function of an integral to the integral of the convex function. It was proved by Jensen in 1906, building on an earlier p ...
, we obtain : D_(P\, Q) = \sum_j (- \log \frac)(p_j) \geq - \log ( \sum_j \frac p_j ) = 0. Jensen's inequality also states that equality holds if and only if, for all ''i'', ''qi'' = (Σ''qj'') ''pi'', i.e. ''p'' = ''q''.


The result

Klein's inequality states that the quantum relative entropy : S(\rho \, \sigma) = \operatorname\rho (\log \rho - \log \sigma). is non-negative in general. It is zero if and only if ''ρ'' = ''σ''. Proof Let ''ρ'' and ''σ'' have spectral decompositions :\rho = \sum_i p_i v_i v_i ^* \; , \; \sigma = \sum_i q_i w_i w_i ^*. So :\log \rho = \sum_i (\log p_i) v_i v_i ^* \; , \; \log \sigma = \sum_i (\log q_i)w_i w_i ^*. Direct calculation gives :S(\rho \, \sigma)= \sum_k p_k \log p_k - \sum_ (p_i \log q_j) , v_i ^* w_j , ^2 :\qquad \quad \; = \sum_i p_i ( \log p_i - \sum_j \log q_j , v_i ^* w_j , ^2) :\qquad \quad \; = \sum_i p_i (\log p_i - \sum_j (\log q_j )P_), where ''Pi j'' = , ''vi*wj'', 2. Since the matrix (''Pi j'')''i j'' is a
doubly stochastic matrix In mathematics, especially in probability and combinatorics, a doubly stochastic matrix (also called bistochastic matrix) is a square matrix X=(x_) of nonnegative real numbers, each of whose rows and columns sums to 1, i.e., :\sum_i x_=\sum_j x_ ...
and -log is a convex function, the above expression is :\geq \sum_i p_i (\log p_i - \log (\sum_j q_j P_)). Define ''r''i = Σ''j''''qj Pi j''. Then is a probability distribution. From the non-negativity of classical relative entropy, we have :S(\rho \, \sigma) \geq \sum_i p_i \log \frac \geq 0. The second part of the claim follows from the fact that, since -log is strictly convex, equality is achieved in : \sum_i p_i (\log p_i - \sum_j (\log q_j )P_) \geq \sum_i p_i (\log p_i - \log (\sum_j q_j P_)) if and only if (''Pi j'') is a
permutation matrix In mathematics, particularly in matrix theory, a permutation matrix is a square binary matrix that has exactly one entry of 1 in each row and each column with all other entries 0. An permutation matrix can represent a permutation of elements. ...
, which implies ''ρ'' = ''σ'', after a suitable labeling of the eigenvectors and .:513


Joint convexity of relative entropy

The relative entropy is jointly convex. For 0\leq \lambda\leq 1 and states \rho_, \sigma_ we have D(\lambda\rho_1+(1-\lambda)\rho_2\, \lambda\sigma_1+(1-\lambda)\sigma_2)\leq \lambda D(\rho_1\, \sigma_1)+(1-\lambda)D(\rho_2\, \sigma_2)


Monotonicity of relative entropy

The relative entropy decreases monotonically under completely positive
trace Trace may refer to: Arts and entertainment Music * ''Trace'' (Son Volt album), 1995 * ''Trace'' (Died Pretty album), 1993 * Trace (band), a Dutch progressive rock band * ''The Trace'' (album), by Nell Other uses in arts and entertainment * ...
preserving (CPTP) operations \mathcal on density matrices, S(\mathcal(\rho)\, \mathcal(\sigma))\leq S(\rho\, \sigma). This inequality is called monotonicity of quantum relative entropy and was first proved by Göran Lindblad.


An entanglement measure

Let a composite quantum system have state space :H = \otimes _k H_k and ''ρ'' be a density matrix acting on ''H''. The relative entropy of entanglement of ''ρ'' is defined by :\; D_ (\rho) = \min_ S(\rho \, \sigma) where the minimum is taken over the family of
separable state In quantum mechanics, separable states are multipartite quantum states that can be written as a convex combination of product states. Product states are multipartite quantum states that can be written as a tensor product of states in each space. ...
s. A physical interpretation of the quantity is the optimal distinguishability of the state ''ρ'' from separable states. Clearly, when ''ρ'' is not entangled :\; D_ (\rho) = 0 by Klein's inequality.


Relation to other quantum information quantities

One reason the quantum relative entropy is useful is that several other important quantum information quantities are special cases of it. Often, theorems are stated in terms of the quantum relative entropy, which lead to immediate corollaries concerning the other quantities. Below, we list some of these relations. Let ''ρ''AB be the joint state of a bipartite system with subsystem ''A'' of dimension ''n''A and ''B'' of dimension ''n''B. Let ''ρ''A, ''ρ''B be the respective reduced states, and ''I''A, ''I''B the respective identities. The maximally mixed states are ''I''A/''n''A and ''I''B/''n''B. Then it is possible to show with direct computation that :S(\rho_ , , I_/n_A) = \mathrm(n_A)- S(\rho_), \; :S(\rho_ , , \rho_ \otimes \rho_) = S(\rho_) + S(\rho_) - S(\rho_) = I(A:B), :S(\rho_ , , \rho_ \otimes I_/n_B) = \mathrm(n_B) + S(\rho_) - S(\rho_) = \mathrm(n_B)- S(B, A), where ''I''(''A'':''B'') is the quantum mutual information and ''S''(''B'', ''A'') is the
quantum conditional entropy The conditional quantum entropy is an entropy measure used in quantum information theory. It is a generalization of the conditional entropy of classical information theory. For a bipartite state \rho^, the conditional entropy is written S(A, B)_ ...
.


References

* {{cite journal , last=Vedral , first=V. , title=The role of relative entropy in quantum information theory , journal=Reviews of Modern Physics , publisher=American Physical Society (APS) , volume=74 , issue=1 , date=8 March 2002 , issn=0034-6861 , doi=10.1103/revmodphys.74.197 , pages=197–234, arxiv=quant-ph/0102094, bibcode=2002RvMP...74..197V , s2cid=6370982 * Marco Tomamichel,
Quantum Information Processing with Finite Resources -- Mathematical Foundations
. arXiv:1504.00233 Quantum mechanical entropy Quantum information theory