Next Article in Journal
Consideration on Singularities in Learning Theory and the Learning Coefficient
Next Article in Special Issue
Local Softening of Information Geometric Indicators of Chaos in Statistical Modeling in the Presence of Quantum-Like Considerations
Previous Article in Journal
Dynamic Distance Measure on Spaces of Isospectral Mixed Quantum States
Previous Article in Special Issue
Information Geometry of Complex Hamiltonians and Exceptional Points
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Correlation Distance and Bounds for Mutual Information

by
Michael J. W. Hall
Centre for Quantum Computation and Communication Technology (Australian Research Council), Centre for Quantum Dynamics, Griffith University, Brisbane, QLD 4111, Australia
Entropy 2013, 15(9), 3698-3713; https://doi.org/10.3390/e15093698
Submission received: 21 June 2013 / Revised: 21 August 2013 / Accepted: 3 September 2013 / Published: 6 September 2013
(This article belongs to the Special Issue Distance in Information and Statistical Physics Volume 2)

Abstract

:
The correlation distance quantifies the statistical independence of two classical or quantum systems, via the distance from their joint state to the product of the marginal states. Tight lower bounds are given for the mutual information between pairs of two-valued classical variables and quantum qubits, in terms of the corresponding classical and quantum correlation distances. These bounds are stronger than the Pinsker inequality (and refinements thereof) for relative entropy. The classical lower bound may be used to quantify properties of statistical models that violate Bell inequalities. Partially entangled qubits can have lower mutual information than can any two-valued classical variables having the same correlation distance. The qubit correlation distance also provides a direct entanglement criterion, related to the spin covariance matrix. Connections of results with classically-correlated quantum states are briefly discussed.
Classification:
PACS 89.70.Cf; 03.67.-a.

1. Introduction

The relative entropy between two probability distributions has many applications in classical and quantum information theory. A number of these applications, including the conditional limit theorem [1], quantum error correction [2], and secure random number generation and communication [3,4], make use of lower bounds on the relative entropy in terms of a suitable distance between the two distributions. The best known such bound is the so-called Pinsker inequality [5]
H ( P Q ) : = j P ( j ) [ log P ( j ) log Q ( j ) ] 1 2 D ( P , Q ) 2 log e
where D ( P , Q ) : = P Q 1 = j | P ( j ) Q ( j ) | is the variational or L1 distance between distributions P and Q. Note that the choice of logarithm base is left open throughout this paper, corresponding to a choice of units. There are a number of such bounds [5], all of which easily generalise to the case of quantum probabilities [2,6,7].
However, in a number of applications of the Pinsker inequality and its quantum analog, a lower bound is in fact only needed for the special case that the relative entropy quantifies the mutual information between two systems. Such applications include, for example, secure random number generation and coding [3,4] (both classical and quantum), and quantum de Finnetti theorems [8]. Since mutual information is a special case of relative entropy, it follows that it may be possible to find strictly stronger lower bounds for mutual information.
Surprisingly little attention appears to have been paid to this possibility of better lower bounds (although upper bounds for mutual information have been investigated [9]). The results of preliminary investigations are given here, with explicit tight lower bounds being obtained for pairs of two-valued classical random variables, and for pairs of quantum qubits with maximally-mixed reduced states.
In the context of mutual information, the corresponding variational distance reduces to the distance between the joint state of the systems and the product of their marginal states, referred to here as the “correlation distance”. It is shown that both the classical and quantum correlation distances are relevant for quantifying properties of quantum entanglement: the former with respect to the classical resources required to simulate entanglement, and the latter as providing a criterion for qubit entanglement. In the quantum case, it is also shown that the minimum value of the mutual information can only be achieved by entangled qubits if the correlation distance is more than ≈ 0 . 72654 .
The main results are given in the following section. Lower bounds on classical and quantum mutual information for two-level systems are derived in Section 3 and Section 5, and an entanglement criterion for qubits in terms of the quantum correlation distance is obtained in Section 4. Connections with classically-correlated quantum states are briefly discussed in Section 6, and conclusions are presented in Section 7.

2. Definitions and Main Results

For two classical random variables A and B, with joint probability distribution P A B ( a , b ) and marginal distributions P A ( a ) and P B ( b ) , the Shannon mutual information and the classical correlation distance are defined respectively by
I ( P A B ) : = H ( P A B P A P B ) = H ( P A ) + H ( P B ) H ( P A B )
C ( P A B ) : = P A B P A P B 1 = a , b P A B ( a , b ) P A ( a ) P B ( b )
where H ( P ) : = j P ( j ) log P ( j ) denotes the Shannon entropy of distribution P. The term “correlation distance” is used for C ( P A B ) , since it inherits all the properties of a distance from the more general variational distance, and clearly vanishes for uncorrelated A and B. Both the mutual information and correlation distance have a minimum value of zero, and for n-valued random variables have maximum values
I ( P A B ) log n , C ( P A B ) 2 ( n 1 ) / n
with saturation corresponding to the maximally correlated case P A B ( a , b ) = n 1 δ a b . Note for n = 2 that C ( P A B ) 1 .
For two quantum systems A and B described by density operator ρ A B and reduced density operators ρ A and ρ B , the corresponding quantum mutual information and quantum correlation distance are analogously defined by
I ( ρ A B ) : = S ( ρ A ) + S ( ρ B ) S ( ρ A B )
C ( ρ A B ) : = ρ A B ρ A ρ B 1 = tr | ρ A B ρ A ρ B |
where S ( ρ ) : = tr [ ρ log ρ ] denotes the von Neumann entropy of density operator ρ. Similarly to the classical case, these are direct measures of the correlation between A and B, vanishing only for uncorrelated A and B. It may be noted that trace distance has recently also been used to distinguish between and quantify quantum and classical contributions to this correlation [10,11]. For n-level quantum systems one has the maximum values
I ( ρ A B ) log n 2 , C ( ρ A B ) 2 ( n 2 1 ) / n 2
with saturation corresponding to maximally entangled states. Thus, comparing with Equation (4), quantum correlations have a quadratic advantage with respect to both mutual information and correlation distance. For example, for n = 2 one has C ( ρ A B ) 3 / 2 , allowing quantum correlation distances that are greater than the corresponding classical maximum value of unity.
In both the classical and quantum cases, one has the lower bound
I 1 2 C 2 log e
for mutual information, as a direct consequence of the Pinsker inequality (1) for classical relative entropies [2,5,6,7]. However, better bounds for mutual information can be obtained, which are stronger than any general inequality for relative entropy and variational distance.
For example, for two-valued classical random variables A and B one has the tight lower bound
I ( P A B ) log 2 H 1 + C ( P A B ) 2 , 1 C ( P A B ) 2
for classical mutual information. This inequality has been previously stated without proof in Reference [12], where it was used to bound the shared information required to classically simulate entangled quantum systems. It is proved in Section 3 below.
In contrast to Pinsker-type inequalities such as Equation (8), the quantum generalisation of Equation (9) is not straightforward. In particular, note for a two-qubit system that one cannot simply replace P A B by ρ A B in Equation (9), as the right hand side would be undefined for C ( ρ A B ) > 1 . This can occur if the qubits are entangled. Indeed, as shown in Section 4, C ( ρ A B ) > 1 is a sufficient condition for the entanglement of two qubits, as is the stronger condition
C ( ρ A B ) > 2 ( 1 tr [ ρ A 2 ] ) ( 1 tr [ ρ B 2 ] )
An explicit expression for the quantum correlation distance for two qubits, in terms of the spin covariance matrix, is also given in Section 4.
It is shown in Section 5 that the quantum equivalent of Equation (9), i.e., a tight lower bound for the quantum mutual information shared by two qubits, is
I ( ρ A B ) log 2 H 1 + C ( ρ A B ) 2 , 1 C ( ρ A B ) 2 , C ( ρ A B ) C 0 log 4 H 1 4 + C ( ρ A B ) 2 , 1 4 C ( ρ A B ) 6 , 1 4 C ( ρ A B ) 6 , 1 4 C ( ρ A B ) 6 , C ( ρ A B ) > C 0
when the reduced density operators are maximally mixed, where C 0 0 . 72654 corresponds to the value of C ( ρ A B ) for which the two expressions are equal. For C ( ρ A B ) > C 0 this lower bound can only be achieved by entangled states, and cannot be achieved by any classical distribution P A B having the same correlation distance. It is also shown that, for C ( ρ A B ) > C 0 , the bound is also tight if only one of the reduced states is maximally mixed. Support is given for the conjecture that the bound in Equation (11) in fact holds for all two-qubit states.
In Section 6 the natural role of “classically-correlated” quantum states, in comparing classical and quantum correlations, is briefly discussed. Such states have the general form ρ A B = j , k P ( j , k ) | j , k j , k | [13], where P ( j , k ) is a classical joint probability distribution and { | j } and { | k } are orthonormal basis sets for the two quantum systems. The lower bound in Equation (11) can be saturated by a classically-correlated state if and only if C C 0 .

3. Tight Lower Bound for Classical Mutual Information

3.1. Derivation of Bound

The tight lower bound in Equation (9) is derived here. The bound is plotted in Figure 1 below (top curve). Also plotted for comparison are the Pinsker lower bound in Equation (8) (bottom curve), and the lower bound following from the best possible generic inequality for relative entropy and variational distance, given in parametric form in Reference [5] (middle curve).
To derive the bound in Equation (9), it is convenient to label the two possible values of A and B by ± 1 . Defining R ( a , b ) : = 4 [ P A B ( a , b ) P A ( a ) P B ( b ) ] , it follows by summing over each of a and b separately that R ( a , b ) + R ( a , b ) = 0 = R ( a , b ) + R ( a , b ) , implying that R ( a , b ) = a b r with r : = R ( 1 , 1 ) . Hence, C ( P A B ) = | r | . Further, writing P A ( a ) = ( 1 + a x ) / 2 and P B ( b ) = ( 1 + b y ) / 2 , for suitable x , y [ 1 , 1 ] , the positivity condition P A B ( a , b ) 0 is equivalent to
| x + y | 1 r + x y 1 | x y |
Now, Equation (9) is equivalent to
f ( r ) : = I ( P A B ) log 2 + H 1 + r 2 , 1 r 2 0
It is easy to check that this inequality is always saturated for the case of maximally-random marginals, i.e., when x = y = 0 . In all other cases, the inequality may be proved by showing that f ( r ) has a unique global minimum value of 0 at r = 0 .
Figure 1. Lower bounds for the classical mutual information between two-valued variables.
Figure 1. Lower bounds for the classical mutual information between two-valued variables.
Entropy 15 03698 g001
In particular, note first that f ( 0 ) = 0 (one has P A B = P A P B in this case, so that the mutual information vanishes). Further, using P A B ( a , b ) = [ ( 1 + a x ) ( 1 + b y ) + a b r ] / 4 , one easily calculates that, using logarithm base e for convenience,
f ( r ) = 1 4 a , b a b log P A B ( a , b ) 1 2 a a log 1 + a r 2 = 1 4 log p A B ( + , + ) p A B ( , ) ( 1 r ) 2 p A B ( + , ) p A B ( , + ) ( 1 + r ) 2
Hence, f ( r ) = 0 if and only if the argument of the logarithm is unity, i.e., if and only if
[ ( 1 + x ) ( 1 + y ) + r ] [ ( 1 x ) ( 1 y ) + r ] ( 1 r ) 2 = [ ( 1 + x ) ( 1 y ) r ] [ ( 1 x ) ( 1 + y ) r ] ( 1 + r ) 2 .
Expanding and simplifying yields two possible solutions: r = 0 , or r = ( x 2 + y 2 x 2 y 2 ) / ( 2 x y ) . However, in the latter case one has
| r + x y | = x 2 + y 2 + x 2 y 2 2 | x y | = α γ + γ 2 1 + γ 2 1
where α and γ denote the arithmetic mean and geometric mean, respectively, of x 2 and y 2 (hence α γ ). This is clearly inconsistent with the positivity condition (12) (unless x = y = 0 , which trivially saturates Equation (13) for all r as noted above). The only remaining solution to f ( r ) = 0 is then r = 0 , implying f ( r ) has a unique maximum or minimum value at r = 0 . Finally, it is easily checked that it is a minimum, since
f ( 0 ) = 1 16 a , b 1 p A ( a ) p B ( b ) 1 = 1 16 P A ( + ) P A ( ) P B ( + ) P B ( ) 1 = 1 ( 1 x 2 ) ( 1 y 2 ) 1 0
(with equality only for the trivially-saturating case x = y = 0 ). Thus, f ( r ) f ( 0 ) = 0 as required.

3.2. Application: Resources for Simulating Bell Inequality Violation

The hallmark feature of quantum correlations is that they cannot be explained by any underlying statistical model that satisfies three physically very plausible properties: (i) no signaling faster than the speed of light; (ii) free choice of measurement settings; and (iii) independence of local outcomes. Various interpretations of quantum mechanics differ in regard to which of these properties should be given up. It is of interest to consider by how much they must be given up, in terms of the information-theoretic resources required to simulate a given quantum correlation [14]. For example, how many bits of communication, or bits of correlation between the source and the measurement settings, or bits of correlation between the outcomes, are required? The lower bound for classical mutual information in Equation (9) is relevant for the last of these questions.
In more detail, if P A B ( a , b ) denotes the joint probability of outcomes a and b, for measurements of variables A and B on respective spacelike-separated systems, and λ denotes any underlying variables relevant to the correlations, then Bayes theorem implies that
P A B ( a , b ) = λ p A B ( λ ) P A B ( a , b | λ )
where summation is replaced by integration over any continuous values of λ. The no-signaling property requires that the underlying marginal distribution of A, p A ( a | λ ) , is independent of whether B or B was measured on the second system (and vice versa), while the free-choice property requires that λ is independent of the choice of the measured variables A and B, i.e., that p A B ( λ ) = p A B ( λ ) for any A , A , B , B . Finally, the outcome independence property requires that any observed correlation between A and B arises from ignorance of the underlying variable, i.e., that P A B ( a , b | λ ) = P A ( a | λ ) P B ( b | λ ) for all A, B and λ. Thus the correlation distance of P A B ( a , b | λ ) vanishes identically:
C ( P A B | λ ) 0
As is well known, the assumption of all three properties implies that two-valued random variables with values ± 1 must satisfy the Bell inequality [15]
A B + A B + A B A B 2
whereas quantum correlations can violate this inequality by as much as a factor of 2 [16]. It follows that quantum correlations can only be modeled by relaxing one or more of the above properties, as has recently been reviewed in detail in Reference [12].
For example, assuming that no-signaling and measurement independence hold (as they do in the standard Copenhagen interpretation of quantum mechanics), and defining C max to be the maximum value of C ( P A B | λ ) over all A, B and λ, it can be shown that Equation (20) generalises to the tight bound [12]
A B + A B + A B A B 4 2 C max
It follows that to simulate a Bell inequality violation A B + A B + A B A B = 2 + V , for some V > 0 , the observers must share random variables having a correlation distance of at least C max 2 V / ( 2 + V ) . Hence, using the classical lower bound Equation (9) (stated without proof in Reference [12]), the observers must share a minimum mutual information of
I min = log 2 H 1 + C max 2 , 1 C max 2 log 2 H 2 + 3 V 4 + 2 V , 2 V 4 + 2 V
Note this reduces to zero in the limit of no violation of Bell inequality (20), i.e., when V = 0 , and reaches a maximum of 1 bit of information in the limit of the maximum possible violation over arbitrary probability distributions, V = 2 . To simulate the maximum quantum violation [16], i.e., V = 2 2 2 , at least 0.264 bits are required.

4. Quantum Correlation Distance and Qubit Entanglement

The positivity condition (12) may be used to show that the classical correlation distance between any pair of two-valued random variables is never greater than unity, i.e., that C ( P A B ) = | r | 1 [12]. In contrast, the quantum correlation distance between a pair of qubits can be greater than unity, with upper bound C ( ρ A B ) 3 / 2 , as per Equation (7).
Non-classical values of the quantum correlation distance are closely related to the quintessential non-classical feature of quantum mechanics: entanglement. In particular, C ( ρ A B ) > 1 is a direct signature of qubit entanglement. Indeed, even correlation distances smaller than unity can imply two qubits are entangled, as per the criterion given in Equation (10) and shown below. An explicit formula for qubit correlation distance in terms of the spin covariance matrix, needed for Section 5, is also obtained below.

4.1. Entanglement Criterion

Recall that the density operator ρ A B of two qubits may always be written in the Fano form [17]
ρ A B = 1 4 I I + u · σ I + I v · σ + j , k σ j σ k σ j σ k = ρ A ρ B + 1 4 j , k T j k σ j σ k
Here I is the unit operator; { σ j } denotes the set of Pauli spin observables on each qubit Hilbert space; the components of the 3-vectors u and v are the spin expectation values u j : = σ j 1 and v : = 1 σ k , for A and B respectively; and T denotes the 3 × 3 spin covariance matrix with coefficients
T j k : = σ j σ k σ j I I σ k
It immediately follows from Equation (23) that the quantum correlation distance may be expressed in terms of the spin covariance matrix as
C ( ρ A B ) = 1 4 tr j , k T j k σ j σ k
This expression will be further simplified in Section 4.2.
Now consider the case where ρ A B is a separable state, i.e., of the unentangled form
ρ A B = λ p ( λ ) τ A ( λ ) ω B ( λ )
for some probability distribution p ( λ ) and local density operators { τ A ( λ ) } , { ω B ( λ ) } . Defining u j ( λ ) : = tr [ τ A ( λ ) σ j ] , v k ( λ ) : = tr [ ω B ( λ ) σ k ] implies u = λ p ( λ ) u ( λ ) and v = λ p ( λ ) v ( λ ) , and substitution into Equation (25) then yields
C ( ρ A B ) = 1 4 λ p ( λ ) [ u ( λ ) u ] · σ [ v ( λ ) v ] · σ 1 1 4 λ p ( λ ) [ u ( λ ) u ] · σ 1 [ v ( λ ) v ] · σ 1 = λ p ( λ ) | u ( λ ) u | | v ( λ ) v | λ p ( λ ) | u ( λ ) u | 2 1 / 2 λ p ( λ ) | v ( λ ) v | 2 1 / 2 = λ p ( λ ) | u ( λ ) | 2 | u | 2 1 / 2 λ p ( λ ) | v ( λ ) | 2 | v | 2 1 / 2 ( 1 u · u ) ( 1 v · v )
Note that second line follows from the properties X + Y 1 X 1 + Y 1 and X Y 1 X 1 Y 1 of the trace norm [18]; the third line using X 1 = tr [ X X ] and the Schwarz inequality; and the last line via | u ( λ ) | , | v ( λ ) | 1 .
Equation (27) holds for all separable qubit states. Hence, a non-classical value of the correlation distance, C ( ρ A B ) > 1 , immediately implies that the qubits must be entangled. More generally, noting that ρ A = 1 2 ( I + u · σ ) and ρ B = 1 2 ( I + v · σ ) , one has tr [ ρ A 2 ] = ( 1 + u · u ) / 2 , tr [ ρ B 2 ] = ( 1 + v · v ) / 2 , and the stronger entanglement criterion (10) immediately follows from Equation (27).
The fact that entanglement between two qubits is necessary (but not sufficient) for C ( ρ A B ) to be greater than the maximum possible value of C ( P A B ) , for two-valued classical variables, is a nice distinction between quantum and classical correlation distances. It would be of interest to determine whether this result generalises to n-level systems. This would follow from the validity of Equation (10) for arbitrary quantum systems.

4.2. Explicit Expression for C ( ρ A B )

To explicitly evaluate C ( ρ A B ) in Equation (25), let T = K D L T denote a singular value decomposition of the spin covariance matrix. Thus, K and L are real orthogonal matrices and D = diag [ t 1 , t 2 , t 3 ] , with the singular values t 1 t 2 t 3 0 corresponding to the square roots of the eigenvalues of T T T . Noting that any 3 × 3 orthogonal matrix is either a rotation matrix, or the product of a rotation matrix with the parity matrix I , one therefore always has a decomposition of the form T = ± K D L T where K and L are now restricted to be rotation matrices. Hence, defining unitary operators U and V corresponding to rotations K and L, via U σ j U = j , j K j j σ j and V σ j V = j , j L j j σ j , and using the invariance of the trace norm under unitary transformations, the quantum correlation distance in Equation (25) can be rewritten as
C ( ρ A B ) = 1 4 tr ± j t j U σ j U V σ j V = 1 4 tr j t j σ j σ j
Determining the eigenvalues of the Hermitian operator j t j σ j σ j is a straightforward 4 × 4 matrix calculation using the standard representation of the Pauli sigma matrices. Summing the absolute values of these eigenvalues then yields the explicit expression
C ( ρ A B ) = 1 4 | t 1 + t 2 + t 3 | + | t 1 + t 2 t 3 | + | t 1 t 2 + t 3 | + | t 1 + t 2 + t 3 | = 1 2 max { t 1 + t 2 + t 3 , 2 t 1 }
for the quantum correlation distance, in terms of the singular values of the spin covariance matrix.
For example, for the Werner state ρ A B = p | ψ ψ | + ( 1 p ) / 4 I I , where | ψ is the singlet state and 1 / 3 p 1 [19], one has T = p I and hence that t 1 = t 2 = t 3 = | p | . The corresponding correlation distance is therefore 3 | p | / 2 , which is greater than the classical maximum of unity for p > 2 / 3 .
Equation (29) also allows the qubit entanglement criterion (10) to be directly compared with the strongest known criterion based on the spin covariance matrix [20]:
t 1 + t 2 + t 3 > 2 ( 1 tr [ ρ A 2 ] ) ( 1 tr [ ρ B 2 ] )
For the above Werner state this criterion is tight, indicating entanglement for p > 1 / 3 . Hence, the main interest in weaker entanglement criteria based on quantum correlation distance lies in their direct connection with non-classical values of the classical correlation distance.

5. Tight Lower Bound for Quantum Mutual Information

Here Equation (11) is derived for the case ρ A = ρ B = 1 2 I . Evidence is provided for the conjecture that Equation (11) in fact holds for all two-qubit states, including a partial generalisation of Equation (11) when only one of ρ A and ρ B is maximally-mixed.

5.1. Derivation for Maximally-Mixed ρ A and ρ B

The tight lower bound for quantum mutual information in Equation (11), for maximally-mixed reduced states, is plotted in Figure 2 below (top solid curve). Also plotted for comparison are the Pinsker lower bound in Equation (8) (bottom solid curve), and classical lower bound in Equation (9) (dashed curve). The dotted vertical line indicates the value of C 0 0 . 72654 in Equation (11). It is seen that quantum correlations can violate the classical lower bound for correlation distances falling between C 0 and 1.
To derive Equation (11) for ρ A = ρ B = 1 2 I , note first that Equation (23) reduces to ρ A B = 1 4 [ I I + j , k T j k σ j σ k ] . By the same argument given in Section 4.2, this can be transformed via local unitary transformations to the state
ρ ˜ A B = 1 4 I I + j r j σ j σ j
where r j = α t j , α = ± 1 , and t 1 t 2 t 3 0 are the singular values of the spin covariance matrix T. Since the quantum mutual information and quantum correlation distance are invariant under local unitary transformations, one has I ( ρ A B ) = I ( ρ ˜ A B ) and C ( ρ A B ) = C ( ρ ˜ A B ) . Hence Equation (11) only needs to be demonstrated for ρ ˜ A B .
Figure 2. Lower bounds for the quantum mutual information between two qubits.
Figure 2. Lower bounds for the quantum mutual information between two qubits.
Entropy 15 03698 g002
The mutual information of ρ ˜ A B is easily evaluated as
I ( ρ ˜ A B ) = S ( ρ ˜ A ) + S ( ρ ˜ B ) S ( ρ ˜ A B ) = log 4 H ( p 0 , p 1 , p 2 , p 3 )
where p 0 = 1 4 ( 1 r 1 r 2 r 3 ) , p 1 = 1 4 ( 1 r 1 + r 2 + r 3 ) , p 2 = 1 4 ( 1 + r 1 r 2 + r 3 ) , p 3 = 1 4 ( 1 + r 1 + r 2 r 3 ) are the eigenvalues of ρ ˜ A B . Inverting the relation between the r j and p j further yields
t j = α r j = α [ 1 2 ( p 0 + p j ) ] , t 1 + t 2 + t 3 = α ( 1 4 p 0 )
and hence the correlation distance follows from Equation (29) as
C ( ρ ˜ A B ) = C : = 1 2 max { α ( 1 4 p 0 ) , α ( 1 4 p 0 + 1 4 p 1 ) }
Equation (32) implies that a tight lower bound for I ( ρ ˜ A B ) corresponds to a tight upper bound for H ( p 0 , p 1 , p 2 , p 3 ) . To determine the maximum value of H ( p 0 , p 1 , p 2 , p 3 ) , for a fixed correlation distance C, consider first the case α = 1 . The ordering and positivity conditions on t j then require
p 1 p 2 p 3 , p 0 + p j 1 2 ( j = 1 , 2 , 3 )
(implying p 0 1 / 4 ). Further, from Equation (34), C = 1 2 max { 1 4 p 0 , 1 4 p 0 + 1 4 p 1 } . Hence, if p 1 1 / 4 , then C = 1 2 ( p 0 + p 1 ) 1 , implying the constraint p 0 + p 1 = ( 1 C ) / 2 . Noting the concavity of entropy, the maximum possible entropy under this constraint corresponds to equal values
p 0 = p 1 = ( 1 C ) / 4 , p 2 = p 3 = ( 1 + C ) / 4
(which are compatible with the above conditions on the p j ). Conversely, if p 1 1 / 4 then C = ( 1 4 p 0 ) / 2 1 / 2 , and hence p 0 = 1 / 4 C / 2 is fixed, implying by concavity that the maximum possible entropy corresponds to
p 0 = 1 / 4 C / 2 , p 1 = p 2 = p 3 = 1 / 4 + C / 6
(which again satisfies the required conditions on the p j ). It follows that the maximum possible entropy is (i) the maximum of the entropies
H 1 ( C ) : = H ( ( 1 C ) / 4 , ( 1 C ) / 4 , ( 1 + C ) / 4 , ( 1 + C ) / 4 )
H 2 ( C ) : = H ( 1 / 4 C / 2 , 1 / 4 + C / 6 , 1 / 4 C / 6 , 1 / 4 C / 6 )
for C 1 / 2 , and (ii) H 1 ( C ) for 1 / 2 < C 1 . However, it is straightforward to show that H 1 ( C ) H 2 ( C ) over their overlapping range. Hence the maximum possible entropy is always H 1 ( C ) for the case α = 1 .
For the case α = 1 , the conditions on t j require that
p 1 p 2 p 3 , p 0 + p j 1 2 ( j = 1 , 2 , 3 )
(implying p 0 1 / 4 ), while from Equation (34) C = 1 2 max { 4 p 0 1 , 4 p 0 1 + 4 p 1 1 ) } . Carrying out a similar analysis to the above, one finds that the maximum possible entropy is (i) the maximum of the entropies H 1 ( C ) and
H 3 ( C ) : = H ( 1 / 4 + Q / 2 , 1 / 4 Q / 6 , 1 / 4 Q / 6 , 1 / 4 Q / 6 )
for C 1 , and (ii) H 3 ( C ) for 1 < C 3 / 2 .
Numerical comparison shows that H 3 ( C ) > H 1 ( C ) for C > C 0 0 . 72654 , and H 3 ( C ) H 1 ( C ) otherwise. Hence, from Equation (32) one has the tight lower bound
I ( ρ ˜ A B ) log 4 H 1 ( C ) , C C 0 log 4 H 3 ( C ) , C > C 0
Since H 1 ( C ) = log 2 + H ( ( 1 C ) / 2 , ( 1 + C ) / 2 ) , it follows that Equation (11) holds for ρ ˜ A B in Equation (31), and hence for all qubit states with maximally-mixed reduced density operators, as claimed.
The states saturating the lower bound in Equations (11) and (42) are easily constructed from the above derivation. In particular, they are given by
ρ ( C ) : = 1 4 I I + C σ 1 σ 1 , C C 0 1 4 I I ( 2 C / 3 ) j σ j σ j , C > C 0
and any local unitary transformations thereof, where the quantum correlation distance of ρ ( C ) is C by construction.
Note that ρ ( C ) is unentangled for C C 0 (it can be written as a mixture of ( 1 / 4 ) I I , | + + | | + + | and | | | | , where σ 1 | ± = ± | ± ). Conversely, ρ ( C ) is an entangled Werner state for C C 0 (with singlet state weighting p = 2 C / 3 > 1 / 3 ). Hence, the lower bound in Equations (11) and (42) can only be achieved by entangled states for C C 0 , and cannot be achieved by any two-valued classical random variables.

5.2. Conjecture

It is conjectured that Equation (11) is in fact a tight lower bound for any two-qubit state. This conjecture would follow immediately if it could be shown that
I ( ρ A B ) I ( ρ A B )
for arbitrary ρ A B , where ρ A B : = ρ A B ρ A ρ B + ( 1 / 4 ) I I , and where it must further be shown that ρ A B is a density operator. The conjecture would then follow since ρ A B is of the form of ρ ˜ A B in Equation (31), and hence I ( ρ A B ) satisfies Equation (42).
Partial support for Equation (44), and hence for the conjecture, is given by noting that any ρ A B and corresponding ρ A B can be brought to the respective forms
ρ A B = ρ A ρ B + 1 4 j r j σ j σ j , ρ A B = 1 4 I I + j r j σ j σ j
via suitable local unitary transformations, similarly to the argument in Section 4.2. Defining the function
F ( r 1 , r 2 , r 3 ) : = I ( ρ A B ) I ( ρ A B )
it is straightforward to show that F = 0 and F / r j = 0 for r 1 = r 2 = r 3 = 0 , consistent with F 0 . However, it remains to be shown that the gradient F / r j = 0 does not vanish for other physically possible values of r j (other than for the trivially saturating case ρ A = ρ B = ( 1 / 2 ) I ).
The above conjecture is further supported by the generalisation of Equation (11) in the following section.

5.3. Generalisation to Maximally-Mixed ρ A or ρ B

It is straightforward to show that the lower bound on quantum mutual information is tight for C C 0 when just one of the mixed density operators is mixed, i.e., if ρ A or ρ B is equal to ( 1 / 2 ) I .
First, since ( 1 / 2 ) I is invariant under unitary transformations, the same argument as in Section 4.2 implies the state can always be transformed by local unitary transformations to the generalised form
ρ ˜ A B = 1 4 ρ ˜ A ρ ˜ B + α j t j σ j σ j
of Equation (31), where either ρ ˜ A or ρ ˜ B equals ( 1 / 2 ) I and α = ± 1 .
Second, let T denote the “twirling” operation, corresponding to applying a random unitary transformation of the form U U [21]. It is easy to check that by definition T ( I I ) = I I , T ( I σ j ) = 0 = T ( σ j 1 ) and T ( σ j σ j ) = T ( σ k σ k ) , for any j and k. Since Werner states are invariant under twirling [19,21], it follows that T ( σ j σ j ) = ( 1 / 3 ) k σ k σ k . Using these properties, one finds that T ( ρ ˜ A ρ ˜ B ) = ( 1 / 4 ) I I if one of ρ ˜ A or ρ ˜ B is maximally mixed, and hence that
T ( ρ ˜ A B ) = 1 4 I I + α t ¯ j σ j σ j = ρ ( 3 α t ¯ / 2 )
where t ¯ : = ( t 1 + t 2 + t 3 ) / 3 and the second equality holds for C C 0 (but not otherwise), with ρ ( C ) defined as per Equation (43). Further, from Equation (29) one has
C ( T ( ρ ˜ A B ) ) = C = 1 2 max { 2 t ¯ , 3 t ¯ } = 3 t ¯ / 2
Recalling that ρ ( C ) saturates Equation (42), an analysis similar to Section 5.1 shows for C C 0 that
I ( T ( ρ ˜ A B ) ) = log 4 H 3 ( α C ) log 4 H 3 ( C )
with equality for α = 1 .
Third, again using T ( ρ ˜ A ρ ˜ B ) = ( 1 / 4 ) I I , and the property that the relative entropy is non-increasing under the twirling operation, it follows that
I ( ρ ˜ A B ) = S ( ρ ˜ A B ρ ˜ A ρ ˜ B ) S ( T ( ρ ˜ A B ) T ( ρ ˜ A ρ ˜ B ) ) = I ( T ( ρ ˜ A B ) ) log 4 H 3 ( C )
for C C 0 . Since Werner states are invariant under twirling, this inequality is tight for α = 1 , being saturated by the choice ρ ˜ A B = ρ ( C ) . Recalling that mutual information and correlation distance are invariant under local unitary operations, the inequality is therefore tight for any ρ A B for which one of ρ A and ρ B is maximally mixed, as claimed.

6. Classically-Correlated Quantum States

It is well known that a quantum system behaves classically if the state and the observables of interest all commute, i.e., if they can be simultaneously diagonalised in some basis. Hence, a joint state will behave classically if the relevant observables of each system commute with each other and the state. It is therefore natural to define ρ A B to be classically correlated if and only if it can be diagonalised in a joint basis [13], i.e., if and only if
ρ A B = j , k P ( j , k ) | j j | | k k |
for some distribution P ( j , k ) and orthonormal basis set { | j | k } . Classical correlation is preserved by tensor products, and by mixtures of commuting states.
While, strictly speaking, a classically-correlated quantum state only behaves classically with respect to observables that are diagonal with respect to | j | k , they also have a number of classical correlation properties with respect to general observables [13,22], briefly noted here.
First, ρ A B above is separable by construction, and hence is unentangled. Second, since it is diagonal in the basis { | j | k } , the mutual information and correlation distance are easily calculated as
I ( ρ A B ) = I ( P ) , C ( ρ A B ) = C ( P )
and hence can only take classical values.
Third, if M and N denote any observables for systems A and B respectively, then their joint statistics are given by
P M N ( m , n ) = j , k p ( m | j ) p ( n | k ) P ( j , k ) = j , k S m , n ; j , k P ( j , k )
where S m , n ; j , k = p ( m | j ) p ( n | k ) is a stochastic matrix with respect to its first and second pairs of indices. Similarly, one finds
P M ( m ) P N ( n ) = j , k S m , n ; j , k P ( j ) P ( k )
for the product of the marginals. Since the classical relative entropy and variational distance can only decrease under the action of a stochastic matrix, it follows that one has the tight inequalities [13,22]
I ( P M N ) I ( P ) = I ( ρ A B ) , C ( P M N ) C ( P ) = C ( ρ A B )
with saturation for M and N diagonal in the bases { | j } and { | k } respectively. Maximising the first of these equalities over M or N immediately implies the well-known result that classically-correlated states have zero quantum discord [22].
Finally, for two-qubit systems, Equation (52) implies that ρ A B is classically correlated if and only if it is equivalent under local unitary transformations to a state of the form
ρ A B = 1 4 ( 1 + x σ 1 ) ( 1 + y σ 1 ) + r σ 1 σ 1
where x , y [ 1 , 1 ] and r satisfies Equation (12). Hence, the mutual information is bounded by the classical lower bound in Equation (9), and ρ ( C ) in Equation (43) is classically correlated for C C 0 . It follows that the lower bound for quantum mutual information in Equation (11) can be attained by classically-correlated states if C C 0 . Conversely, the minimum possible bound cannot be reached by any classically-correlated two-qubit state if C > C 0 .

7. Conclusions

Lower bounds for mutual information have been obtained that are stronger than those obtainable from general bounds for relative entropy and variational distance. Unlike the Pinsker inequality in Equation (8), the quantum form of these bounds is not a simple generalisation of the classical form.
Similarly to the case of upper bounds for (classical) mutual information [9], the tight lower bounds obtained here depend on the dimension of the systems. The results of this paper represent a preliminary investigation largely confined to two-valued classical variables and qubits. It would be of interest to generalise both the classical and quantum cases, and to further investigate connections between them.
Open questions include whether a quantum correlation distance greater than the corresponding maximum classical correlation distance is a signature of entanglement for higher-dimensional systems, and whether the related qubit entanglement criterion in Equation (10) holds more generally. The conjecture in Section 5.2, as to whether the quantum lower bound in Equation (11) is valid for all two-qubit states, also remains to be settled. Finally, it would be of interest to generalise and to better understand the role of the transition from classically-correlated states to entangled states in saturating information bounds, in the light of Equation (43) for qubits.

Acknowledgements

This research was supported by the ARC Centre of Excellence CE110001027.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2006; Chapter 11. [Google Scholar]
  2. Schumacher, B.; Westmoreland, M.D. Approximate quantum error correction. Quantum Inf. Process. 2002, 1, 5–12. [Google Scholar] [CrossRef]
  3. Hayashi, M. Large deviation analysis for classical and quantum security via approximate smoothing. Quantum Phys. 2013. [Google Scholar]
  4. He, X.; Yener, A. Strong secrecy and reliable Byzantine detection in the presence of an untrusted relay. IEEE Trans. Inf. Theory 2013, 59, 177–192. [Google Scholar] [CrossRef]
  5. Fedotov, A.A.; Harremöes, P.; Tøpsoe, F. Refinements of Pinsker’s inequality. IEEE Trans. Inf. Theory 2003, 49, 1491–1498. [Google Scholar] [CrossRef]
  6. Hiai, F.; Ohya, M.; Tsukada, M. Sufficiency, KMS conditions and relative entropy in von Neumann algebras. Pac. J. Math. 1981, 96, 99–109. [Google Scholar] [CrossRef]
  7. Rastegin, A.E. Fano type quantum inequalities in terms of q-entropies. Quantum Inf. Process. 2012, 11, 1895–1910. [Google Scholar] [CrossRef]
  8. Brandão, F.G.S.L.; Harrow, A.W. Quantum de Finetti theorems under local measurements with applications. Quantum Phys. 2013. [Google Scholar]
  9. Zhang, Z. Estimating mutual information via Kolmogorov distance. IEEE Trans. Inf. Theory 2007, 53, 3280–3282. [Google Scholar] [CrossRef]
  10. Aaronson, B.; Franco, R.L.; Compagno, G.; Adesso, G. Hierarchy and dynamics of trace distance correlations. Quantum Phys. 2013. [Google Scholar]
  11. Paula, F.M.; Montealegre, J.D.; Saguia, A.; de Oliveira, T.R.; Sarandy, M.S. Geometric classical and total correlations via trace distance. Quantum Phys. 2013. [Google Scholar]
  12. Hall, M.J.W. Relaxed Bell inequalities and Kochen-Specker theorems. Phys. Rev. A 2012, 84, 022102. [Google Scholar] [CrossRef]
  13. Piani, M.; Horodecki, P.; Horodecki, R. No-local-broadcasting theorem for multipartite quantum correlations. Phys. Rev. Lett. 2008, 100, 090502. [Google Scholar] [CrossRef] [PubMed]
  14. Toner, B.F.; Bacon, D. Communication cost of simulating Bell correlations. Phys. Rev. Lett. 2003, 91, 187904. [Google Scholar] [CrossRef] [PubMed]
  15. Clauser, J.F.; Horne, M.A.; Shimony, A.; Holt, R.A. Proposed experiment to test local hidden-variable theories. Phys. Rev. Lett. 1969, 23, 880–884. [Google Scholar] [CrossRef]
  16. Csirel’son, B.S. Quantum generalizations of Bell’s inequality. Lett. Math. Phys. 1980, 4, 93–100. [Google Scholar] [CrossRef]
  17. Fano, U. Pairs of two-level systems. Rev. Mod. Phys. 1983, 55, 855–874. [Google Scholar] [CrossRef]
  18. Bhatia, R. Matrix Analysis; Springer-Verlag: New York, NY, USA, 1997; Section IV.2. [Google Scholar]
  19. Werner, R.F. Quantum states with Einstein-Podolsky-Rosen correlations admitting a hidden-variable model. Phys. Rev. A 1989, 40, 4277–4281. [Google Scholar] [CrossRef] [PubMed]
  20. Zhang, C.J.; Zhang, Y.S.; Zhang, S.; Guo, G.C. Entanglement detection beyond the computable cross-norm or realignment criterion. Phys. Rev. A 2008, 77, 060301(R). [Google Scholar] [CrossRef]
  21. Bennett, C.H.; DiVincenzo, D.P. Mixed-state entanglement and quantum error correction. Phys. Rev. A 1996, 54, 3824–3851. [Google Scholar] [CrossRef] [PubMed]
  22. Wu, S.; Poulsen, U.V.; Mølmer, K. Correlations in local measurements on a quantum state, and complementarity as an explanation of nonclassicality. Phys. Rev. A 2009, 80, 032319. [Google Scholar] [CrossRef]

Share and Cite

MDPI and ACS Style

Hall, M.J.W. Correlation Distance and Bounds for Mutual Information. Entropy 2013, 15, 3698-3713. https://doi.org/10.3390/e15093698

AMA Style

Hall MJW. Correlation Distance and Bounds for Mutual Information. Entropy. 2013; 15(9):3698-3713. https://doi.org/10.3390/e15093698

Chicago/Turabian Style

Hall, Michael J. W. 2013. "Correlation Distance and Bounds for Mutual Information" Entropy 15, no. 9: 3698-3713. https://doi.org/10.3390/e15093698

APA Style

Hall, M. J. W. (2013). Correlation Distance and Bounds for Mutual Information. Entropy, 15(9), 3698-3713. https://doi.org/10.3390/e15093698

Article Metrics

Back to TopTop