Next Article in Journal
Design of Optimal Rainfall Monitoring Network Using Radar and Road Networks
Next Article in Special Issue
A Method to Compute the Schrieffer–Wolff Generator for Analysis of Quantum Memory
Previous Article in Journal
A Generative Adversarial Network for Infrared and Visible Image Fusion Based on Semantic Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Classical Capacity of General Quantum Gaussian Measurement

Steklov Mathematical Institute, RAS, 119991 Moscow, Russia
Entropy 2021, 23(3), 377; https://doi.org/10.3390/e23030377
Submission received: 2 March 2021 / Revised: 19 March 2021 / Accepted: 19 March 2021 / Published: 21 March 2021
(This article belongs to the Special Issue Quantum Communication, Quantum Radar, and Quantum Cipher)

Abstract

:
In this paper, we consider the classical capacity problem for Gaussian measurement channels. We establish Gaussianity of the average state of the optimal ensemble in the general case and discuss the Hypothesis of Gaussian Maximizers concerning the structure of the ensemble. Then, we consider the case of one mode in detail, including the dual problem of accessible information of a Gaussian ensemble. Our findings are relevant to practical situations in quantum communications where the receiver is Gaussian (say, a general-dyne detection) and concatenation of the Gaussian channel and the receiver can be considered as one Gaussian measurement channel. Our efforts in this and preceding papers are then aimed at establishing full Gaussianity of the optimal ensemble (usually taken as an assumption) in such schemes.

1. Introduction

From the viewpoint of information theory, measurements are hybrid communication channels that transform input quantum states into classical output data. As such, they are described by the classical information capacity which is the most fundamental quantity characterizing their ultimate information-processing performance [1,2,3,4]. Channels with continuous output, such as bosonic Gaussian measurements, do not admit direct embedding into properly quantum channels and, hence, require separate treatment. In particular, their output entropy is the Shannon differential entropy, instead of the quantum entropy, which completely changes the pattern of the capacity formulas. The classical capacity of multimode Gaussian measurement channels was computed in Reference [5] under so-called threshold condition (which includes phase-insensitive or gauge covariant channels as a special case [6]). The essence of this condition is that it reduces the classical capacity problem to the minimum output differential entropy problem solved in Reference [7] (in the context of quantum Gaussian channels, a similar condition was introduced and studied in References [8,9]; also see references therein).
In this paper, we approach the classical capacity problem for Gaussian measurement channels without imposing any kind of threshold condition. In particular, in the framework of quantum communication, this means that both (noisy) heterodyne and (noisy/noiseless) homodyne measurements [10,11] are treated from a common viewpoint. We prove Gaussianity of the average state of the optimal ensemble in general and discuss the Hypothesis of Gaussian Maximizers (HGM) concerning the structure of the ensemble. The proof uses the approach of the paper of Wolf, Giedke, and Cirac [12] applied to the convex closure of the output differential entropy. Then, we discuss the case of one mode in detail, including the dual problem of accessible information of a Gaussian ensemble.
In quantum communications, there are several studies of the classical capacity in the transmission scheme where not only the Gaussian channel but also the receiver is fixed, and the optimization is performed over certain set of the input ensembles (see References [10,13,14,15] and references therein). These studies are practically important in view of greater complexity of the optimal receiver in the Quantum Channel Coding (HSW) theorem (see, e.g., Reference [16]). Our findings are relevant to such a situation where the receiver is Gaussian and concatenation of the channel and the receiver can be considered as one Gaussian measurement channel. Our efforts in this and preceding papers are then aimed at establishing full Gaussianity of the optimal ensemble (usually taken as a key assumption) in such schemes.

2. The Measurement Channel and Its Classical Capacity

An ensemble E = π ( d x ) , ρ ( x ) consists of probability measure π ( d x ) on a standard measurable space X and a measurable family of density operators (quantum states) x ρ ( x ) on the Hilbert space H of the quantum system. The average state of the ensemble is the barycenter of this measure:
ρ ¯ E = X ρ ( x ) π ( d x ) ,
the integral existing in the strong sense in the Banach space of trace-class operators on H .
Let M = { M ( d y ) } be an observable (POVM) on H with the outcome standard measurable space Y . There exists a σ finite measure μ ( d y ) such that, for any density operator ρ , the probability measure Tr ρ M ( d y ) is absolutely continuous w.r.t. μ ( d y ) , thus having the probability density p ρ ( y ) (one can take μ ( d y ) = Tr ρ 0 M ( d y ) , where ρ 0 is a nondegenerate density operator). The affine map M : ρ p ρ ( · ) will be called the measurement channel.
The joint probability distribution of x , y on X × Y is uniquely defined by the relation
P ( A × B ) = A π ( d x ) Tr ρ ( x ) M ( B ) = A B p ρ ( x ) ( y ) π ( d x ) μ ( d y ) ,
where A is an arbitrary Borel subset of X , and B is that of Y . The classical Shannon information between x , y is equal to
I ( E , M ) = π ( d x ) μ ( d y ) p ρ ( x ) ( y ) log p ρ ( x ) ( y ) p ρ ¯ E ( y ) .
In what follows, we will consider POVMs having (uniformly) bounded operator density, M ( d y ) = m ( y ) μ ( d y ) , with m ( y ) b , so that the probability densities p ρ ( y ) = Tr ρ m ( y ) are uniformly bounded, 0 p ρ ( y ) b . (The probability densities corresponding to Gaussian observables we will be dealing with possess this property). Moreover, without loss of generality [6] we can assume b = 1 . Then, the output differential entropy
h M ( ρ ) = p ρ ( y ) log p ρ ( y ) μ ( d y )
is well defined with values in [ 0 , + ] (see Reference [6] for the details). The output differential entropy is concave lower semicontinuous (w.r.t. trace norm) functional of a density operator ρ . The concavity follows from the fact that the function p p log p , p [ 0 , 1 ] is concave. Lower semicontinuity follows by an application of the Fatou-Lebesgue lemma from the fact that this function is nonnegative, continuous, and p ρ ( y ) p σ ( y ) ρ σ 1 .
Next, we define the convex closure of the output differential entropy (1):
e M ( ρ ) = inf E : ρ ¯ E = ρ h M ( ρ ( x ) ) π ( d x ) ,
which is the “measurement channel analog” of the convex closure of the output entropy for a quantum channel [17].
Lemma 1.
The functional e M ( ρ ) is convex, lower semicontinuous and strongly superadditive:
e M 1 M 2 ( ρ 12 ) e M 1 ( ρ 1 ) + e M 2 ( ρ 2 ) .
As it is well known, the property (3) along with the definition (2) imply additivity: if ρ 12 = ρ 1 ρ 2 then
e M 1 M 2 ( ρ 12 ) = e M 1 ( ρ 1 ) + e M 2 ( ρ 2 ) .
Proof. 
The lower semicontinuity follows from the similar property of the output differential entropy much in the same way as in the case of quantum channels, treated in Reference [17], Proposition 4; also see Reference [18], Proposition 1.
Let us prove strong superadditivity. Let
ρ 12 = ρ 12 ( x ) π ( d x )
be a decomposition of a density operator ρ 12 on H 1 H 2 , then
p M 1 M 2 ( y 1 , y 2 | x ) = Tr ρ 12 ( x ) m 1 ( y 1 ) m 2 ( y 2 ) = Tr ρ 1 ( x ) m 1 ( y 1 ) Tr ρ 2 ( y 1 , x ) m 2 ( y 2 ) = p M 1 ( y 1 | x ) p M 2 ( y 2 | y 1 , x ) ,
where ρ 1 ( x ) = Tr 2 ρ 12 ( x ) , ρ 2 ( y 1 , x ) = Tr 1 ρ 12 ( x ) m 1 ( y 1 ) I 2 Tr ρ 12 ( x ) m 1 ( y 1 ) I 2 , so that
Tr ρ 12 ( x ) m 1 ( y 1 ) I 2 = Tr ρ 1 ( x ) m 1 ( y 1 ) = p M 1 ( y 1 | x ) ,
and ρ 2 = ρ 2 ( y 1 , x ) p M 1 ( y 1 | x ) π ( d x ) μ 1 ( d y 1 ) while ρ 1 = ρ 1 ( x ) π ( d x ) . It follows that:
h ( Y 1 , Y 2 | X ) h M 1 M 2 ( ρ 12 ( x ) ) π ( d x ) = h M 1 ( ρ 1 ( x ) ) π ( d x ) + h M 2 ( ρ 2 ( y 1 , x ) ) p M 1 ( y 1 | x ) π ( d x ) μ 1 ( d y 1 ) = h ( Y 1 | X ) + h ( Y 2 | Y 1 , X ) ,
and, whence taking the infimum over decompositions (5), we obtain (3). □
Let H be a Hamiltonian in the Hilbert space H of the quantum system, E a positive number. Then, the energy-constrained classical capacity of the channel M is equal to
C ( M , H , E ) = sup E : Tr ρ ¯ E H E I ( E , M ) ,
where maximization is over the input ensembles of states E satisfying the energy constraint Tr ρ ¯ E H E , as shown in Reference [5], proposition 1.
If h M ( ρ ¯ E ) < + , then
I ( E , M ) = h M ( ρ ¯ E ) h M ( ρ ( x ) ) π ( d x ) .
Note that the measurement channel is entanglement-breaking [16]; hence, its classical capacity is additive and is given by the one-shot expression (6). By using (7), (2), we obtain
C ( M , H , E ) = sup ρ : Tr ρ H E h M ( ρ ) e M ( ρ ) .

3. Gaussian Maximizers for Multimode Bosonic Gaussian Observable

Consider now multimode bosonic Gaussian system with the quadratic Hamiltonian H = R ϵ R t , where ϵ > 0 is the energy matrix, and R = q 1 , p 1 , , q s , p s is the row vector of the bosonic position-momentum observables, satisfying the canonical commutation relation
[ R t , R ] = i Δ I , Δ = diag 0 1 1 0 1 , , s ¯ ,
(see, e.g., References [11,16]). This describes quantization of a linear classical system with s degrees of freedom, such as finite number of physically relevant electromagnetic modes on the receiver’s aperture in quantum optics.
From now on, we will consider only states with finite second moments. By S ( α ) , we denote the set of all states ρ with the fixed correlation matrix
α = Re Tr R t ρ R .
For centered states (i.e., states with vanishing first moments), the covariance matrix and the matrix of second moments coincide. We denote by ρ α centered Gaussian state with the correlation matrix α ± i / 2 Δ . For states ρ S ( α ) , we have h M ( ρ ) h M ( ρ α ) < + , by the maximum entropy principle.
The energy constraint reduces to
Sp α ϵ E .
(We denote Sp trace of s × s -matrices as distinct from trace of operators on H .)
For a fixed correlation matrix α , we will study the α -constrained capacity
C ( M ; α ) = sup E : ρ ¯ E S ( α ) I ( E , M ) = sup ρ S ( α ) h M ( ρ ) e M ( ρ ) .
With the Hamiltonian H = R ϵ R t , the energy-constrained classical capacity of observable M is
C ( M ; H , E ) = sup α : Sp α ϵ E C ( M ; α ) .
We will be interested in the approximate position-momentum measurement (observable, POVM)
M ( d 2 s z ) = D ( z ) ρ β D ( z ) * d 2 s z 2 π s
where ρ β is centered Gaussian density operator with the covariance matrix β and
D ( z ) = exp i j = 1 s y j q j x j p j , z = x 1 , y 1 , , x s , y s t R 2 s
are the unitary displacement operators. Thus, μ ( d z ) = d 2 s z 2 π s and the operator-valued density of POVM (11) is m ( z ) = D ( z ) ρ β D ( z ) * . In quantum optics, some authors [11,19] call such measurements (noisy) general-dyne detections.
In what follows, we will consider n independent copies of our bosonic system on the Hilbert space H n . We will supply all the quantities related to k - th copy ( k = 1 , , n ) with upper index ( k ) , and we will use tilde to denote quantities related to the whole collection on n copies. Thus,
z ˜ = z ( 1 ) z ( n ) , D ( z ˜ ) = D ( z ( 1 ) ) D ( z ( n ) )
and
M n ( d z ˜ ) = m ˜ ( z ˜ ) μ ˜ ( d z ˜ ) = m ( z ( 1 ) ) m ( z ( n ) ) μ ( d z ( 1 ) ) μ ( d z ( n ) ) .
Lemma 2.
Let O = O k l k , l = 1 , , n be a real orthogonal n × n matrix and U—the unitary operator on H n implementing the linear symplectic transformation
R ˜ = R ( 1 ) , , R ( n ) R ˜ O ,
so that
U * D ( z ˜ ) U = D ( O z ˜ ) .
Then, for any state ρ ˜ on H n ,
e M n ( ρ ˜ ) = e M n ( U ρ ˜ U * ) .
Proof. 
The covariance matrix β ˜ of ρ β n is block-diagonal, β ˜ = [ δ k l β ] k , l = 1 , , n ; hence, O t β ˜ O = β ˜ . Thus, we have U * ρ β n U = ρ β n , and taking into account (12),
U * m ˜ ( z ˜ ) U = D ( O z ˜ ) ρ β n D ( O z ˜ ) * = m ˜ ( O z ˜ ) .
Therefore, for any state σ ˜ on H n , the output probability density of the measurement channel M ˜ = M n corresponding to the input state U σ ˜ U * is
p U σ ˜ U * ( z ˜ ) = Tr U σ ˜ U * m ˜ ( z ˜ ) = Tr σ ˜ m ˜ ( O z ˜ ) = p σ ˜ ( O z ˜ ) .
Hence, by using orthogonal invariance of the Lebesgue measure,
h M n ( U σ ˜ U * ) = h M n ( σ ˜ ) .
If ρ ˜ = X ρ ˜ ( x ) π ( d x ) , then U ρ ˜ U * = X U ρ ˜ ( x ) U * π ( d x ) , and taking σ ˜ = ρ ˜ ( x ) in the previous formula, we deduce
X h M n ( U ρ ˜ ( x ) U * ) π ( d x ) = X h M n ( ρ ˜ ( x ) ) π ( d x ) ;
hence, (13) follows. □
Lemma 3.
Let M be the Gaussian measurement (11). For any state ρ with finite second moments, e M ( ρ ) e M ( ρ α ) , where α is the covariance matrix of ρ.
Proof. 
The proof follows the pattern of Lemma 1 from the paper of Wolf, Giedke, and Cirac [12]. Without loss of generality, we can assume that ρ is centered. We have
e M ( ρ ) = ( 1 ) 1 n e M n ( ρ n ) = ( 2 ) 1 n e M n ( ρ ˜ ) ( 3 ) 1 n k = 1 n e M ( ρ ˜ ( k ) ) ,
where ρ ˜ = U ρ n U * with symplectic unitary U in H n , corresponding to an orthogonal matrix O as in Lemma 2, and ρ ˜ ( k ) is the k - th partial state of ρ ˜ .
Step (1) follows from the additivity (4). Step (2) follows from Lemma 2, and step (3) follows from the superadditivity of e M (Lemma 1). The final step of the proof,
lim inf n 1 n k = 1 n e M ( ρ ˜ ( k ) ) e M ( ρ α ) ,
uses ingeniously constructed U from Reference [12] and lower semicontinuity of e M (Lemma 1). Namely, n = 2 m , and U corresponds via (12) to the following special orthogonal matrix
O = O k l k , l = 1 , , n = H m , H = 1 2 1 1 1 1 .
Every row of the n × n matrix O, except the first one which has all the elements 1, has n / 2 = 2 m 1 elements equal to 1 and n / 2 elements equal to −1. Then, the quantum characteristic function of the states ρ ˜ ( k ) , k = 2 , , n is equal to ϕ ( z / n ) n / 2 ϕ ( z / n ) n / 2 , where ϕ ( z ) is the quantum characteristic function of the state ρ . This allows to apply Quantum Central Limit Theorem [20] to show that ρ ˜ ( k ) ρ α as n , in a uniform way, implying (16); see Reference [12] for details. □
Theorem 1.
The optimizing density operator ρ in (10) is the (centered) Gaussian density operator ρ α :
C ( M ; α ) = h M ( ρ α ) e M ( ρ α ) ,
and, hence,
C ( M , H , E ) = max α : Sp α ϵ E C ( M ; α ) = max α : Sp α ϵ E h M ( ρ α ) e M ( ρ α ) .
Proof. 
Lemma 3 implies that, for any ρ with finite second moments, e M ( ρ ) e M ( ρ α ) , where α is the covariance matrix of ρ . On the other hand, by the maximum entropy principle, h M ( ρ ) h M ( ρ α ) . Hence, (17) is maximized by a Gaussian density operator. □
Remark 1.
The proof of Lemma 2 and, hence, of Theorem 1 can be extended to a general Gaussian observable M in the sense of References [16,21], defined via operator-valued characteristic function of the form
ϕ M ( w ) = exp i R K w 1 2 w t γ w ,
where K is a scaling matrix, γ is the measurement noise covariance matrix, and γ ± i 2 K t Δ K . Then, the Fourier transform of the measurement probability density p ρ ( z ) is equal to Tr ρ ϕ M ( w ) , and one can use this function to obtain generalization of the relation (14) for the measurement probability densities. The case (11) corresponds to the type 1 Gaussian observable [21] with K = I 2 s , γ = β . However, (19) also includes type 2 and 3 observables (noisy and noiseless multimode homodyning), in which case K is a projection onto an isotropic subspace of Z (i.e., one on which the symplectic form Δ vanish.)
Remark 2.
Theorem 1 establishes Gaussianity of the average state of the optimal ensemble for a general Gaussian measurement channel. However, Gaussian average state can appear in a non-Gaussian ensemble. An immediate example is thermal state represented as a mixture of the Fock states with geometric distribution. Thus, Theorem 1 does not necessarily imply full Gaussianity of the optimal ensemble as formulated in the following conjecture.
Hypothesis of Gaussian Maximizers (HGM).
Let M be an arbitrary Gaussian measurement channel. Then, there exists an optimal Gaussian ensemble for the convex closure of the output differential entropy (2) with Gaussian ρ and, hence, for the energy-constrained classical capacity (6) of the channel M. More explicitly, the ensemble consists of (properly squeezed) coherent states with the displacement parameter having Gaussian probability distribution.
For Gaussian measurement channels of the type 1 (essentially of the form (11), see Reference [21] for complete classification) and Gaussian states ρ α satisfying the “threshold condition”, we have
e M ( ρ α ) = min ρ h M ( ρ ) ,
with the minimum attained on a squeezed coherent state, which implies the validity of the HGM and an efficient computation of C ( M , H , E ) ; see Reference [5]. On the other hand, the problem remains open in the case where the “threshold condition” is violated, and in particular, for all Gaussian measurement channels of the type 2 (noisy homodyning), with the generic example of the energy-constrained approximate measurement of the position q 1 , , q s subject to Gaussian noise (see Reference [22], where the entanglement-assisted capacity of such a measurement was computed). In the following section, we will touch upon the HGM in this case for one mode system.

4. Gaussian Measurements in One Mode

Our framework in this section will be one bosonic mode described by the canonical position and momentum operators q , p . We recall that
D ( x , y ) = exp i y q x p , x , y R
are the unitary displacement operators.
We will be interested in the observable
M ( d x d y ) = D ( x , y ) ρ β D ( x , y ) * d x d y 2 π ,
where ρ β is centered Gaussian density operator with the covariance matrix
β = β q 0 0 β p ; β q β p 1 4 .
Let ρ α be a centered Gaussian density operator with the covariance matrix
α = α q 0 0 α p .
The problem is, to compute e M ( ρ α ) and, hence, the classical capacity C ( M , H , E ) for the oscillator Hamiltonian H = 1 2 q 2 + p 2 (as shown in the Appendix of Reference [22], we can restrict to Gaussian states ρ α with the diagonal covariance matrix in this case). The energy constraint (9) takes the form
α q + α p 2 E .
The measurement channel corresponding to POVM (21) acts on the centered Gaussian state ρ α by the formula
M : ρ α p ρ α ( x , y ) = 1 2 π α q + β q α p + β p exp x 2 2 α q + β q y 2 2 α p + β p ,
so that
h M ( ρ α ) = 1 2 log α q + β q α p + β p + c .
In this expression, c is a fixed constant depending on the normalization of the underlying measure μ in (1). It does not enter the information quantities which are differences of the two differential entropies.
Assuming validity of the HGM, we will optimize over ensembles of squeezed coherent states
ρ x , y = D ( x , y ) ρ Λ D ( x , y ) * , ( x , y ) R 2 ,
where ρ Λ is centered Gaussian state with correlation matrix Λ = δ 0 0 1 / 4 δ , and the vector ( x , y ) has centered Gaussian distribution with covariance matrix γ q 0 0 γ p . Then, the average state ρ ¯ E of the ensemble is centered Gaussian ρ α with the covariance matrix (23), where
α q = γ q + δ , α p = γ p + 1 / 4 δ ;
hence,
1 4 α p δ α q .
For this ensemble,
h M ( ρ x , y ) π ( d x d y ) = h M ( ρ Λ ) = 1 2 log δ + β q 1 / 4 δ + β p + c .
Then, the hypothetical value:
e M ( ρ α ) = min 1 / 4 α p δ α q 1 2 log δ + β q 1 / 4 δ + β p + c .
The derivative of the minimized expression vanishes for δ = 1 2 β q β p . Thus, depending on the position of this value with respect to the interval (27), we obtain three possibilities):
Here, the column C corresponds to the case where the “threshold condition” holds, implying (20). Then the full validity of the HGM in much more general multimode situation was established in Reference [5]. All the quantities in this column, as well as the value of C ( M , H , E ) in the central column of Table 2, were obtained in that paper as an example. On the other hand, the HGM remains open in the cases of mutually symmetric columns L and R (for the derivation of the quantities in column L of Table 1 and Table 2 see Appendix A).
Maximizing C ( M ; α ) over α q , α p which satisfy the energy constraint (24) (with the equality): α q + α p = 2 E , we obtain C ( M , H , E ) depending on the signal energy E and the measurement noise variances β q , β p :
Table 2. The values of the capacity C ( M , H , E ) .
Table 2. The values of the capacity C ( M , H , E ) .
L: HGM openC: HGM valid [5]R: HGM open
β q β p ; E < E β p , β q E E β p , β q E β q , β p β p β q ; E < E β q , β p
log 1 + 8 E β q + 4 β q 2 1 2 β q log E + β q + β p / 2 β q β p + 1 / 2 log 1 + 8 E β p + 4 β p 2 1 2 β p
where we introduced the “energy threshold function”
E β 1 , β 2 = 1 2 β 1 β 2 + β 1 β 2 .
In the gauge invariant case when β q = β p = β , the threshold condition amounts to E 1 / 2 , which is fulfilled by definition, and the capacity formula gives the expression log E + β β + 1 / 2 equivalent to one obtained in Hall’s 1994 paper [13].
Let us stress that, opposite to column C, the values of C ( M , H , E ) in the L and R columns are hypothetic, conditional upon validity of the HGM. Looking into the left column, one can see that C ( M ; α ) and C ( M , H , E ) do not depend at all on β p . Thus, we can let the variance of the momentum p measurement noise β p + , and, in fact, set β p = + , which is equivalent to the approximate measurement only of the position q described by POVM
M ( d x ) = exp q x 2 2 β q d x 2 π β q = D ( x , 0 ) e q 2 / 2 β q D ( x , 0 ) * d x 2 π β q ,
which belongs to type 2 according to the classification of Reference [21]. In other words, one makes the “classical” measurement of the observable
X = q + ξ , ξ N ( 0 , β q ) ,
with the quantum energy constraint Tr ρ ( q 2 + p 2 ) 2 E .
The measurement channel corresponding to POVM (29) acts on the centered Gaussian state ρ α by the formula
M : ρ α p ρ α ( x ) = 1 2 π α q + β q exp x 2 2 α q + β q .
In this case, we have
h M ( ρ α ) = 1 2 log α q + β q + c ,
e M ( ρ α ) = 1 2 log 1 / 4 α p + β q + c ,
which differ from the values in the case of finite β p + by the absence of the factor α p + β p under the logarithms, while the difference C ( M ; α ) = h M ( ρ α ) e M ( ρ α ) and the capacity C ( M , H , E ) have the same expressions as in that case (column L).
For β q = 0 (sharp position measurement, type 3 of Reference [21]), the HGM is valid with
C ( M , H , E ) = log 2 E .
This follows from the general upper bound (Figure 1)
C ( M , H , E ) log 1 + E 1 / 2 β q + 1 / 2 = log 2 ( E + β q ) 1 + 2 β q
for β q 0 (Equation (28) in Reference [23]; also see Equation (5.39) in Reference [10]).

5. The Dual Problem: Accessible Information

Let us sketch here ensemble-observable duality [1,2,4] (see Reference [6] for details of mathematically rigorous description in the infinite dimensional case).
Let E = π ( d x ) , ρ ( x ) be an ensemble, μ ( d y ) a σ finite measure and M = M ( d y ) an observable having operator density m ( y ) = M ( d y ) / μ ( d y ) with values in the algebra of bounded operators in H . The dual pair ensemble-observable E , M is defined by the relations
E : π ( d y ) = Tr ρ ¯ E M ( d y ) , ρ ( y ) = ρ ¯ E 1 / 2 m ( y ) ρ ¯ E 1 / 2 Tr ρ ¯ E m ( y ) ;
M : M ( d x ) = ρ ¯ E 1 / 2 ρ ( x ) ρ ¯ E 1 / 2 π ( d x ) .
Then, the average states of both ensembles coincide
ρ ¯ E = ρ ¯ E
and the joint distribution of x , y is the same for both pairs ( E , M ) and ( E , M ) so that
I ( E , M ) = I ( E , M ) .
Moreover,
sup M I ( E , M ) = sup E : ρ ¯ E = ρ ¯ E I ( E , M ) ,
where the supremum in the right-hand side is taken over all ensembles E satisfying the condition ρ ¯ E = ρ ¯ E . It can be shown (Reference [6], Proposition 4), that the supremum in the lefthand side remains the same if it is taken over all observables M (not only of the special kind with the density we started with), and then it is called the accessible information A ( E ) of the ensemble E . Thus,
A ( E ) = sup E : ρ ¯ E = ρ ¯ E I ( E , M ) .
Since the application of the duality to the pair E , M results in the initial pair E , M , we also have
A ( E ) = sup M I ( E , M ) = sup E : ρ ¯ E = ρ ¯ E I ( E , M ) .
Coming to the case of bosonic mode, we fix the Gaussian state ρ α and restrict to ensembles E with ρ ¯ E = ρ α . Let M be the measurement channel corresponding to POVM (21). Then, according to formulas (34), the dual ensemble E = p ( x , y ) , ρ ( x , y ) , where p ( x , y ) is the Gaussian probability density (25) and
ρ ( x , y ) = p ( x , y ) 1 ρ α D ( x , y ) ρ β D ( x , y ) * ρ α .
By using the formula for ρ 1 ρ 2 ρ 1 , where ρ 1 , ρ 2 are Gaussian operators (see Reference [24] and also Corollary in the Appendix of Reference [25]), we obtain
ρ ( x , y ) = D ( x , y ) ρ α D ( x , y ) * = ρ α ( x , y ) ,
where
α = α γ , γ = κ α + β 1 κ , x y = κ α + β 1 x y ,
and
κ = I + 2 α Δ 1 2 α = α I + 2 Δ 1 α 2 .
Since x y t N ( 0 , α + β ) , then, from second and third equations in (39), we obtain x y t N ( 0 , κ α + β 1 κ ) = N ( 0 , γ ) . By denoting p γ ( x , y ) , the density of this normal distribution, we can equivalently rewrite the ensemble E as E = p γ ( x , y ) , ρ α ( x , y ) with the average state ρ α , α = α + γ . Then, HGM is equivalent to the statement
A ( E ) = C M ; α ,
where the values of C M ; α are given in Table 1; however, they should be reexpressed in terms of the ensemble parameters γ , α . In Reference [25], we treated the case C in multimode situation, establishing that the optimal measurement is Gaussian, and described it. Here, we will discuss the case L (R is similar) and show that, for large β p (including β p = + ), the HGM is equivalent to the following: the value of the accessible information
A ( E ) = C M ; α = 1 2 log α q + β q 1 4 α p + β q
is attained on the sharp position measurement M 0 ( d ξ ) = | ξ ξ | d ξ (in fact, this refers to the whole domain L: 1 2 β q β p < 1 4 α p , which, however, has rather cumbersome description in the new variables γ , α , cf. Reference [25]).
In the one mode case we are considering, the matrix α is given by (23), β —by (22), and Δ = 0 1 1 0 , so that 2 Δ 1 α 2 = 4 α q α p I . Computations according to (39) and (40) give
α = α q 0 0 α p = α q β q + 1 / 4 α p α q + β q 0 0 α p β p + 1 / 4 α q α p + β p .
But under the sharp position measurement M 0 ( d ξ ) = | ξ ξ | d ξ , one has (in the formulas below, p ( ξ ) = N ( m , α ) means that p ( ξ ) is Gaussian probability density with mean m and variance α ):
p ( ξ | x , y ) = ξ | ρ α ( x , y ) | ξ = N ( x , α q ) ,
while ξ | ρ α | ξ = N ( 0 , α q ) (note that ρ ¯ E = ρ ¯ E = ρ α ), and
I E , M 0 = 1 2 log α q + γ q log α q = 1 2 log α q log α q β q + 1 / 4 α p α q + β q = 1 2 log α q + β q β q + 1 / 4 α p ,
which is identical to the expression in (41).
In the case of the position measurement channel M corresponding to POVM (29) ( β p = + ) , we have α p = α p ; otherwise, the argument is essentially the same. Thus, we obtain that the HGM concerning e M ( ρ ) in case L is equivalent to the following:
The accessible information of a Gaussian ensemble E = p ( x ) , ρ ( x ) , where
p ( x ) = N ( 0 , γ q ) , ρ ( x ) = D ( x , 0 ) ρ α D ( x , 0 ) * ,
is given by the expression (43) and attained on the sharp position measurement M 0 ( d x ) = | ξ ξ | d ξ .

6. Discussion

In this paper, we investigated the classical capacity problem for Gaussian measurement channels. We established Gaussianity of the average state of the optimal ensemble in full generality and discussed the Hypothesis of Gaussian Maximizers concerning the detailed structure of the ensemble. Gaussian systems form the backbone of information theory with continuous variables, both in the classical and in the quantum case. Starting from them, other, non-linear models can be constructed and investigated. Therefore, the quantum Gaussian models must be studied exhaustively. Despite the progress made, there are still intriguing gaps along this way. A major problem remains the proof (or refutation) of the hypothesis of Gaussian optimizers for various entropy characteristics of quantum Gaussian systems and channels. So far, the proof of this hypothesis in special cases required tricky and special constructions, such as in the path-breaking paper [7] concerning gauge-covariant channels, or in Section 3 of the present work concerning general Gaussian measurement channels. It seems plausible that quantum Gaussian systems may have some as yet undiscovered structural property, from which a proof of this hypothesis in its maximum generality would follow in a natural way.

Funding

This work was performed at the Steklov International Mathematical Center and supported by the Ministry of Science and Higher Education of the Russian Federation (agreement no. 075-15-2019-1614).

Acknowledgments

The author is grateful to M. J. W. Hall for sending a copy of his paper [13], and to M. E. Shirokov for the comments improving the presentation.

Conflicts of Interest

The author declares no conflict of interest.

Appendix A. Case L in Tables 1 and 2

By taking the Gaussian ensemble parameters in (28) as
δ = 1 / 4 α p , γ p = 0 , γ q = α q 1 / 4 α p ,
we get the hypothetic value
e M ( ρ α ) = 1 2 log 1 4 α p + β q α p + β p + c ,
hence taking into account (26),
C G a u s s ( M ; α ) = h M ( ρ α ) e M ( ρ α ) = 1 2 log α q + β q 1 4 α p + β q .
The Gaussian constrained capacity is
C G a u s s ( M , H , E ) = max α q + α q 2 E 1 2 log α q + β q log 1 / 4 α p + β q = max α p 1 2 log 2 E α p + β q log 1 / 4 α p + β q ,
where, in the second line, we took the maximal value α q = 2 E α p . Differentiating, we obtain the equation for the optimal value α p :
4 β q α p 2 + 2 α p 2 E + β q = 0 ,
the positive solution of which is
α p = 1 4 β q 1 + 8 E β q + 4 β q 2 1 ,
whence
C G a u s s ( M , H , E ) = log 1 + 8 E β q + 4 β q 2 1 2 β q .
The parameters of the optimal Gaussian ensemble are obtained by substituting the value (A5) into (A1) with α q = 2 E α p .
The above derivation concerns the measurement (21) ( β p < ) . The case of the measurement (29) ( β p = + ) is treated similarly, with (A2), (26) replaced by (32), (31). Notably, in this case, the expression (A6) coincides with the one obtained in Reference [13] by optimizing the information from applying sharp position measurement to noisy optimally squeezed states (the author is indebted to M. J. W. Hall for this observation).

References

  1. Hall, M.J.W. Quantum information and correlation bounds. Phys. Rev. A 1997, 55, 1050–2947. [Google Scholar] [CrossRef]
  2. Dall’Arno, M.; D’Ariano, G.M.; Sacchi, M.F. Informational power of quantum measurements. Phys. Rev. A 2011, 83, 062304. [Google Scholar] [CrossRef] [Green Version]
  3. Oreshkov, O.; Calsamiglia, J.; Munoz-Tapia, R.; Bagan, E. Optimal signal states for quantum detectors. New J. Phys. 2011, 13, 073032. [Google Scholar] [CrossRef]
  4. Holevo, A.S. Information capacity of quantum observable. Probl. Inform. Transm. 2012, 48, 1–10. [Google Scholar] [CrossRef] [Green Version]
  5. Holevo, A.S.; Kuznetsova, A.A. Information capacity of continuous variable measurement channel. J. Phys. A Math. Theor. 2020, 53, 175304. [Google Scholar] [CrossRef] [Green Version]
  6. Holevo, A.S. Gaussian maximizers for quantum Gaussian observables and ensembles. IEEE Trans. Inform. Theory 2020, 66, 5634–5641. [Google Scholar] [CrossRef] [Green Version]
  7. Giovannetti, V.; Holevo, A.S.; Mari, A. Majorization and additivity for multimode bosonic Gaussian channels. Theor. Math. Phys. 2015, 182, 284–293. [Google Scholar] [CrossRef]
  8. Schäfer, J.; Karpov, E.; Garćia-Patrón, R.; Pilyavets, O.V.; Cerf, N.J. Equivalence Relations for the Classical Capacity of Single-Mode Gaussian Quantum Channels. Phys. Rev. Lett. 2013, 111, 030503. [Google Scholar] [CrossRef] [Green Version]
  9. Holevo, A.S. On the constrained classical capacity of infinite-dimensional covariant channels. J. Math. Phys. 2016, 57, 15203. [Google Scholar] [CrossRef] [Green Version]
  10. Caves, C.M.; Drummond, P.D. Quantum limits on bosonic communication rates. Rev. Mod. Phys. 1994, 68, 481–537. [Google Scholar] [CrossRef]
  11. Serafini, A. Quantum Continuous Variables: A Primer of Theoretical Methods; CRC Press: Boca Raton, FL, USA, 2017. [Google Scholar]
  12. Wolf, M.M.; Giedke, G.; Cirac, J.I. Extremality of Gaussian quantum states. Phys. Rev. Lett. 2006, 96, 080502. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Hall, M.J.W. Gaussian noise and quantum optical communication. Phys. Rev. A 1994, 50, 3295–3303. [Google Scholar] [CrossRef] [PubMed]
  14. Takeoka, M.; Guha, S. Capacity of optical communication in loss and noise with general Gaussian receivers. Phys. Rev. A 2014, 89, 042309. [Google Scholar] [CrossRef] [Green Version]
  15. Lee, J.; Ji, S.W.; Park, J.; Nha, H. Gaussian benchmark for optical communication aiming towards ultimate capacity. Phys. Rev. A 2016, 93, 050302. [Google Scholar] [CrossRef] [Green Version]
  16. Holevo, A.S. Quantum Systems, Channels, Information: A Mathematical Introduction, 2nd ed.; De Gruyter: Berlin, Germany; Boston, MA, USA, 2019. [Google Scholar]
  17. Shirokov, M.E. On entropic quantities related to the classical capacity of infinite dimensional quantum channels. Theor. Probab. Appl. 2007, 52, 250–276. [Google Scholar] [CrossRef] [Green Version]
  18. Shirokov, M.E. On properties of the space of quantum states and their application to the construction of entanglement monotones. Izv. Math. 2010, 74, 849–882. [Google Scholar] [CrossRef] [Green Version]
  19. Wiseman, H.M.; Milburn, G.J. Quantum Measurement and Control; Cambridge University Press: New York, NY, USA, 2010. [Google Scholar]
  20. Cushen, C.D.; Hudson, R.L. A quantum mechanical central limit theorem. J. Appl. Prob. 1971, 8, 454–469. [Google Scholar] [CrossRef]
  21. Holevo, A.S. The structure of general quantum Gaussian observable. arXiv 2020, arXiv:2007.02340. [Google Scholar]
  22. Holevo, A.S.; Yashin, V.I. Maximum information gain of approximate quantum position measurement. Quantum Inf. Process. 2021, 20, 97. [Google Scholar] [CrossRef]
  23. Hall, M.J.W. Information exclusion principle for complementary observables. Phys. Rev. Lett. 1995, 74, 3307. [Google Scholar] [CrossRef] [PubMed]
  24. Lami, L.; Das, S.; Wilde, M.M. Approximate reversal of quantum Gaussian dynamics. J. Phys. A 2018, 51, 125301. [Google Scholar] [CrossRef] [Green Version]
  25. Holevo, A.S. Accessible information of a general quantum Gaussian ensemble. arXiv 2021, arXiv:2102.01981. [Google Scholar]
Figure 1. (color online) The Gaussian classical capacity (A6) and the upper bound (33) ( β = 1 ).
Figure 1. (color online) The Gaussian classical capacity (A6) and the upper bound (33) ( β = 1 ).
Entropy 23 00377 g001
Table 1. The three parameter ranges.
Table 1. The three parameter ranges.
rangeL: 1 2 β q β p < 1 4 α p C: 1 4 α p 1 2 β q β p α q R: α q < 1 2 β q β p
HGMopenvalidopen
δ o p t 1 / 4 α p 1 2 β q β p α q
e M ( ρ α ) c 1 2 log 1 4 α p + β q log β q β p + 1 / 2 1 2 log 1 4 α q + β p
× α p + β p ] × α q + β q ]
C ( M ; α ) 1 2 log α q + β q 1 4 α p + β q 1 2 log α q + β q α p + β p β q β p + 1 / 2 2 1 2 log α p + β p 1 4 α q + β p
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Holevo, A. On the Classical Capacity of General Quantum Gaussian Measurement. Entropy 2021, 23, 377. https://doi.org/10.3390/e23030377

AMA Style

Holevo A. On the Classical Capacity of General Quantum Gaussian Measurement. Entropy. 2021; 23(3):377. https://doi.org/10.3390/e23030377

Chicago/Turabian Style

Holevo, Alexander. 2021. "On the Classical Capacity of General Quantum Gaussian Measurement" Entropy 23, no. 3: 377. https://doi.org/10.3390/e23030377

APA Style

Holevo, A. (2021). On the Classical Capacity of General Quantum Gaussian Measurement. Entropy, 23(3), 377. https://doi.org/10.3390/e23030377

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop