1. Introduction
From the viewpoint of information theory, measurements are hybrid communication channels that transform input quantum states into classical output data. As such, they are described by the classical information capacity which is the most fundamental quantity characterizing their ultimate information-processing performance [
1,
2,
3,
4]. Channels with continuous output, such as bosonic Gaussian measurements, do not admit direct embedding into properly quantum channels and, hence, require separate treatment. In particular, their output entropy is the Shannon differential entropy, instead of the quantum entropy, which completely changes the pattern of the capacity formulas. The classical capacity of multimode Gaussian measurement channels was computed in Reference [
5] under so-called threshold condition (which includes phase-insensitive or gauge covariant channels as a special case [
6]). The essence of this condition is that it reduces the classical capacity problem to the minimum output differential entropy problem solved in Reference [
7] (in the context of quantum Gaussian channels, a similar condition was introduced and studied in References [
8,
9]; also see references therein).
In this paper, we approach the classical capacity problem for Gaussian measurement channels without imposing any kind of threshold condition. In particular, in the framework of quantum communication, this means that both (noisy) heterodyne and (noisy/noiseless) homodyne measurements [
10,
11] are treated from a common viewpoint. We prove Gaussianity of the average state of the optimal ensemble in general and discuss the Hypothesis of Gaussian Maximizers (HGM) concerning the structure of the ensemble. The proof uses the approach of the paper of Wolf, Giedke, and Cirac [
12] applied to the convex closure of the output differential entropy. Then, we discuss the case of one mode in detail, including the dual problem of accessible information of a Gaussian ensemble.
In quantum communications, there are several studies of the classical capacity in the transmission scheme where not only the Gaussian channel but also the receiver is fixed, and the optimization is performed over certain set of the input ensembles (see References [
10,
13,
14,
15] and references therein). These studies are practically important in view of greater complexity of the optimal receiver in the Quantum Channel Coding (HSW) theorem (see, e.g., Reference [
16]). Our findings are relevant to such a situation where the receiver is Gaussian and concatenation of the channel and the receiver can be considered as one Gaussian measurement channel. Our efforts in this and preceding papers are then aimed at establishing full Gaussianity of the optimal ensemble (usually taken as a key assumption) in such schemes.
2. The Measurement Channel and Its Classical Capacity
An
ensemble consists of probability measure
on a standard measurable space
and a measurable family of density operators (quantum states)
on the Hilbert space
of the quantum system. The
average state of the ensemble is the barycenter of this measure:
the integral existing in the strong sense in the Banach space of trace-class operators on
.
Let be an observable (POVM) on with the outcome standard measurable space . There exists a finite measure such that, for any density operator , the probability measure is absolutely continuous w.r.t. thus having the probability density (one can take , where is a nondegenerate density operator). The affine map will be called the measurement channel.
The joint probability distribution of
on
is uniquely defined by the relation
where
A is an arbitrary Borel subset of
, and
B is that of
The classical Shannon information between
is equal to
In what follows, we will consider POVMs having (uniformly) bounded operator density,
with
so that the probability densities
are uniformly bounded,
. (The probability densities corresponding to Gaussian observables we will be dealing with possess this property). Moreover, without loss of generality [
6] we can assume
Then, the output differential entropy
is well defined with values in
(see Reference [
6] for the details). The output differential entropy is concave lower semicontinuous (w.r.t. trace norm) functional of a density operator
. The concavity follows from the fact that the function
is concave. Lower semicontinuity follows by an application of the Fatou-Lebesgue lemma from the fact that this function is nonnegative, continuous, and
Next, we define the
convex closure of the output differential entropy (
1):
which is the “measurement channel analog” of the convex closure of the output entropy for a quantum channel [
17].
Lemma 1. The functional is convex, lower semicontinuous and strongly superadditive: As it is well known, the property (
3) along with the definition (
2) imply
additivity: if
then
Proof. The lower semicontinuity follows from the similar property of the output differential entropy much in the same way as in the case of quantum channels, treated in Reference [
17], Proposition 4; also see Reference [
18], Proposition 1.
Let us prove strong superadditivity. Let
be a decomposition of a density operator
on
, then
where
so that
and
while
It follows that:
and, whence taking the infimum over decompositions (
5), we obtain (
3). □
Let
H be a Hamiltonian in the Hilbert space
of the quantum system,
E a positive number. Then, the
energy-constrained classical capacity of the channel
M is equal to
where maximization is over the input ensembles of states
satisfying the energy constraint
, as shown in Reference [
5], proposition 1.
If
, then
Note that the measurement channel is entanglement-breaking [
16]; hence, its classical capacity is additive and is given by the one-shot expression (
6). By using (
7), (
2), we obtain
3. Gaussian Maximizers for Multimode Bosonic Gaussian Observable
Consider now multimode bosonic Gaussian system with the quadratic Hamiltonian
where
is the energy matrix, and
is the row vector of the bosonic position-momentum observables, satisfying the canonical commutation relation
(see, e.g., References [
11,
16]). This describes quantization of a linear classical system with
s degrees of freedom, such as finite number of physically relevant electromagnetic modes on the receiver’s aperture in quantum optics.
From now on, we will consider only states with finite second moments. By
, we denote the set of all states
with the fixed correlation matrix
For
centered states (i.e., states with vanishing first moments), the covariance matrix and the matrix of second moments coincide. We denote by
centered Gaussian state with the correlation matrix
. For states
, we have
by the maximum entropy principle.
The energy constraint reduces to
(We denote Sp trace of
-matrices as distinct from trace of operators on
.)
For a fixed correlation matrix
, we will study the
-
constrained capacity
With the Hamiltonian
the
energy-constrained classical capacity of observable
M is
We will be interested in the approximate position-momentum measurement (observable, POVM)
where
is centered Gaussian density operator with the covariance matrix
and
are the unitary displacement operators. Thus,
and the operator-valued density of POVM (
11) is
In quantum optics, some authors [
11,
19] call such measurements (noisy) general-dyne detections.
In what follows, we will consider
n independent copies of our bosonic system on the Hilbert space
We will supply all the quantities related to
th copy (
) with upper index
, and we will use tilde to denote quantities related to the whole collection on
n copies. Thus,
and
Lemma 2. Let be a real orthogonal matrix and U—the unitary operator on implementing the linear symplectic transformationso thatThen, for any state on , Proof. The covariance matrix
of
is block-diagonal,
; hence,
. Thus, we have
and taking into account (
12),
Therefore, for any state
on
, the output probability density of the measurement channel
corresponding to the input state
is
Hence, by using orthogonal invariance of the Lebesgue measure,
If
then
and taking
in the previous formula, we deduce
hence, (
13) follows. □
Lemma 3. Let M be the Gaussian measurement (11). For any state ρ with finite second moments, , where α is the covariance matrix of ρ. Proof. The proof follows the pattern of Lemma 1 from the paper of Wolf, Giedke, and Cirac [
12]. Without loss of generality, we can assume that
is centered. We have
where
with symplectic unitary
U in
corresponding to an orthogonal matrix
O as in Lemma 2, and
is the
th partial state of
Step (1) follows from the additivity (
4). Step (2) follows from Lemma 2, and step (3) follows from the superadditivity of
(Lemma 1). The final step of the proof,
uses ingeniously constructed
U from Reference [
12] and lower semicontinuity of
(Lemma 1). Namely,
and
U corresponds via (
12) to the following special orthogonal matrix
Every row of the
matrix
O, except the first one which has all the elements 1, has
elements equal to 1 and
elements equal to −1. Then, the quantum characteristic function of the states
is equal to
, where
is the quantum characteristic function of the state
This allows to apply Quantum Central Limit Theorem [
20] to show that
as
in a uniform way, implying (
16); see Reference [
12] for details. □
Theorem 1. The optimizing density operator ρ in (10) is the (centered) Gaussian density operator and, hence, Proof. Lemma 3 implies that, for any
with finite second moments,
, where
is the covariance matrix of
. On the other hand, by the maximum entropy principle,
. Hence, (
17) is maximized by a Gaussian density operator. □
Remark 1. The proof of Lemma 2 and, hence, of Theorem 1 can be extended to a general Gaussian observable M in the sense of References [16,21], defined via operator-valued characteristic function of the formwhere K is a scaling matrix, γ is the measurement noise covariance matrix, and . Then, the Fourier transform of the measurement probability density is equal to , and one can use this function to obtain generalization of the relation (14) for the measurement probability densities. The case (11) corresponds to the type 1 Gaussian observable [21] with . However, (19) also includes type 2 and 3 observables (noisy and noiseless multimode homodyning), in which case K is a projection onto an isotropic subspace of Z (i.e., one on which the symplectic form Δ
vanish.) Remark 2. Theorem 1 establishes Gaussianity of the average state of the optimal ensemble for a general Gaussian measurement channel. However, Gaussian average state can appear in a non-Gaussian ensemble. An immediate example is thermal state represented as a mixture of the Fock states with geometric distribution. Thus, Theorem 1 does not necessarily imply full Gaussianity of the optimal ensemble as formulated in the following conjecture.
Hypothesis of Gaussian Maximizers (HGM). Let M be an arbitrary Gaussian measurement channel. Then, there exists an optimal Gaussian ensemble for the convex closure of the output differential entropy (2) with Gaussian ρ and, hence, for the energy-constrained classical capacity (6) of the channel M. More explicitly, the ensemble consists of (properly squeezed) coherent states with the displacement parameter having Gaussian probability distribution. For Gaussian measurement channels of the type 1 (essentially of the form (
11), see Reference [
21] for complete classification) and Gaussian states
satisfying the “threshold condition”, we have
with the minimum attained on a squeezed coherent state, which implies the validity of the HGM and an efficient computation of
; see Reference [
5]. On the other hand, the problem remains open in the case where the “threshold condition” is violated, and in particular, for all Gaussian measurement channels of the type 2 (noisy homodyning), with the generic example of the energy-constrained approximate measurement of the position
subject to Gaussian noise (see Reference [
22], where the entanglement-assisted capacity of such a measurement was computed). In the following section, we will touch upon the HGM in this case for one mode system.
4. Gaussian Measurements in One Mode
Our framework in this section will be one bosonic mode described by the canonical position and momentum operators
. We recall that
are the unitary displacement operators.
We will be interested in the observable
where
is centered Gaussian density operator with the covariance matrix
Let
be a centered Gaussian density operator with the covariance matrix
The problem is, to compute
and, hence, the classical capacity
for the oscillator Hamiltonian
(as shown in the Appendix of Reference [
22], we can restrict to Gaussian states
with the diagonal covariance matrix in this case). The energy constraint (
9) takes the form
The measurement channel corresponding to POVM (
21) acts on the centered Gaussian state
by the formula
so that
In this expression,
c is a fixed constant depending on the normalization of the underlying measure
in (
1). It does not enter the information quantities which are differences of the two differential entropies.
Assuming validity of the HGM, we will optimize over ensembles of squeezed coherent states
where
is centered Gaussian state with correlation matrix
and the vector
has centered Gaussian distribution with covariance matrix
Then, the average state
of the ensemble is centered Gaussian
with the covariance matrix (
23), where
hence,
For this ensemble,
Then, the hypothetical value:
The derivative of the minimized expression vanishes for
Thus, depending on the position of this value with respect to the interval (
27), we obtain three possibilities):
Here, the column C corresponds to the case where the “threshold condition” holds, implying (
20). Then the full validity of the HGM in much more general multimode situation was established in Reference [
5]. All the quantities in this column, as well as the value of
in the central column of Table 2, were obtained in that paper as an example. On the other hand, the HGM remains open in the cases of mutually symmetric columns L and R (for the derivation of the quantities in column L of
Table 1 and
Table 2 see
Appendix A).
Maximizing
over
which satisfy the energy constraint (
24) (with the equality):
, we obtain
depending on the signal energy
E and the measurement noise variances
Table 2.
The values of the capacity .
Table 2.
The values of the capacity .
L: HGM open | C: HGM valid [5] | R: HGM open |
| | |
| | |
where we introduced the “energy threshold function”
In the gauge invariant case when
, the threshold condition amounts to
, which is fulfilled by definition, and the capacity formula gives the expression
equivalent to one obtained in Hall’s 1994 paper [
13].
Let us stress that, opposite to column C, the values of
in the L and R columns are hypothetic, conditional upon validity of the HGM. Looking into the left column, one can see that
and
do not depend at all on
Thus, we can let the variance of the momentum
p measurement noise
and, in fact, set
which is equivalent to the approximate measurement only of the position
q described by POVM
which belongs to type 2 according to the classification of Reference [
21]. In other words, one makes the “classical” measurement of the observable
with the quantum energy constraint
.
The measurement channel corresponding to POVM (
29) acts on the centered Gaussian state
by the formula
In this case, we have
which differ from the values in the case of finite
by the absence of the factor
under the logarithms, while the difference
and the capacity
have the same expressions as in that case (column L).
For
(sharp position measurement, type 3 of Reference [
21]), the HGM is valid with
This follows from the general upper bound (
Figure 1)
for
(Equation (
28) in Reference [
23]; also see Equation (5.39) in Reference [
10]).
5. The Dual Problem: Accessible Information
Let us sketch here
ensemble-observable duality [
1,
2,
4] (see Reference [
6] for details of mathematically rigorous description in the infinite dimensional case).
Let
be an ensemble,
a
finite measure and
an observable having operator density
with values in the algebra of bounded operators in
. The dual pair ensemble-observable
is defined by the relations
Then, the average states of both ensembles coincide
and the joint distribution of
is the same for both pairs
and
so that
Moreover,
where the supremum in the right-hand side is taken over all ensembles
satisfying the condition
. It can be shown (Reference [
6], Proposition 4), that the supremum in the lefthand side remains the same if it is taken over
all observables
M (not only of the special kind with the density we started with), and then it is called the
accessible information of the ensemble
. Thus,
Since the application of the duality to the pair
results in the initial pair
we also have
Coming to the case of bosonic mode, we fix the Gaussian state
and restrict to ensembles
with
Let
M be the measurement channel corresponding to POVM (
21). Then, according to formulas (
34), the dual ensemble
where
is the Gaussian probability density (
25) and
By using the formula for
, where
are Gaussian operators (see Reference [
24] and also Corollary in the Appendix of Reference [
25]), we obtain
where
and
Since
then, from second and third equations in (
39), we obtain
By denoting
, the density of this normal distribution, we can equivalently rewrite the ensemble
as
with the average state
Then, HGM is equivalent to the statement
where the values of
are given in
Table 1; however, they should be reexpressed in terms of the ensemble parameters
. In Reference [
25], we treated the case C in multimode situation, establishing that the optimal measurement is Gaussian, and described it. Here, we will discuss the case L (R is similar) and show that, for large
(including
), the HGM is equivalent to the following: the value of the accessible information
is attained on the sharp position measurement
(in fact, this refers to the whole domain L:
which, however, has rather cumbersome description in the new variables
, cf. Reference [
25]).
In the one mode case we are considering, the matrix
is given by (
23),
—by (
22), and
so that
Computations according to (
39) and (
40) give
But under the sharp position measurement
one has (in the formulas below,
means that
is Gaussian probability density with mean
m and variance
):
while
(note that
), and
which is identical to the expression in (
41).
In the case of the position measurement channel
M corresponding to POVM (
29) (
, we have
otherwise, the argument is essentially the same. Thus, we obtain that the HGM concerning
in case L is equivalent to the following:
The accessible information of a Gaussian ensemble whereis given by the expression (43) and attained on the sharp position measurement