Next Article in Journal
Active Fault Isolation for Multimode Fault Systems Based on a Set Separation Indicator
Next Article in Special Issue
Covariance-Matrix-Based Criteria for Network Entanglement
Previous Article in Journal
Induced Relaxation Enhances the Cardiorespiratory Dynamics in COVID-19 Survivors
Previous Article in Special Issue
Considerations on the Relativity of Quantum Irrealism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Information Transfer and the Uniform Measure over Probability Space

by
William K. Wootters
Department of Physics, Williams College, Williamstown, MA 01267, USA
Entropy 2023, 25(6), 875; https://doi.org/10.3390/e25060875
Submission received: 1 April 2023 / Revised: 29 May 2023 / Accepted: 29 May 2023 / Published: 30 May 2023
(This article belongs to the Special Issue Quantum Correlations, Contextuality, and Quantum Nonlocality)

Abstract

:
For a quantum system with a d-dimensional Hilbert space, suppose a pure state | ψ is subjected to a complete orthogonal measurement. The measurement effectively maps | ψ to a point ( p 1 , , p d ) in the appropriate probability simplex. It is a known fact—which depends crucially on the complex nature of the system’s Hilbert space—that if | ψ is distributed uniformly over the unit sphere, then the resulting ordered set ( p 1 , , p d ) is distributed uniformly over the probability simplex; that is, the resulting measure on the simplex is proportional to d p 1 d p d 1 . In this paper we ask whether there is some foundational significance to this uniform measure. In particular, we ask whether it is the optimal measure for the transmission of information from a preparation to a measurement in some suitably defined scenario. We identify a scenario in which this is indeed the case, but our results suggest that an underlying real-Hilbert-space structure would be needed to realize the optimization in a natural way.

1. Introduction

This paper brings together two questions pertaining to the structure of quantum theory. The questions concern (i) a quantum-mechanically natural measure on the probability simplex, and (ii) the optimization of the transfer of information. I now present the two questions.
(i)
A quantum-mechanically natural measure on the probability simplex
Given a quantum system with a d-dimensional Hilbert space, let us suppose its state is pure but otherwise unknown, by which we mean that it is distributed uniformly over the unit sphere in the Hilbert space. (The uniform distribution is the unique distribution that is invariant under all unitary transformations). Now suppose we perform on the system a complete orthogonal measurement, represented by an ordered set of d orthogonal projection operators ( Π 1 , , Π d ) . For any given pure state | ψ , the measurement maps the state to an ordered set of probabilities, namely, the probabilities of the outcomes of the measurement if the state is | ψ . The mapping is given by
| ψ ( p 1 , , p d ) = ( ψ | Π 1 | ψ , , ψ | Π d | ψ ) .
The ordered set ( p 1 , , p d ) resides in the ( d 1 ) -dimensional simplex Σ d defined by the conditions p j 0 and p 1 + + p d = 1 . Now, if our knowledge of | ψ is characterized by the uniform distribution over the unit sphere, what distribution, or measure, on the simplex Σ d characterizes our knowledge of the probabilities ( p 1 , , p d ) ? This question was answered in 1974 by Sýkora [1]. It turns out that the resulting measure is the uniform measure. That is, if we parameterize the points of the simplex by the first d 1 probabilities p 1 , , p d 1 , the induced measure d μ is given by
d μ = ( d 1 ) ! d p 1 d p d 1 ,
which is normalized in the sense that
Σ d d μ = 1 .
For example, if d = 2 , we start with a random state uniformly distributed over the surface of the Bloch sphere. Suppose we perform on this state the orthogonal measurement whose first and second outcomes are represented respectively by the north and south poles of the sphere. Then the probability p 1 of the first outcome is p 1 = 1 2 ( 1 + cos θ ) , where θ is the angle between the north pole and the point representing the state being measured. The area element on the sphere can be written as d ( cos θ ) d ϕ , and the range of the azimuthal angle ϕ is independent of θ ; so for our random state, the quantity cos θ is distributed uniformly between 1 and + 1 . It follows that p 1 is distributed uniformly over the interval 0 p 1 1 .
The uniformity of the distribution over the probability simplex is by no means a trivial or obviously foreordained result. If quantum theory had been based on a real Hilbert space instead of a complex Hilbert space, the uniform measure on the set of pure states would have produced a rather non-uniform measure on the probability simplex. Specifically, we would have obtained the normalized measure d ν given by [2]
d ν = Γ ( d / 2 ) π d / 2 · 1 p 1 p d d p 1 d p d 1 ,
which favors values of ( p 1 , , p n ) that are close to the edge of the simplex, where one or more of the p j s is small. The simplicity of the actual result (2) raises the following question: does the uniformity of this quantum-mechanically natural measure over the probability simplex hold any deep significance regarding the structure of quantum theory? Is there a fundamental principle implying that quantum theory must give rise to this measure? If so, then that principle would, for example, constitute a way of understanding why quantum theory is based on a complex rather than a real Hilbert space. (It might, for example, substitute for the principle of local tomography, which has frequently been used in reconstruction programs to rule out the real-vector-space theory [3,4]).
(ii)
The optimization of the transfer of information
Under unitary evolution in quantum theory, information is conveyed perfectly from the past to the future. A pure state remains pure, and a mixed state does not change its entropy. In this sense, nature at a fundamental level does not waste information.
However, our actual experience of the quantum world comes to us by way of measurements, for which the appearance of a definite outcome is not described unitarily. Rather, it is described probabilistically. (This is the case even if one regards the underlying dynamics to be unitary, as in the Everett interpretation). The outcome of a measurement does not give us complete information about the prior state of the measured system, and the outcome itself constitutes novel information that was not present earlier. These facts suggest the following question: in a description of quantum phenomena that includes probabilistic measurements, is there any sense in which information is conveyed optimally from the past to the future?

Relation between the Two Questions

I have posed versions of the information-optimization question in previous papers [2,5]. In the statements of the question I have considered, the problem boils down to determining the optimal form of the measure on the probability simplex Σ d induced by the uniform distribution over the pure states. The main upshot of my earlier work on this subject—which is consistent with the work of others [6,7,8,9]—has been that the measure that optimizes the transfer of information is the one given in Equation (4), not the one given in Equation (2). That is, if quantum theory were based on a real vector space rather than a complex vector space, there would indeed be a sense in which information is conveyed optimally. However, this intriguing result is spoiled by the fact that actual probability amplitudes are complex.
In the simplest example of optimal information transfer in a real-amplitude context, we imagine a beam of photons emerging from a linearly polarizing filter and then being subjected to a horizontal-vs-vertical polarization measurement. (This counts as a real-Hilbert-space example because the linear polarizations can be represented in such a space). A person stationed at the measuring site can gain information about, say, the probability p v of the vertical outcome by counting how many photons yield the vertical and horizontal outcomes. However, as long as the number of photons is finite, the amount of information gained in this way is limited. Not surprisingly, the amount of information gained depends on the prior measure one assumes for the value of p v , and it turns out that in a suitable limit as the number of photons goes to infinity, the information is maximized if this prior measure is taken to be
d ν = 1 π p v ( 1 p v ) d p v ,
which is the special case of Equation (4) appropriate for d = 2 .
It is not hard to see that the measure (5) is the one the experimenter would naturally use if she had no initial information about the orientation of the polarizing filter from which the beam of photons emerged. The probability of the vertical outcome is p v = cos 2 θ , where θ is the angle between the filter’s favored axis and the vertical axis. In the real-amplitude theory, one would naturally assume that θ is uniformly distributed over its range, which we can take to run from 0 to π . The normalized prior measure would then be
1 π d θ = 2 π d p v d θ 1 d p v = 2 π · 1 | 2 cos θ sin θ | d p v = 1 π p v ( 1 p v ) d p v ,
in agreement with Equation (5). (The factor of 2 in the first step appears because there are generically two values of θ in the interval 0 θ π that yield a given value of p v ). So this is an example, in the real-amplitude context, in which the natural prior measure is also the one that maximizes the amount of information conveyed.
Our first aim in this paper is to identify a different optimal-information-transfer problem, ideally not too complicated, whose solution is not the measure given by Equation (4) but rather the one given by Equation (2), that is, the one we associate with standard, complex-amplitude quantum theory. This aim is an unusual one in a certain respect. Normally, one is given a problem and one wants to find the solution. Here, we are given a solution—the uniform measure over the probability simplex—and we are looking for a problem that has this measure as its solution. This backward-sounding procedure, however, is implicit in the question we asked earlier: in a description of quantum phenomena that includes probabilistic measurements, is there any sense in which information is conveyed optimally? We are given quantum theory, along with the associated measure (2) over the probability simplex, and we are looking for a problem that will serve to identify an interesting feature of this measure and hence an interesting feature of quantum theory.
We will indeed find an information-optimization problem that is solved by the measure of Equation (2). We present this problem in Section 2. In Section 3, we solve the problem and show that it leads us to the uniform measure, as promised. Then in Section 4, we ask whether we can interpret the problem within the framework of quantum theory. That is, we ask whether there is a quantum scenario in which we can see that information is conveyed optimally, without having to restrict ourselves artificially to quantum states with only real amplitudes as in the linear polarization example. We will find that there is indeed such a scenario, but it is not as straightforward as one might have liked. Finally in Section 5 we ask what insights this exercise gives us into the structure of quantum theory.
The problem we are about to lay out in Section 2 will be couched in terms of Shannon information—specifically the combination of Shannon entropies that defines the mutual information—and one might wonder whether Shannon information is the best way to quantify information in the setting we are considering. Shannon information is intimately related to asymptotic coding efficiency, which will not figure into our problem. However, Shannon information can also be thought of as the unique measure of information that satisfies certain natural assumptions having nothing directly to do with coding [10]. It is for this latter reason that we feel the measure is justified for our problem.
It is worth noting that there are close connections between mutual information, which we will be using here, and Fisher information, which is widely used in statistics [6,7,8,9]. For the discrete probability distribution ( p 1 , p 2 , , p d ) , regarded as parameterized by the first d 1 probabilities, the determinant of the Fisher information matrix J is det J = 1 / ( p 1 p 2 p d ) . It is no accident that the distribution (5) that maximizes the experimenter’s expected gain in information in our polarization example is proportional to the square root of det J (for the case d = 2 , in which case J is simply a scalar function). In general, the distribution proportional to det J is known as the Jeffreys prior [11]. As a prior distribution for Bayesian inference, it was originally motivated by invariance arguments, but it can also be shown to maximize information in certain settings [6,9]. Though we will not be explicitly concerned with Fisher information in what follows, some of the arguments in Section 3 below are similar to what can be found in these earlier papers.
Note that in identifying the uniform measure (2) with standard, complex-amplitude quantum theory, we are placing pure states and orthogonal measurements at the center of our argument: we obtain the quantum mechanical measure by starting with the unitarily invariant measure over pure states, and the probabilities in question are the probabilities of the outcomes of a complete orthogonal measurement. The centrality of pure states and complete orthogonal measurements in this paper reflects my own view that these structures are more fundamental than mixed states and generalized measurements. Mixed states and general POVM measurements can of course be obtained from their purer counterparts by embedding the system of interest in a larger system and ultimately ignoring part of this larger system.

2. The Problem

We now state our information-optimization problem. Though we are ultimately interested in gaining insight into quantum theory, the problem itself is not quantum mechanical. It is to be understood as an abstract communication problem.
The scenario begins with Alice, who sends a beam of N information-bearing particles to Bob. Bob has a measuring device that separates this beam into d subbeams, recording, for each particle, the subbeam into which it was sent (the record is the outcome of Bob’s measurement). The d subbeams are then sent on to participants named Carol-1, , Carol-d, each of whom makes a binary measurement on each of the particles she receives. (In Section 4 and Section 5, we will discuss the significance of these binary measurements. For now, we simply take them as part of the statement of the problem). All the measuring devices in this problem—Bob’s device and the device of each of the Carols—generate their outputs probabilistically, and all the information that is conveyed is encoded in the probabilities of the outcomes. We can imagine Alice having control over the probabilities of Bob’s outcomes, so that Alice is conveying information to Bob. If we wish, we can imagine Alice also having control over the probabilities of Carol-j’s outcomes. Alternatively, though, we could imagine that it is Bob who controls Carol-j’s probabilities; the statement of the problem is independent of the source of the information.
Let p j be the probability of the jth outcome of Bob’s measurement, and let n j be the number of times this outcome occurs when Bob measures the N particles arriving from Alice. Bob’s goal is to estimate p = ( p 1 , , p n ) by observing n = ( n 1 , , n d ) , but again, because N is finite, Bob’s information about p is limited. We quantify this information as the mutual information between p and n :
I B = I ( p : n ) ,
which is the average amount of information Bob gains about p upon observing n . (In our notation, we do not distinguish between a random variable and a particular value that the random variable might take, relying instead on the context to make clear which meaning is intended. In the expression I ( p : n ) , the symbols p and n are understood to name random variables). This mutual information depends on Bob’s prior measure on the probability simplex Σ d , which we write as ρ ( p ) d p , where d p = d p 1 d p d 1 .
Let q j and 1 q j be the probabilities of the two possible outcomes of Carol-j’s measurement, and let m j and n j m j be the numbers of occurrences of these outcomes when Carol-j measures the n j particles sent to her by Bob. The amount of information obtained by Carol-j is quantified by
I C j = I ( q j : n j , m j ) .
This mutual information depends on the function ρ ( p ) , and it also depends on the function σ ( q j ) , which characterizes Carol-j’s prior measure σ ( q j ) d q j over her one-dimensional probability simplex Σ 2 . (We assume that the same function σ applies to each of the Carols).
We are interested in maximizing the sum of the information gained by Bob and the information gained by all the Carols. So we construct the quantity
I = I B + j I C j .
This quantity depends on N, and we would like to take the limit as N goes to infinity, but we need to take into account that I itself approaches infinity in this limit, growing as ( d 1 2 ) ln N . So we aim to maximize the quantity I ˜ , defined by
I ˜ = lim N I ( d 1 2 ) ln N .
Our problem, then, is to find the functions ρ ( p ) and σ ( q ) that maximize I ˜ .
In order for the argument in the following section to be mathematically sound, we need to assume that the functions ρ and σ are reasonably well behaved. We assume that both functions are differentiable in the interiors of their respective simplices. We also assume that the derivatives do not diverge too rapidly near the edges of the simplices. Specifically, we assume there are positive numbers α and β , with β < 2 , such that | d σ / d q | < α / [ q ( 1 q ) ] β and | ρ / p j | < α / ( p 1 p 2 p d ) β for j = 1 , , d 1 . [12]. (One can imagine an alternative formulation of the problem that removes the temptation to introduce such extra assumptions: let ρ N and σ N be the optimal priors when there are exactly N particles. Then find suitable limiting measures ρ and σ as N approaches infinity. I expect the results would be the same as in the present paper, but I will have to leave that problem for future work).

3. Solving the Problem

Here we give the argument leading to the conclusion that the optimal function ρ for our problem is the constant function ρ ( p ) = ( d 1 ) ! . We will also find the optimal function σ .
We begin by estimating the mutual information I B = I ( p : n ) ; the error caused by using our estimate rather than the exact value will become negligible as N approaches infinity. First, although we are thinking of I ( p : n ) as the average amount of information Bob gains about p by observing n , the mutual information is symmetric, so that we can compute it as the average amount of information one would gain about n upon learning the value of p . This interpretation is the one suggested by the formula
I B = I ( p : n ) = I ( n : p ) = H ( n ) H ( n | p ) .
Here H ( n ) is the entropy of the prior probability distribution of n , which we label P ( n ) , and H ( n | p ) is the conditional entropy of n given p . The exact formula for H ( n ) is
H ( n ) = n P ( n ) ln P ( n ) .
when N is very large, the vector n is likely to be very close to N p ; so the function P ( n ) in that case is largely determined by the function ρ ( p ) and can be well approximated by
P ( n ) 1 N d 1 ρ ( n / N ) .
Thus we can approximate H ( n ) as
H ( n ) ( d 1 ) ln N Σ d ρ ( p ) ln ρ ( p ) d p .
Meanwhile, the formula for H ( n | p ) is
H ( n | p ) = Σ d n P ( n | p ) ln P ( n | p ) ρ ( p ) d p ,
where P ( n | p ) is the probability of the outcome n if the probabilities are those in p . This probability is given by a multinomial distribution, whose entropy, for large N, is well approximated as [13]
n P ( n | p ) ln P ( n | p ) d 1 2 ln ( 2 π e N ) + 1 2 ln ( p 1 p d ) .
We can therefore write
H ( n | p ) d 1 2 ln ( 2 π e N ) + 1 2 Σ d ρ ( p ) ln ( p 1 p d ) d p .
Combining Equations (14) and (17), we have
I B = I ( p : n ) d 1 2 ln N 2 π e Σ d ρ ( p ) ln ρ ( p ) d p 1 2 Σ d ρ ( p ) ln ( p 1 p d ) d p .
Before we continue with our calculation, we pause to note that if we were simply to maximize I B , then for very large N we would be maximizing the right-hand side of Equation (18). One can show that the distribution ρ ( p ) that achieves the maximum in this case is the one given by Equation (4), that is, the distribution associated with a real Hilbert space [2]. So, the fact that we include in Equation (9) the information transmitted by Bob to the Carols will make a crucial difference in determining what function ρ ( p ) emerges as optimal.
Continuing now with our calculation, we compute I C j , which we can write as
I C j = I ( q j : n j , m j ) = H ( n j , m j ) H ( n j , m j | q j ) .
The entropies on the right-hand side can be expanded via the additivity of Shannon entropy.
H ( n j , m j ) = H ( n j ) + H ( m j | n j ) H ( n j , m j | q j ) = H ( n j | q j ) + H ( m j | n j , q j ) .
Now, the number of particles, n j , entering Carol-j’s apparatus has nothing to do with the probability q j of the first outcome of Carol-j’s measurement. So H ( n j | q j ) is equal to H ( n j ) , and those two terms cancel when we carry out the subtraction in Equation (19). Thus we have
I C j = H ( m j | n j ) H ( m j | n j , q j ) .
This equation is similar in form to the Equation (11) defining I B , but there is now an additional conditioning (on n j ). We use this similarity to find an approximate formula for I C j analogous to our Equation (18) for I B . The role of N is now played by n j , and the role of ρ ( p ) is now played by σ ( q j ) . The differences are these: (i) whereas in the calculation of I B the number of incoming particles (N) was fixed, now we need to average over n j , and (ii) Carol-j’s measurement has only two outcomes, so we replace d with 2. This gives us
I C j 1 2 ln n j 2 π e 0 1 σ ( q j ) ln σ ( q j ) d q j 1 2 0 1 σ ( q j ) ln [ q j ( 1 q j ) ] d q j ,
where the angular brackets indicate an average over the prior distribution of n j . Of the three terms inside the brackets, only the first term will need to be averaged in this way. Using the fact that for large enough N, the value of n j will likely be very close to N p j , we can write
ln n j Σ d ρ ( p ) ln ( N p j ) d p = ln N + Σ d ρ ( p ) ln ( p j ) d p .
So we have
I C j 1 2 ln N 2 π e + 1 2 Σ d ρ ( p ) ln ( p j ) d p 0 1 σ ( q j ) ln σ ( q j ) d q j 1 2 0 1 σ ( q j ) ln [ q j ( 1 q j ) ] d q j .
Summing over all the Carols, we obtain
j = 1 d I C j d 2 ln N 2 π e + 1 2 Σ d ρ ( p ) ln ( p 1 p d ) d p d 0 1 σ ( q ) ln σ ( q ) d q d 2 0 1 σ ( q ) ln [ q ( 1 q ) ] d q .
Combining this result with Equation (18) and the definition (9) gives us
I d 1 2 ln N 2 π e Σ d ρ ( p ) ln ρ ( p ) d p d 0 1 σ ( q ) ln σ ( q ) d q d 2 0 1 σ ( q ) ln [ q ( 1 q ) ] d q .
Finally, we subtract ( d 1 2 ) ln N to obtain I ˜ as defined in Equation (10):
I ˜ = d 1 2 ln 1 2 π e Σ d ρ ( p ) ln ρ ( p ) d p d 0 1 σ ( q ) ln σ ( q ) d q d 2 0 1 σ ( q ) ln [ q ( 1 q ) ] d q .
This last expression is exact, because the limit N that is part of the definition of I ˜ eliminates the errors we have made in our approximations.
Our expression for I ˜ depends on ρ only in the second term. So the function ρ ( p ) that maximizes I ˜ is the one that maximizes
Σ d ρ ( p ) ln ρ ( p ) d p ,
subject, of course, to the condition that Σ d ρ ( p ) d p = 1 . The solution to this problem is well known: the integral in Equation (28) is uniquely maximized by the uniform distribution over the probability simplex. (This follows from the strict concavity of the function x ln x for x > 0 ). Recall that this is the outcome we were aiming for. So we have indeed identified an information-optimization problem that has the uniform distribution as its solution.
We can also easily find the optimal function σ ( q ) . According to Equation (27), it is the function that maximizes the quantity
0 1 σ ln σ + 1 2 σ ln [ q ( 1 q ) ] d q
subject to the condition 0 1 σ ( q ) d q = 1 . We find this function by introducing a Lagrange multiplier λ and considering the quantity
S = 0 1 σ ln σ + 1 2 σ ln [ q ( 1 q ) ] + λ σ d q .
Computing the variation δ S and setting it equal to zero, we find that
σ ( q ) = A q ( 1 q )
for some constant A. We normalize σ by setting A equal to 1 / π :
σ ( q ) = 1 π q ( 1 q ) ,
thereby arriving at a distribution we have seen before, in Equation (5). So, while the optimal ρ ( p ) is the uniform distribution characteristic of complex-amplitude quantum theory, the optimal σ ( q ) is the one arising naturally in a two-dimension real-amplitude version of quantum theory.
It is worth highlighting the crucial cancellation that made the uniform distribution the optimal choice for ρ ( p ) : the integral over p in our expression (25) for j I C j cancelled the last term in our expression (18) for I B . It was this term that favored a distribution more heavily weighted toward the edges of the probability simplex, leading to the distribution (4).
Let us try to get an intuitive understanding of this cancellation. The reason the term 1 2 ρ ( p ) ln ( p 1 p d ) d p appears in our expression for I B is that the statistical fluctuations are smaller near the edges of the probability simplex than near the center, in the sense that the observed frequencies of occurrence of the outcomes will typically deviate less from the outcomes’ probabilities. For the purpose of sending information to Bob, this fact makes it desirable to favor the parts of the simplex that are close to the edges. On the other hand, the amount of information Carol-j obtains includes a term proportional to ln n j , reflecting the fact that she gains more information if Bob sends her more particles to measure. The term + 1 2 ρ ( p ) ln ( p 1 p d ) d p appearing in Equation (25) results from a kind of compromise among all the Carols; it favors values of p such that Bob sends a respectable number of particles to each of the Carols, that is, values lying close to the center of the probability simplex. It is the competition between these two tendencies—that is, (i) the edge preference that is beneficial for communicating to Bob and (ii) the center preference that is beneficial for communicating to the Carols—that finally leads to the optimality of the uniform distribution.

4. Does This Information Optimization Appear in Quantum Theory?

In looking for a quantum mechanical realization of the information-optimizing scenario we imagined in Section 2 and Section 3, let us begin by assuming Alice is sending Bob a beam of d-dimensional particles, each prepared in the same pure state | ψ A . Bob performs on these particles a complete orthogonal measurement ( Π 1 , , Π d ) , the probabilities of the outcomes being
p = ( p 1 , , p d ) = ( ψ A | Π 1 | ψ A , , ψ A | Π d | ψ A ) .
So far, so good—but what, then, are the Carols doing? Here we consider two possibilities.
1.
Carol-j tests for another pure state
Let us make the standard assumption that if a particle yields the jth outcome of Bob’s measurement—and is therefore sent to Carol-j—it has been collapsed into the pure state picked out by Π j . Suppose that Carol-j’s measurement is a test for some pure state | ψ C j unrelated to Π j , and q j is the probability that a particle sent by Bob will pass the test. Then
q j = ψ C j | Π j | ψ C j .
This is a perfectly reasonable quantum mechanical scenario that one can easily imagine setting up in a lab.
For this experiment, there are quantum-mechanically natural prior measures that Bob and the Carols could use. Are they the ones that maximize the quantity I ˜ ? For Bob, the answer is yes: assuming a uniform distribution of Alice’s state | ψ over the sphere of pure states, we know that Bob’s prior measure for the ordered set p is the uniform measure over Σ d , which, as we have seen, is the optimal measure for maximizing I ˜ .
However, for the Carols, the answer is no. If we assume Carol-j is initially completely ignorant of the state picked out by Bob’s projector Π j , then if she were performing a complete orthogonal measurement, her prior measure on the ( d 1 ) -dimensional probability simplex would be the uniform measure. Since she is not performing a complete measurement but is only distinguishing between | ψ C j and the ( d 1 ) -dimensional subspace orthogonal to | ψ C j , the probability distribution σ ( q j ) she should use is the marginal distribution obtained by integrating the uniform distribution over the probabilities associated with the potential outcomes that she is not distinguishing. One finds that the result of this marginalization is
σ ( q j ) = ( d 1 ) ( 1 q j ) d 2 .
(Smaller values of q j are favored—for values of d greater than 2—because they leave a larger volume of Σ d available for the other probabilities. The exponent d 2 appears because, with q j fixed, the other probabilities occupy a space of just d 2 dimensions). This is very different from the information-optimizing distribution of Equation (32). So this example does not constitute a realization of our information-optimizing scenario, even though Bob’s ρ ( p ) is the one that maximizes I ˜ .
2.
Carol-j measures a “phase rebit”
In the preceding example, we were imagining Carol acquiring information that was created by Bob’s measurement. We now consider the possibility that she is acquiring information sent by Alice but left unmeasured by Bob.
For each of the pure states distinguished by Bob’s measurement, let us choose a definite state vector | ψ B j to represent this state, so that Π j = | ψ B j ψ B j | . (Our choice here is simply a matter of choosing a definite overall phase for the vector). Bob’s measurement is sensitive to part of the information contained in | ψ A but not to all of it. Specifically, it is sensitive to the magnitudes of the complex numbers ψ B j | ψ A —the squares of these magnitudes are the probabilities of Bob’s outcomes—but it is not sensitive to the phases of these numbers.
Let us now pretend, contrary to the actual laws of quantum theory, that these phases are preserved in the particles Bob sends to the Carols and that the Carols can gain access to them. Focusing on Carol-j, let us represent the phase of ψ B j | ψ A as a unit vector in the complex plane. Moreover, let us treat this complex plane as the state space of a rebit, that is, a binary quantum object in real-vector-space quantum theory. Then Carol-j can make an orthogonal measurement that distinguishes, say, the real axis from the imaginary axis. The probabilities of Carol-j’s outcomes would then be cos 2 ϕ j and sin 2 ϕ j , where ϕ j is the phase of ψ B j | ψ A . If, as we have been assuming, Alice’s state vector | ψ A is initially uniformly distributed over the unit sphere in Hilbert space, then ϕ j is uniformly distributed over the interval 0 ϕ j < 2 π . By the same argument we used in our polarization example in the Introduction, this makes Carol-j’s prior distribution σ ( q j ) equal to
σ ( q j ) = 1 π q j ( 1 q j ) ,
which is the distribution that maximizes I ˜ .
So this example agrees perfectly with our information-optimizing scenario. It suffers, though, from the flaw that it cannot be carried out in real life.

An Actual Quantum Mechanical Experiment

Of course phases can be measured in the context of an interference experiment. We now show how the preceding, impossible scenario can be turned into an experiment allowed by quantum theory.
We begin by separating the vector | ψ A into a “magnitude part” and a “phase part.” Using the notation introduced in the preceding example, we write
ψ B j | ψ A = e i ϕ j | ψ B j | ψ A | .
Let us now introduce a state vector | ψ ˜ A and a unitary transformation Φ defined as follows:
| ψ ˜ A = j | ψ B j | ψ B j | ψ A |
and
Φ = k e i ϕ k | ψ B k ψ B k | .
One can check that
| ψ A = Φ | ψ ˜ A .
So Alice could prepare the state | ψ A by first preparing the state | ψ ˜ A and then applying the transformation Φ .
Our interference experiment will be set up as a modified Mach–Zehnder interferometer. In this experiment, in addition to the d-dimensional quantum system to be prepared by Alice, there is a binary degree of freedom consisting of two possible paths that the particles might take. The path we label | 0 consists of the particles’ initial path together with the continuation of this path via transmissions (not reflections) through two symmetric beam splitters. The path we label | 1 is the path that begins with reflection at the first beam splitter and continues with transmission through the second. The layout is illustrated in Figure 1 for the case in which the main quantum system is photon polarization, for which d has the value 2.
Now, instead of having Alice prepare the state | ψ A at the outset, we have her prepare the state | ψ ˜ A initially and then apply the unitary transformation Φ only in path | 0 just after the first beam splitter. Thus the state | ψ A is prepared only in that path. Then, in each of the two paths emerging from the second beam splitter, Bob performs his complete orthogonal measurement. Since he performs this measurement in two places, there are a total of 2 d possible outcomes of the experiment for each incoming particle.
To compute the probabilities of the outcomes, we introduce a unitary transformation V representing the action of each beam splitter:
V = 1 2 1 i i 1 .
We now keep track of the effects of the three successive transformations—the first beam splitter, the application of Φ in path | 0 , and the second beam splitter—to obtain the state | Ψ that will be measured by Bob. Here we take the main quantum system to be the first factor in each tensor product and the binary path variable to be the second factor.
| Ψ = ( I V ) ( Φ | 0 0 | + I | 1 1 | ) ( I V ) ( | ψ ˜ A | 0 ) .
Plugging in the definitions of | ψ ˜ A , V, and Φ , we find that
| Ψ = 1 2 j | ψ B j | ψ A | · | ψ B j e i ϕ j 1 | 0 + i e i ϕ j + 1 | 1 .
From this we can compute the probabilities of the 2 d possible outcomes of the experiment.
( probability of outcome j in path | 0 ) = | ψ B j | ψ A | 2 sin 2 ( ϕ j / 2 ) ( probability of outcome j in path | 1 ) = | ψ B j | ψ A | 2 cos 2 ( ϕ j / 2 )
Let us think of the values | ψ B j | ψ A | 2 as the probabilities of the outcomes of Bob’s measurement, and the values sin 2 ( ϕ j / 2 ) and cos 2 ( ϕ j / 2 ) as the probabilities of the outcomes of Carol-j’s measurement. Then the probabilities in Equation (44), while not exactly the ones we encountered in our illegal example of the “phase rebits,” differ from those probabilities only by the factor 1 / 2 inside the sine and cosine. If ϕ j is uniformly distributed, as we are assuming, then sin 2 ϕ j and sin 2 ( ϕ j / 2 ) both inherit the same distribution over the interval [ 0 , 1 ] , namely, the one given by Equation (36). So in this experiment, Bob’s ρ ( p ) and Carol-j’s σ ( q j ) both turn out to be the functions that maximize the quantity I ˜ .
Let us identify the components of the experiment associated with the roles of Alice, Bob, and the Carols. As we have already suggested, Alice controls the device that prepares the initial state | ψ ˜ A and the device that implements the unitary transformation Φ . To identify the Bob and Carol components, it helps to imagine each of the d-outcome measurements as consisting of a splitting device S k , which merely splits the beam into d subbeams, and a set of d detectors D j k . Here the index k takes the values 0 and 1 and labels the path on which the splitting device and the detectors appear. Then we can locate our participants’ portions of the apparatus as follows:
Alice:Devices associated with | ψ ˜ A and Φ .
Bob: S 0 , S 1 , and all the detectors.
Carol-j:The two beam splitters, plus detectors D j 0 and D j 1 .
Note that our participants are not well separated from each other. Two of the overlaps are particularly noteworthy: (i) Alice must know in advance what measurement Bob will make, because her unitary transformation Φ needs to be diagonal in the preferred basis of that measurement; (ii) in order to make their phase measurements, the Carols must interrupt Alice’s preparation by inserting their initial beam splitter. The latter of these two overlaps is what prevents us from interpreting the experiment as a clean preparation followed by a measurement.
Thus we have imagined a genuine quantum mechanical experiment in which information is transmitted optimally from the sender (Alice) to the receivers (Bob and the Carols), but the experiment involves a basis-specific separation between the “magnitude” and “phase” components of the prepared state, and the receivers need to intervene in the middle of the preparation in order to be able, later, to extract certain features of the transmitted information.

5. Discussion

It is an intriguing fact that, according to the quantum theory of a system with a finite-dimensional Hilbert space, a uniform distribution over the unit sphere of pure states induces the uniform distribution over the probability simplex. This fact could conceivably help us understand why quantum theory has the structure it has, if we can attribute to the latter distribution some foundational significance. In this paper, we have explored the possibility that the uniform distribution over the probability simplex might have the significance of allowing information to be transmitted optimally, in some appropriate understanding of that phrase.
We have identified a specific information-optimization problem for which the uniform distribution over the simplex does emerge as part of the solution. The problem begins with a simple communication scenario between a sender, Alice, and a receiver, Bob. It continues by assuming that for each outcome of Bob’s measurement, there is a further binary measurement, and the information provided by these binary measurements is included in the tally. The solution of this problem consists of two distributions: a distribution ρ ( p ) over Σ d , where d is the number of outcomes of Bob’s measurement, and a distribution σ ( q ) over the one-dimensional simplex Σ 2 . The first of these is indeed the uniform distribution. The other is the distribution given in Equation (32).
It is interesting that each of these distributions, taken separately, appears naturally in quantum theory. The uniform distribution has, of course, been the focus of this paper. Meanwhile, the distribution (32), in addition to arising in the linear polarization problem as discussed in the Introduction, arises also in the simplest possible interference experiment. Consider a basic Mach–Zehnder interferometer with a simple phase shifter in one of the two paths. (This is a special case of the interference experiment described in Section 4, with d having the value 1). Let q and 1 q be the probabilities of the two possible paths emerging from the second beam splitter. If the phase ϕ that is added by the phase shifter is uniformly distributed over the interval 0 ϕ < 2 π , then q is distributed according to the function
σ ( q ) = 1 π q ( 1 q ) ,
which is the optimal σ ( q ) for our communication problem.
The interference experiment we have presented in Section 4 brings these two optimal functions, ρ ( p ) and σ ( q ) , together by embedding a basic prepare-and-measure experiment into a Mach–Zehnder interferometer. While this experiment does indeed exhibit optimal information transfer from the sender to the receivers, it does so at the cost of requiring the receivers to intervene in the middle of the preparation procedure. So the experiment seems a bit artificial, just as the polarization experiment we described in the Introduction feels artificial in that it requires ignoring the possibility of circular or elliptical polarization.
Essentially, what the additional binary measurements in our problem are doing, in the quantum context, is to turn the real and imaginary parts of each quantum amplitude into two dimensions of a real vector space. The interference experiment of Section 4 is basically a scheme for making standard quantum theory look like real-vector-space quantum theory—the interference between the two paths provides a way of accessing those real and imaginary components. As we have noted in the Introduction, it was already known that in real-vector-space quantum theory, the transfer of information from Alice to Bob (with no extra measurements after Bob’s measurement) is optimal. So we are effectively observing the same result again in this paper. What is different here is that our framing of the communication problem, which separates Bob from the Carols, allows us to see the how the uniform distribution over Σ d , which is characteristic of standard, complex-amplitude quantum theory, emerges from the extremization.
Still, though, this line of thinking—that is, asking about optimal information transfer under the assumption that the state being transmitted is pure—continues to single out the real-vector-space theory as special in a certain sense. It is in this theory that a complete orthogonal measurement accesses every dimension of the manifold of pure states (as opposed to the complex theory, in which such a measurement misses the phase dimensions), and it does so in a way that allows information to be conveyed optimally. One can imagine reconstructing quantum theory by starting with an information-optimizing theory but then imposing a rule that forbids measurement of the “phase rebits.” That is to say, one could start with Bob and the Carols, but then assume that no one can actually carry out the Carols’ measurements. Indeed, something along these lines can be seen in Goyal’s 2010 reconstruction of quantum theory, in which one assumes the existence of a binary degree of freedom that is an essential part of the theory but is inaccessible [14].
Very recently, theoretical and experimental advances have shown quite clearly that, within the standard framework of quantum axioms, real-vector-space quantum theory can be ruled out [15,16,17,18], and yet we have known since the work of Stueckelberg that it is possible to express quantum theory purely in terms of a real Hilbert space [19]. These two facts are not inconsistent, because Stueckelberg’s formulation does not respect the usual tensor-product rule for describing composite systems. The work presented in the present paper gives us reason not to set real-vector-space theories entirely aside.

Funding

This research received no external funding.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Sýkora, S. Quantum theory and Bayesian inference problems. J. Stat. Phys. 1974, 11, 17–27. [Google Scholar] [CrossRef]
  2. Wootters, W.K. Communicating through probabilities: Does quantum theory optimize the transfer of information? Entropy 2013, 15, 3130–3147. [Google Scholar] [CrossRef]
  3. Chiribella, G.; D’Ariano, G.M.; Perinotti, P. Informational derivation of quantum theory. Phys. Rev. A 2011, 84, 012311. [Google Scholar] [CrossRef]
  4. Hardy, L. Quantum theory from five reasonable axioms. arXiv 2021, arXiv:quant-ph/0101012. [Google Scholar]
  5. Wootters, W.K. Optimal Information Transfer and Real-Vector-Space Quantum Theory. In Quantum Theory: Informational Foundations and Foils; Spekkens, R.W., Chiribella, D., Eds.; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  6. Bernardo, J.M. Reference Posterior Distributions for Bayesian Inference. J. R. Stat. Soc. Ser. 1979, 41, 113–147. [Google Scholar] [CrossRef]
  7. Clarke, B.S.; Barron, A.R. Information theoretic asymptotics of Bayes methods. IEEE Trans. Inf. Theory 1990, 36, 453–471. [Google Scholar] [CrossRef]
  8. Rissanen, J. Fisher information and stochastic complexity. IEEE Trans. Inf. Theory 1996, 42, 40–47. [Google Scholar] [CrossRef]
  9. Brunel, N.; Nadal, J.-P. Mutual information, Fisher information, and Population Coding. Neural Comput. 1998, 10, 1731–1757. [Google Scholar] [CrossRef]
  10. Khinchin, A. Mathematical Foundations of Information Theory; Courier: San Francisco, CA, USA, 1957. [Google Scholar]
  11. Jeffreys, H. An invariant form for the prior probability in estimation problems. Proc. R. Soc. London Ser. Math. Phys. Sci. 1946, 186, 453–461. [Google Scholar]
  12. Wootters, W.K. The Acquisition of Information from Quantum Measurements. Ph.D. Thesis, The University of Texas at Austin, Austin, TX, USA, 1980. [Google Scholar]
  13. Kaji, Y. Converging Bounds on the Entropy of Multinomial Distributions. In Proceedings of the 2016 International Symposium on Information Theory, Monterey, CA, USA, 30 October–2 November 2016; pp. 433–437. [Google Scholar]
  14. Goyal, P. From information geometry to quantum theory. New J. Phys. 2010, 12, 023012. [Google Scholar] [CrossRef]
  15. Renou, M.D.; Trillo, D.; Weilenmann, M.; Le, T.P.; Tavakoli, A.; Gisin, N.; Acin, A.; Navascués, M. Quantum theory based on real numbers can be experimentally falsified. Nature 2021, 600, 625–629. [Google Scholar] [CrossRef] [PubMed]
  16. Chen, M.-C.; Wang, C.; Liu, F.-M.; Wang, J.-W.; Ying, C.; Shang, Z.-X.; Wu, Y.; Gong, M.; Deng, H.; Liang, F.-T.; et al. Ruling out real-valued standard formalism of quantum theory. Phys. Rev. Lett. 2022, 128, 040403. [Google Scholar] [CrossRef] [PubMed]
  17. Li, Z.-D.; Mao, Y.-L.; Weilenmann, M.; Tavakoli, A.; Chen, H.; Feng, L.; Yang, S.-J.; Renou, M.-O.; Trillo, D.; Le, T.P.; et al. Testing real quantum theory in an optical quantum network. Phys. Rev. Lett. 2022, 128, 040402. [Google Scholar] [CrossRef] [PubMed]
  18. Wu, D.; Jiang, Y.-F.; Gu, X.-M.; Huang, L.; Bai, B.; Sun, Q.-C.; Gong, S.-Q.; Mao, Y.; Zhong, H.-S.; Chen, M.-C.; et al. Experimental refutation of real-valued quantum mechanics under strict locality conditions. Phys. Rev. Lett. 2022, 129, 140401. [Google Scholar] [CrossRef] [PubMed]
  19. Stueckelberg, E.C.G. Quantum theory in real Hilbert space. Helv. Phys. Acta 1960, 33, 727–752. [Google Scholar]
Figure 1. An interference experiment in which information is transferred optimally from sender to receivers. The sender (Alice) controls the linearly polarizing filter and the birefringent medium; the preferred axes of this medium are horizontal and vertical. The receivers (Bob, Carol-1, and Carol-2) have set up the beam splitters (BS), the polarizing beam splitters (PBS), and the detectors. The polarizing beam splitters have the same preferred axes as the birefringent medium.
Figure 1. An interference experiment in which information is transferred optimally from sender to receivers. The sender (Alice) controls the linearly polarizing filter and the birefringent medium; the preferred axes of this medium are horizontal and vertical. The receivers (Bob, Carol-1, and Carol-2) have set up the beam splitters (BS), the polarizing beam splitters (PBS), and the detectors. The polarizing beam splitters have the same preferred axes as the birefringent medium.
Entropy 25 00875 g001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wootters, W.K. Optimal Information Transfer and the Uniform Measure over Probability Space. Entropy 2023, 25, 875. https://doi.org/10.3390/e25060875

AMA Style

Wootters WK. Optimal Information Transfer and the Uniform Measure over Probability Space. Entropy. 2023; 25(6):875. https://doi.org/10.3390/e25060875

Chicago/Turabian Style

Wootters, William K. 2023. "Optimal Information Transfer and the Uniform Measure over Probability Space" Entropy 25, no. 6: 875. https://doi.org/10.3390/e25060875

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop