1. Introduction, and Parental Advisory
Some of those who consider themselves true physicists may find what I have to say in this paper at best meaningless or stupid, at worst heretical. I am happy with the notions that quantum mechanics is non-local, that the physics of quantum mechanics is not time-reversible, and that it involves irreducible randomness. It seems to me (a mathematician and an applied statistician) that that is the message which quantum mechanics has been shouting at us since its birth: namely, that that is what reality truly is like. Obviously, this contradicts what by the end of the 19th century was dogmatic. The wonderful thing of the Växjö conferences has been that people interested in quantum foundations have been able to say such terrible things to one another and remain friends.
I want to introduce my take on ideas of Belavkin (2007) [
1] and Landsman (1995) [
2] which I find very closely related and very attractive. Belavkin is unfortunately no longer with us; Landsman by now has quite different ideas; see their book Landsman (2017) [
3].
The paper is about the Schrödinger cat paradox, which, I believe, is indeed truly only a paradox, since (as a purely mathematical issue) it can be easily resolved. This does have consequences for physics. We have a mathematical framework which might be applied to a physical context in different ways. How to apply it in a given context is a question for physicists to decide. Converting a bug into a feature, if there is some choice as to how a certain framework can be applied to a real world physical problem, one can also say that this offers a challenge to experimenters to find out whether experimental results can be used to restrict that choice. In our case, if wave function collapse is real and non-local and due to gravity, experiment should be able to tell us more about it.
It is very important to distinguish between the mathematical models which mathematical physicists develop, and the real world itself. The models are supposed to describe the real world, and aid us in engineering it. Such a model helps one to understand the phenomenon being modelled, in the sense that it makes one comfortable with it and aids in dealing with it, but it is clear that quantum mechanics contains elements which seem to conflict with very basic human understanding. We evolved in a real world with a lot of randomness, but we find randomness horrific, since our evolutionary speciality is to predict what might happen next and base our actions on choices between different possible futures. We associate randomness with the choices of the Gods, and they may not be benevolent.
Both Belavkin and Landsman argued for including the mathematical existence of a classical-like world as part of the axiomatic foundations of quantum mechanics. For them, classicality is not merely an emergent phenomenon. Landsman (1995) went on to distinguish two kinds of classicality, or attitudes to it. He wrote “those believing that the classical world exists intrinsically and absolutely (such persons later termed by them B-realists) are advised against reading this paper”. He adopts a milder position, calling it that of an A-realist: we live in a classical world but to give it special status is like insisting that the Earth is the centre of the universe. The B-realists are accused of living under some kind of hallucination. Landsman presents arguments pointing in a particular direction to a resolution of the measurement problem which at least would satisfy the A-realists. We point out in this paper that the theory earlier developed by Belavkin (surveyed in their 2007 paper) seems to complete Landsman’s program or at least exhibits a “realisation” satisfying their desiderata. At the same time it seems that this completion of the program ends up giving both A- and B-realists equal licence to accuse the others of living under hallucinations. I think that this distinction is a philosophical distinction and as much an aesthetic issue as anything else.
The theory is presented in the context of a standard (separable) Hilbert space description of quantum theory. The author is well aware that to go further one will sooner or later have to leave Hilbert spaces behind one, and enter a more abstract and exotic mathematical universe, to do justice to the nature of our real universe.
2. The Basic Framework
Belavkin’s eventum mechanics, developed in the 80’s, and a recent exposition of which is given by Belavkin (2007), has been created in an attempt to resolve the Schrödinger cat problem by showing that measurement, and random collapse of the wave function, can be seen as the result of a deterministic unitary evolution as long as one recognises that this evolution must take place in a mixed classical-quantum system. The collapse of the wave function is the stochastic result of a deterministic, unitary evolution in a situation where there is a quantum interaction between a quantum system and a classical system. Classicality corresponds to a superselection rule, saying that not all observables (in the sense of bounded operators) are actually observable (in the sense that quantum superposition of certain states cannot occur, or at least, can never be detected). The essential and unorthodox aspect of the theory is that it is time irreversible. Unitarity is retained but in the Heisenberg picture, the time evolution of the relevant observables is an endomorphism, not an isomorphism.
We present Belavkin’s basic theory in the most simple possible mathematical context, involving nothing more elaborate than separable Hilbert spaces and and their tensor product . There will be a unitary operation U and an initial pure (vector) state defined on . The initial state, together with iterated application of U on the initial state vector, define a discrete time evolution of the system. The evolution can also be traced backwards in discrete time, through iterates of . To be honest, though, we do need to make use of some elementary properties of von Neumann algebras. Here is the definition. Given a collection of bounded operators on , its commutant is the collection of all bounded operators on which commute with every element of . A von Neumann algebra is a collection of bounded operators on the Hilbert space, whose bicommutant is equal to itself: . Concrete examples are given later.
We suppose that has a particular orthonormal basis denoted by , where the index set is finite or countably infinite. Let denote an arbitrary (finite or countably infinite) orthonormal basis of so that the kets form an o.n.b. of . The coordinate is supposed to indicate the (classical) state of the real world, which evolves stochastically by its interaction with an underlying quantum world. In order for this to be meaningful we must make some assumptions relating U to the product structure of and to the preferred basis of .
Later we will discuss the more general situation in which is not necessarily (a priori) of a product form, and in which a “preferred basis” of part of this space emerges, though in general not uniquely, from the other physical information about the system: specification of U.
We next introduce certain algebras of bounded operators on
,
and their product
. A *-algebra of bounded operators of a Hilbert space
is a subset of
, closed under addition, multiplication (composition of operators), scalar multiplication with complex numbers, and the involution * (adjoint). It is called a
-algebra when it is a closed subset of
with respect to the operator norm; hence it is also complete for the topology induced by this norm. It is called a von Neumann algebra it is furthermore closed with respect to the weakest topology making the “matrix elements”
continuous for all
,
. According to von Neumann’s bicommutant theorem, this is equivalent to
. Abstract versions of all these objects also exist; in particular, the abstract version of a von Neumann algebra is called a
-algebra. The reason we must insist on von Neumann algebras is that normal states—states satisfying a natural continuity property—can be represented by density matrices: trace class operators on
. It has been said that
-algebras form the right context for non-commutative geometry; von Neumann algebras the context for non-commutative probability. Firstly, define
Using a prime to denote the commutant of a set of bounded operators, i.e., the set of bounded operators each of which commutes with everything in the first set, one can verify by direct calculation that
and hence
It follows from von Neumann’s double commutant theorem that
is a von Neumann algebra: that is to say, a
-algebra which is closed under the weak norm topology. The elements of
all commute with one another, and
is maximal in the sense that no other element of
commutes with all of
.
The tensor product of von Neumann algebras is the smallest von Neumann algebra containing the tensor products of individual elements of the two algebras.
is the von Neumann algebra of all complex multiples of the identity on
. One can verify by direct calculation (or by appeal to general theory of von Neumann algebras: the commutant of a tensor product is the tensor product of the commutants) that
Note that
is an algebra of commuting observables, and as such it is maximal. We also have
We shall refer to the elements of as observables. The word observable is just used for convenience. We are setting up a toy universe in which there are no external observers or measurements. There are just places or sectors in this universe, which we will call “worlds”, which have rich enough properties that they support “life as we know it”. One can imagine physical objects called observers living in such a world (maybe even imagining that with free will they can choose to do various different measurements), who see a consistent stochastically time-evolving environment. Yet these observers and their measurement devices are subject to the same laws of quantum physics as everything else.
We think of as being a set of beables, that is to say, physical quantities which can be given definite values. Very soon we will add a condition so that becomes a set of viables, that is to say, the beables also have definite time trajectories, or lives. We see as a possible (classical) world which can be found within the quantum universe given by U and .
The commutant of , the algebra , is the corresponding set of predictables. As we will see, these observables have definite probability distributions relative to , and moreover are used to predict the stochastic future of the beables.
Now consider a unitary operator
U acting on
; so
. A pure state vector
is mapped by
U to
. We will use
U to define a discrete time dynamics on our system, corresponding to iterated application of
U. We say that this time evolution is
compatible with the classical-quantum pair
,
if
which can be easily shown to be equivalent to
The key to this equivalence is to note that since U maps products to products, it maps commuting variables to commuting variables. Note that in the Heisenberg picture, an observable A is mapped to by one forwards time-step, and an observable C is mapped to by one backwards time-step. Thus, the assumption just made states that future predicables are also predictable, and past beables are also beable. Rather neatly, one need only assume that future predictables are also predictable or that past beables are also beable. A classically-minded physicist would prefer the latter; a quantum-minded physicist the former. The other property comes for free.
Belavkin uses the word non-demolition property, or causality property, rather than compatibility. We here use the word compatibility because later we will consider and U as given, the essential physics of our toy universe. The Hilbert space will not a priori be given any special structure. Given and U, a particular choice of von Neumann subalgebras having the properties , , if also compatible with U, then supports a viable quantum-classical world which is part of the universe. There can be many such worlds.
Incidentally, one can again start with assumptions for the classical-minded and assumptions for the quantum-minded. The classical-minded physicist starts with a commuting von Neumann algebra , defines ; he automatically obtains and the other relations between and . The quantum minded physicist starts with a von Neumann algebra which is such that is commutative; she similarly obtains all the other relations between the two algebras.
A state on the product space can be represented in the ordinary way with a (bipartite) density matrix , and the expectation value of an arbitary observable B is . We want to restrict the state to the algebra . We wish to think of the state abstractly as the mapping from a given sub-algebra (in particular, ) of observables to their expectation values. In that context, different density matrices can be indistinguishable from one another, i.e., generate the same expectation values.
Denote by the subspace of all vectors of the product space of the form , with arbitrary, and as the orthogonal projection onto this space, then with respect to the algebra , a bipartite density matrix cannot be distinguished from . Normalizing each component of this sum, the state lives on the subspace which is just a copy of . Write ) and as the just mentioned normalized state, thought of as a density matrix on . Together, these remarks mean that any quantum state on can be represented uniquely with the density matrix , where , and the form a probability distribution on . The states are pure states on —they cannot be written as mixtures of other states (where states are taken to be “expectation values” defined on ).
can also be called a sector of and corresponds to a superselection rule: quantum superpositions between different eigenstates of a corresponding (possibly unbounded) observable X on are impossible. In our product construction, a particular superselection rule was put into the model by hand. However, if we just consider and U as given, different choices of and can be identified, compatible with the given and U. Thus, different, mutually incompatible, superselection rules can be identified. However, if one would make some partial requirements on or , it is possible that their complete identity would then be fixed; in other words, a superselection rule can emerge from the physics. We need in advance to specify time, the unitary U, and we need to specify a weak kind of locality corresponding to a notion of space, in the form of some commutation properties. Principles of causality then determine what the classical world looks like.
Inspection of how an observable in transforms under U reveals that U must have the following structure: written out blockwise with blocks , which are operators on , indexed row-wise and column-wise by , for each y there must exist at least one x such that ; for each x there exists exactly one y with . Thus, there exists a function f from onto itself such that if and only iff . (For any , the observable must transform into a linear combination of observables of the same type.)
The unitarity of U implies that the satisfy , if , if , . Conversely, given any f and satisfying these properties, we can reconstruct U, compatible with .
The (forwards) Heisenberg evolution of observables in
can be described through the mapping
The (forwards) Schrödinger evolution of states on
is similarly described through
The interpretation of this change of state is that the classical coordinate y jumps to one of the coordinates x with with probability while the state on is transformed into , normalized.
I view Equation (
1), the most important result in this paper, as describing spontaneous collapse. In fact, the formula quite simply expresses iterations of the Lüders-von Neumann rule for the probabilities of outcomes and resulting transformation of quantum state, applied to successive observation of a sequence of observables of components of a composite quantum system. The probabilities are those given by the Born rule, the state is transformed by projection onto the eigenstate corresponding to which eigenvalue
x has been realised (by the Lüders-von Neumann rule). Back to von Neumann (1932) [
4]. However, now, it is not the experimenter or observer who chooses a measurement to make. The sequence of “measurements” being made is determined by the unitary evolution of the wave function of the universe; classical reality is realised step by step in a stochastic way with probability laws determined by the same underlying deterministic laws; the wave function undergoing repeated random partial collapses.
We see that the more simple situation in which f is not only onto but also one-to-one is so simple as to be completely uninteresting: the classical part follows a deterministic path, according to the iterates of the inverse of f; in each classical state a corresponding unitary evolution takes place of the state of the quantum part. The evolution can be termed autonomous in the sense that the classical world follows a deterministic path not influenced in any way by the state of the quantum world.
So the interesting situation is that f is onto but not one-to-one. This has some immediate consequences: first of all, it forces to be infinite, and secondly, because , where the sum can be over several x, and at the same time if , the matrix is not itself unitary when y is the image of several x. Thus, the space must be infinite dimensional.
Though the forward evolution of the classical part is stochastic, its backward history is deterministic: if U has been applied repeatedly bringing us into the classical state x, the classical history is given by the iterates of f on x. In terms of observables, maps classical observables to classical observables in the (reversed time) Heisenberg picture. The classical observables commute with everything, and can all be assigned values simultaneously. In particular, the whole past trajectory of the classical system up to the present time is itself classical.
These features of the model are retained when we drop the special product structure of the Hilbert space . One point of this analysis is to show by elementary and direct means that the features exhibited by various toy models are generic to the approach. In particular, we can always identify some kind of shift operator—acting on classical trajectories—which is the source of the quantum-classical interaction in the model. The future of the trajectory is hidden in the quantum future; the past of the trajectory is fixed in the classical present.
4. Some Examples
4.1. Representing a CP Map
One can embed an arbitrary CP map (taking quantum states to mixed classical-quantum) into eventum mechanics (allowing to extract both the measurement outcome and the transformed state). In this paper, we will only do a simple example, with only a shift on infinite chains of two-level systems.
The basic trick goes something like this. Let be the Hilbert space of the system being transformed and/or measured by the CP map. Consider an operator sum representation with matrices such that ; i.e., for simplicity and this outcome space is finite or countably infinite. The CP map produces the classical outcome x and transforms to the state , normalized, with probability equal to the normalization constant, where is the state (which is arbitrary) in of the system being transformed. I add to this a Hilbert space of the apparatus and a Hilbert space of the environment. The space of the apparatus will be simply . For the environment, take an infinite collection of copies of , indexed by . The tensor product of all these spaces is not separable but we restrict attention to a part of the space, namely , the closure of the span of the countably many orthonormal vectors for which all but finitely many of the coordinates are equal to a special value 0. To simplify the construction let us suppose that is not a possible value of the outcome of the measurement, i.e., . Otherwise we simply extend by adding one point different from those already present and call it 0. The environment component will be considered as the product of two parts, , by writing . We now have got a large, separable Hilbert space for system, apparatus and environment, where the environment again is the product of two parts thought of as classical and quantum, respectively. The algebra of observables of the joint system will be that generated by taking arbitrary quantum observables of the system, apparatus, and quantum environment, together with only classical observables (diagonal in the specified basis) of the classical environment.
The centre of this algebra can be identified with the classical observables on . The classical states of this algebra correspond to infinite sequences of elements of indexed by the negative integer , which only have a finite number of elements unequal to the special element 0.
The initial state of apparatus will be the state in which , and that of the environment will have for all .
We now describe a unitary mapping on the product system, as the composition of the following three maps, each working on different parts of the system. We describe the mapping in the Schrödinger picture, as unitary maps to be applied consecutively (on the left) to a vector in the large product system. First there is a unitary mapping of to itself satisfying . As specified so far, the map preserves inner-products, and it can be extended in many ways to be unitary on the entire space . This part will be very familiar as one of the many ways to express a CP map as a unitary mapping on a larger space followed by a projective measurement of part of the space (or by tracing out the complementary part).
Next we copy the measurement outcome, though still thought of as quantum (superposition is still possible) into the quantum part of the environment, and specifically the component of the quantum environment. The unitary achieving this can be taken to be any unitary taking to where the two indices stand for and the zero’th component of .
Finally we apply the unitary mapping to which performs the left-shift, taking to . The composition of these three maps is called U. It operates as required on the observables of the joint system, since in the Heisenberg picture we first have the right shift, shifting the classical observable at position into a larger quantum space, while the subsequent steps are unitary mappings acting on quantum observables only and leaving classical observables unchanged.
Back in the Schrödinger picture take the initial state of the combined system to be , where we abuse notation by writing just the state vector rather than the corresponding density matrix. Applying U to this state converts it into the mixture, with probabilities , of the state which is the product of , normalized, of the system, together with the pure state of apparatus and environment with , , all other components equal to 0.
To make the description of the measurement make sense when U is iterated, we note that the apparatus contains a quantum memory of whether or not it has already been used, depending on whether or not. We can append to our prescription to steps 1 and 2 above, that when initially is not 0, nothing happens at all, while step 3 is unchanged. Alternatively, the classical part of the environment also contains a memory of whether or not a measurement already took place, so we can achieve the same effect by letting it control what happens in steps 1 and 2. In that case we could also delete one of our copies of and reduces steps 1 and 2 to a single step, by effectively taking the zero’th component to be the apparatus rather than part of the quantum environment.
After applying U any number of times, the classical environment is in a classical state which tells us when the measurement took place and what the outcome was; the probability of any particular outcome is what we require. After the measurement there is no entanglement of system, apparatus and environment; there is just a classical correlation. The system could be detached and measured another way in another measurement apparatus; the CP map we have implemented can also be seen as a quantum channel.
There are three unsatisfactory features of this model, apart from the huge amount of freedom which is left in how to completely specify the unitary maps involved.
The first is the huge size of the quantum environment—infinitely many copies of the apparatus space, even if the apparatus can be taken finite dimensional. However we know that we need to presume infinite dimensionality of the quantum part of the world, and the block-wise description which we have of possible U shows that something like a shift of an infinite sequence is going to be unavoidable. Since the model allows iteration of the unitary map, we have to allow infinitely many branches in the classical outcomes, and these have to be the reflection of an infinite possibility of branching in the quantum part. Moreover, every attractive measurement model so far discussed in the literature either needs an infinite system to begin with, or speaks about limiting properties of systems of larger and larger numbers of copies of finite systems. Such models typically give the nice results they do only in the large time limit. The Belavkin model could be thought of as an attempt to complete or extend the existing classes of models, so that behaviour which formerly could only be approximated arbitrarily well, can now also be exhibited exactly, inside the model.
The second unsatisfactory feature of the model is the fact that the initial state of the environment must be taken as fixed. Many attractive attempts to model measurement actually work by assuming an environment consisting of many small systems in a mixed state. One keeps inside a traditional framework with a unitary mapping on a completely quantum space of observables, generating classical probability at a macroscopic level from classical probability inserted at a microscopic level. These arguments usually involve some kind of averaging over microscopic degrees of freedom to destroy quantum coherence between different macroscopic states. Though physically appealing (Landsman quotes van Kampen as having said that someone who does not accept this does not understand what physics is) this argument is metaphysically very unsatisfactory, and still leaves the question open as to whether mathematically attractive models can be built within which the limit has been attained. However, at least the approach does respect the fact that a macroscopic apparatus and its even larger environment is never going to be in a very special, controlled, initial state. The answer to this from the Belavkin side must be that there do exist special states within macroscopic quantum systems. Is this the vacuum state of some quantum field? From this point of view, the quantum environment is actually at a very deep level inside of the systems being studied, and represents simply the effects of pure quantum noise from deeply microscopic levels.
The third apparently unsatisfactory feature from Landsman’s point of view is that the unitary mapping
U is strongly related to (or constrained by) the chosen algebra of beables. Landsman would rather find the algebra of beables emerge from the description of
U. That is indeed a feature of the toy models we have discussed, but in an abstract approach one is completely free to start from specification of
U and then identify possible, and indeed incompatible, algebras of beables. For the same unitary mapping
U there do exist different algebras with different centres, corresponding to rather different kinds of observers who are completely incompatible with one another, as Landsman would appreciate. We return to this discussion in
Section 5.
4.2. The Geiger Counter
Basic books on quantum physics never give a model for the Geiger counter, supposed to give a click on the detection of a radioactive emission coming from a single atom. Yet this is presumably the apparatus in the Schrödinger cat story, which is supposed to detect whether or not an emission occurs in a certain time interval, and according to this trigger the poisoning or not of the cat. It is possible to give a simple Belavkin-type model for this situation, where an atom starts in the initial state
. With probability
it delivers a macroscopic signal at a geometrically distributed random time (exponentially distributed in the continuous time limit), with probability
it never emits a signal. If after any time the atom has already given a signal, the atom will be in the state
. If after a long time it still has not emitted a signal, it will still be in a superposition of
and
, but with more and more weight on
as the length of time we have waited (and still nothing has happened) increases. The model has been worked out in detail by Feenstra (2009) [
5], and by Brown (2021) [
6].
4.3. Continuous Time
A decent (CP) continuous time measurement (and state transformation) process can also be represented in the Belavkin picture. The time shift operation becomes more natural than ever, the classical and quantum environment become larger but also more physically interpretable. It seems that as we scale up the model towards reasonable levels of space-time complexity, what initially seems like an excess of hidden layers in model, some of them too narrowly prescribed, others embarassingly free, comes into a decent balance with what has to be in the model anyway. Feenstra (2009) works out how the master equation from quantum optics can be neatly expressed as a Belavkin model, not surprisingly in view of Belavkin’s work on continuous time measurement and control of quantum systems in quantum optics.