9 Reasons Quantum Mechanics is Incomplete - Comments

In summary: I don't know...the lead singer for a band that sounds like it borrows from Pink Floyd and Radiohead.In summary, Dirac believed that QM is just a natural outgrowth of classical physics where some of its assumptions are relaxed. However, because QM is still a physical theory, it is incomplete.
  • #36
Hi,

is it not possible to verify the completeness of any mathematical axiomatization of quantum mechanics?

http://www.iecl.univ-lorraine.fr/~Wolfgang.Bertram/WB-TB.pdf

/Patrick
 
Physics news on Phys.org
  • #37
DarMM said:
I think this mixes up the incompleteness of the Standard Model (mass and coupling parameters) with the incompleteness of Quantum Theory in general (measurement problem, meaning of contextuality).
In general that's true, but in my view of Bohmian mechanics those two types of incompleteness are closely related. See the paper linked in my signature.
 
  • Like
Likes DarMM and bhobba
  • #38
microsansfil said:
is it not possible to verify the completeness of any mathematical axiomatization of quantum mechanics?

It is not possible to verify the axiomatic completeness of any system as strong as arithmetic. Note - this is verifying completeness in the Godel sense because we always have true statements that can't be proven in the system. But that has nothing to do with if a system is not complete because its axioms may contain a hidden circularity with regard to observations, which is what people that talk about QM completeness mean. You are falling for a common semantic error - because two different things use the same name you context shift. Such reasoning is invalid.

Thanks
Bill
 
Last edited:
  • #39
ftr said:
SM is part of QM, measurement problem is an "interpretation" issue but mass and coupling is real/clear issue. I think understanding will come when the later is addressed first.
The measurement problem is a problem in QM that interpretations try to solve, but it is a real issue as much as the masses and couplings. It might be the case that the explanation of couplings and masses is related to the measurement problem or it might not, from where we stand now they are two separate problems.
 
  • Like
Likes bhobba and Demystifier
  • #40
  • #41
Demystifier said:
I use it colloquially.

How do we measure the distance between physical theory and the ontology of nature, without having access to it? So, therefore, any physical theory will be incomplete?

/Patrick
 
  • #42
"There is an objective reality out there". I don't think that applies to the relational interpretation. And what about the "new" Copenhagen interpretation with decoherence taken into account?
 
  • #43
Regarding “the incompleteness of QM interpretations”, there is, to my mind, a subtle hint given by David Bohm and Basil J. Hiley in “The Undivided Universe: An Ontological Interpretation of Quantum Theory”:

"Several physicists have already suggested that quantum mechanics and consciousness are closely related and that the understanding of the quantum formalism requires that ultimately we bring in consciousness in some role or other (e.g. Wigner [17], Everett [18] and Squires [19]). Throughout this book it has been our position that the quantum theory itself can be understood without bringing in consciousness and that as far as research in physics is concerned, at least in the present general period, this is probably the best approach. However, the intuition that consciousness and quantum theory are in some sense related seems to be a good one, and for this reason we feel that it is appropriate to include in this book a discussion of what this relationship might be." [emphasis added by LJ]
 
  • Like
Likes Demystifier and AlexCaledin
  • #44
There is one reason that quantum mechanics is a complete physical theory. Quantum mechanics makes a class of predictions. Those are probabilistic predictions but since you can embed logic within probability by restricting to specific cases of p=0 or 1, this is not a failing on the part of QM. And quantum mechanics also specifies that there are no stronger predictions that can be made. The uncertainty principle defined the end of the question of what more can be known about what will happen.

So any improvement upon the predictions of QM must necessarily violate QM. That's fine, GR breaks SR except as a limiting case. SR breaks Newtonian mechanics except at a limiting case. QM is not necessarily the end of all theories but it is complete in the sense that no more exact theory can be given that incorporates all of QM's predictions.

Now if you feel it is "incomplete" because it doesn't provide for a specific observably distinguishable ontological model then I would point out to you that QM is not an ontological theory. It is a description of what happens (and how often) not of what is. That's the mistake people make in misinterpreting Copenhagen. Yes the wave function collapses but the wave function is not an ontological representative. It is a QM version of a classical probability distribution. Classical distributions collapse all the time.

The dissatisfaction people have with the Copenhagen interpretation which is the positivistic interpretation is the same sort of dissatisfaction they have with the non-absolute time in the Twins paradox which makes them one to throw in an aether (with necessarily screwy dynamical properties) in order to feel better about the relative ages of twins. The relativity of time and simultaneity is likewise the positivistic interpretation eliminating as meaningless that which cannot be observed, namely the luminiferous aether.

There is no fixed objective reality in QM. There are a set of relative frame dependent realities with well defined relativity transformations relating them. Fantasies about pilot waves and infinite worlds are no more helpful than questions about the unobservable properties of the aether.
 
  • Like
Likes DarMM, dextercioby, jim mcnamara and 3 others
  • #45
jambaugh said:
Yes the wave function collapses
What is the mathematical description of the collapse process?

Let's say we're going to do a Stern-Gerlach experiment. We isolate the lab completely from any external environment. We put the lab (and everything in it, including the experimenters) in an initial pure state, with the experiment about to begin. We time-evolve forward unitarily with the hamiltonian of the Standard Model.

I claim that the state of the lab splits into an "up" branch (where the measured spin was up, the dial points up, the brains of the experimenters in the lab think they saw the dial point up, etc) plus a "down" branch. (I believe this part must be correct, but maybe you disagree.)

So then, when does collapse occur? And what is the mathematical description of it?
 
Last edited:
  • #46
Avodyne said:
What is the mathematical description of the collapse process?
If you look at @jambaugh 's post he's clearly taking an epistemic view of the wavefunction where collapse is just (generalized) Bayesian conditioning.
 
  • #47
I don't see why it matters whether one views the quantum state as epistemic or ontological. QFT predicts a two-branched state for the post-measurement Stern-Gerlach lab (and the people in it). At what point in time, and according to what mathematical rule, do we throw away one branch?
 
  • #48
Avodyne said:
I don't see why it matters whether one views the quantum state as epistemic or ontological.
If the quantum state is epistemic then throwing away one branch doesn't correspond to a physical process which needs a more detailed mathematical description, it's simply (generalized) Bayesian updating.
 
  • #49
What is "(generalized)" Bayesian updating? (I know what ordinary Bayesian updating is.) Do you have a reference?
 
  • #50
It's just how Lüders rule is viewed in epistemic approaches, i.e. it represents a generalized form of Bayesian conditioning.
 
  • #52
Avodyne said:
Lüders rule, as I understand it (see below), seems to require an observer outside the system being measured. This is not the case in QFT, where observers and everything else are comprised of excitations of the quantum fields.

http://philsci-archive.pitt.edu/4111/1/Lueders_rule_BuschLahti.pdf
Well, for quantum field theory you may need the thermal interpretation, where observers can be part of the system and collapse happens as an approximate effect due to the ignored environment.
 
Last edited:
  • #53
Avodyne said:
Lüders rule, as I understand it (see below), seems to require an observer outside the system being measured. This is not the case in QFT, where observers and everything else are comprised of excitations of the quantum fields.

http://philsci-archive.pitt.edu/4111/1/Lueders_rule_BuschLahti.pdf
How does this affect an epistemic view? Even if you include the observer themselves within the quantum description there is no need for the quantum state to be a physical/ontic object and thus no need for a dynamical or mathematically detailed description of where the other branches "go".

For example in a dice roll there are probabilities for each roll ##p(1)##, ##p(2)## and so on. I could also include the observer in this:
$$p(1, Observer\hspace{2pt} sees \hspace{2pt}1)$$
$$p(2, Observer\hspace{2pt} sees\hspace{2pt} 2)$$

However there is no need to explain where ##p(2, Observer\hspace{2pt} sees\hspace{2pt} 2)## "goes" upon the outcome ##1##.

Since observers can be included within the system even in QM, this isn't particular to QFT.
 
  • #54
DarMM said:
How does this affect an epistemic view?
Because you still need rules to decide when to modify the state by chopping off branches. If you do it under the wrong circumstances, you will destroy observable quantum coherence effects.

Psi-epistemic interpretations have all sorts of other problems as well:

https://arxiv.org/abs/1303.2834
 
  • #55
Avodyne said:
Because you still need rules to decide when to modify the state by chopping off branches. If you do it under the wrong circumstances, you will destroy observable quantum coherence effects.
You remove the branches related only to unobtained outcomes for the systems you actually observe. This is not in contradiction with a superobserver retaining a superposed description of your entire lab. See for example Richard Healey's "The quantum revolution in philosophy" Chapter 11 for a good summary of this.

Avodyne said:
Psi-epistemic interpretations have all sorts of other problems as well:

https://arxiv.org/abs/1303.2834
That paper concerns psi-epistemic views where there is an underlying ontic space obeying the ontological framework axioms, just like the PBR theorem. It says nothing about epistemic theories in general such as for example acausal views. It's not really an issue for any of the epistemic views actually being worked on. Especially ones where there is no underlying ontic hidden variable space.
 
  • #56
I consider orthodox quantum mechanics to be incomplete (or maybe inconsistent) and I don't think it really makes any difference whether you consider the wave function to be epistemic or ontological. Wigner's friend or Frauchiger and Renner shows this incompleteness.

Rather than going through the F&R argument again, let's just take something much simpler. Alice and Bob prepare an electron so that it is a superposition of spin-up and spin-down. Alice measures the spin. According to orthodox quantum mechanics, she either gets spin-up with such-and-such probability, or she gets spin-down with such-and-such probability.

But Bob, considering the composite system of Alice + electron, will predict that that composite system will evolve into a superposition of "the electron is spin-up and Alice measured spin-up" and "the electron is spin-down and Alice measured spin-down". From Bob's point of view, Alice doesn't get a definite result, but somehow turns herself into a superposition of both results.

So it seems to me that either you're led to (1) the "fantasy" that is many-worlds, or (2) quantum mechanics is inconsistent, or (3) somehow Alice being in a superposition of states doesn't preclude her being in a definite state.

The latter possibility is realized by the Bohmian interpretation, but I would say that if you take that way out, then you are admitting that standard quantum mechanics is incomplete, because the Bohmian interpretation has additional variables (the actual positions of particles) that are not present in standard QM.

Calling the wave function "epistemic" rather than "ontological" doesn't actually accomplish anything, in my opinion. So you want to interpret Alice in her superposed state as being epistemological? Does that mean that she's ACTUALLY in some definite state, and we just don't know which? The problem with that is that there is actually a difference between "Alice is in this state or the other, we just don't know which" and "Alice is in a superposition of this state and that state". I don't think you can consistently treat the latter as epistemological. In any case, if you say that the superposition reflects your lack of information about the "true" state of Alice, that to me means again that quantum mechanics is incomplete, since it presupposes a "true" state.
 
  • Like
Likes eloheim
  • #57
Parsing your description of the classic thought experiment, StevenDaryl

Alice and Bob prepare an electron so that it is a superposition of spin-up and spin-down. [\quote]
fine so far.
Alice measures the spin. According to orthodox quantum mechanics, she either gets spin-up with such-and-such probability, or she gets spin-down with such-and-such probability.
Yes and so you should then taking into account that she made this measurement and recorded her result in a classical copiable memory by describing Alice + electron via a density operator:
[tex] \rho =\rho_A\otimes \left( \begin{array}{cc} 0.5&0\\0&0.5\end{array}\right)[/tex]
But Bob, considering the composite system of Alice + electron, will predict that that composite system will evolve into a superposition of "the electron is spin-up and Alice measured spin-up" and "the electron is spin-down and Alice measured spin-down".
If Bob does this he is making a grave mistake. Alice is not at zero entropy and so cannot be described sharply. He, if he wishes to describe Alice plus electron before the physical and dynamic act of measurement Alice will make occurs would presumably use a density operator to include the entropy dump Alice necessarily must utilize to amplify the electron's spin component into a macroscopic record.
From Bob's point of view, Alice doesn't get a definite result, but somehow turns herself into a superposition of both results.
From Bob's point of view, if he has correctly represented the measurement process of Alice, if he correctly represents the dynamic that makes her recording of the electron spin possible, he will end up with a density operator of the form I mentioned above. In short Alice as an observer necessarily must interact with the system entropically and decoher it. At this point Bob is not describing this Alice but any such Alice electron pair given the same or equivalent initial conditions. Then Bob classically measures the record of Alice and collapses the classical probability distribution down to a single classical sharp state just the same way that the LOTTO commission does each time they bounce their ping pong balls and collapse the value distribution over the tickets to sharp "win" or "lose" states.
 
  • Like
Likes DarMM
  • #58
Well first of all after going through the paper more I don't think FR adds anything beyond Wigner's friend.

Secondly an epistemic account will by nature be taking the view that QM is incomplete as it views a central object in it as non-representational. What the epistemic approach is supposed to add is not a complete and closed account of the ontology of nature (explicitly because they are epistemic). Rather they seek to provide an explanation of several of the properties of quantum states by saying that they are epistemic objects. So for example teleportation, local indistinguishability, no cloning, remote steering, interference, non-commutativity are viewed as having a more natural explanation if you understand ##\psi## as epistemic.

However saying that these properties of ##\psi## have a more natural explanation if ##\psi## is a sort of generalized probability distribution, does not mean you're claiming to know the fundamental ontology of nature. In fact you're explicitly not, since you are providing an epistemic account.

The two main classes of epistemic views then differ on what the "completion" of QM will be like. Roughly speaking ##\psi##-statistical views will see the underlying theory as having some strange ontology like retro or acausality. ##\psi##-epistemic/doxastic views like QBism, Bub or Copenhagen view a completion as impossible for reasons that differ between them.

However to reiterate, by their nature they deny that QM is a complete representational account of the world. The simply seek to explain the properties of ##\psi##.

stevendaryl said:
I don't think you can consistently treat the latter as epistemological
Spekken's toy model gives an epistemic set of states that have interference at the observer + system level, while having a definite outcome at the system level. Superpostion of the total system isn't incompatible with definite outcomes for subsystems. If it was epistemic views would be finished.
To explain briefly in Spekkens toy model superposition arises from the structure of the space of probability distributions under the presence of an epistemic limit. When considered over "system + environment" that limit has a different form than over just "system", so we have a superposition for the former, but a definite outcome for the latter.
Actual epistemic views are more complicated than this, but it gives a good example of how superpositions are compatible with definite subsystem outcomes.
 
  • #59
jambaugh said:
Yes and so you should then taking into account that she made this measurement and recorded her result in a classical copiable memory by describing Alice + electron via a density operator:

No, you shouldn't. That's introducing an ad hoc element to quantum mechanics. Through unitary evolution a pure state never turns into a mixed state. It becomes a mixed state through a modeling decision on OUR part. We can trace over unobservable degrees of freedom and produce a mixed state, but as I said, that's something WE do. It's not a physical effect.
 
  • #60
Let me add again, the collapse is not mysterious for classical probability distributions because they too are fundamentally (in the frequentist's intepretation of probability) phenomenological descriptions. One is not describing the state of a six sided die, one is describing the behavior of that die when tossed. The phenomena of its relative frequencies of outcomes when tossed repeatedly.

I make this distinction in my Prob Stat classes by beginning the semester having the students estimate the probability that at least two class members have the same birthday. I give them two guesses which is a major hint, and write my two guesses down on a piece of paper. We then go through and verify if it is the case (highly probable for my classes of 30 or more students). I then ask them again, what's the probability as I turn over my guesses "Either it's P=1 or P=0". The point here is you can calculate the probability that an arbitrary set of 30 people will have a common birthday but that is not a description of the reality of our set of students, it is a description of the class of such sets for which ours is an instance. Saying our specific group has a certain probability distribution is shorthand for saying that group belongs to a class of such groups with the given probability distribution.
 
  • #61
stevendaryl said:
No, you shouldn't. That's introducing an ad hoc element to quantum mechanics. Through unitary evolution a pure state never turns into a mixed state. It becomes a mixed state through a modeling decision on OUR part. We can trace over unobservable degrees of freedom and produce a mixed state, but as I said, that's something WE do. It's not a physical effect.
Right, but to give a valid pure state description of a measuring device requires you break the device by measuring it so exactly its entropy becomes zero. This precludes it being a measuring device. That's the fallacy of Wigner's friend arguments.
When you include the measuring device prior to measurement you have automatically started with a "mixed state" because to measure it must carry with it an entropy dump. From Bob's perspective the system evolves unitarily and by virtue of the measurment having occurred the description will necessarily involve entanglement between measuring device and system measured. The entropy of the composite doesn't change but the sum of the sub-entropies does go up due to this entanglement. The entropy dump and measurement record by themselves have higher entropy, the measured system by itself has completely decohered because it has been maximally measured i.e. maximally entangled with the measuring device.
 
  • #62
jambaugh said:
Right, but to give a valid pure state description of a measuring device requires you break the device by measuring it so exactly its entropy becomes zero. This precludes it being a measuring device. That's the fallacy of Wigner's friend arguments.

I consider that completely bogus. You don't have to know the initial state precisely in order to do reasoning about it. You do conditional reasoning: If it is in state x, then it will evolve into state y.

From Bob's perspective the system evolves unitarily and by virtue of the measurment having occurred the description will necessarily involve entanglement between measuring device and system measured.

From Bob's perspective, it's false that Alice got spin up or spin down. So what, exactly is the content of the claim that a measurement produces one of the eigenvalues with such-and-such a probability?
 
Last edited:
  • #63
stevendaryl said:
I consider that completely bogus. You don't have to know the initial state precisely in order to do reasoning about it. You do conditional reasoning: If it is in state x, then it will evolve into state y.
You're thinking in terms of classical states. You disagree with getting a mixed state after the fact. I point out that your problem is starting with as sharp state and claiming part of the system is able to make a measurement.

Ok. Let me try it without that, basically a deferred measurement scenario let "Alice" be a simple enough system to be sharply described in a pragmatic sense. To say "Alice measured the electron" Alice and the electron become maximally entangled. Bob's proper description of the Alice-electron composite necessarily represents this and so Bob cannot describe it in terms of Alice's state and the electron's state. To speak of Alice alone he must trace over the electron's density operator (while the composite is still sharp) yielding a mixed state of Alice with 50-50 percent probabilities of having a "up" vs a "down" record of the electron state. Likewise to describe the electron alone Bob must trace over Alice's density matrix yielding again a mixed state of the electron in that 50-50 classical superposition of outcomes. And we're back to playing Lotto.

What I'm doing here is basically letting Alice borrow Bob's entropy dump. It's not Wigner's friend but an indirect measurement. I would not describe Alice as having made a measurement here because she's too damned cold. But it shows again that Yes the composite system as a whole evolves unitarily but the factor system does not since it becomes entangled with the other factor system.

The "measuring system" must have enough spare degrees of freedom to make the states of nearly maximal entanglement much more likely than a spontaneous return to un-entangled. But it is this inevitable entanglement that is the 2nd law of thermodynamics in action. It is the fact that though Me plus the photon that bounced off my head and is heading toward Alpha Centauri may be evolving unitarily, I cannot access the photon any more since I am now outside its future light cone. I must discard any hope of recovering its part of our entangled information and trace it out. I alone now have a "mixed state".

( It grates on me to use this "my state" language because it is the very source of confusion by pretending one is making ontic descriptions using QM. But it is the language of the intended audience.)
 
  • #64
jambaugh said:
You're thinking in terms of classical states.[/QUOTES]

That's not true. Schrodinger's equation tells us how to evolve a quantum state from one time to another.

I point out that your problem is starting with as sharp state and claiming part of the system is able to make a measurement.

Why don't you restate what you're claiming without mentioning the word "measurement"? A measurement is nothing more than interaction that causes a macroscopic variable (such as whether there is a spot on a photographic plate at a particular location) to become correlated with a microscopic variable, such as the z-component of the electron's spin. You can (presumably) derive from physics the fact that if a spin-up electron passes through the magnetic field then it will be deflected in one direction and will produce a spot over here. If a spin-down electron passes through the magnetic field then it will be deflected in the opposite direction and will produce a spot over there.

But the linearity of quantum mechanics implies that a superposition of spin-up and spin-down will result in a superposition of the dot being here and the dot being there. (Okay, since the macroscopic device interacts with the rest of the universe, the prediction would be that the whole universe ends up in a superposition of states. Thus Many-Worlds, fantasy or not, is a consequence of quantum mechanics. It really is. Unless you want to add hidden variables that say that only one of the possibilities is "real").

Ok. Let me try it without that, basically a deferred measurement scenario let "Alice" be a simple enough system to be sharply described in a pragmatic sense. To say "Alice measured the electron" Alice and the electron become maximally entangled. Bob's proper description of the Alice-electron composite necessarily represents this and so Bob cannot describe it in terms of Alice's state and the electron's state. To speak of Alice alone he must trace over the electron's density operator (while the composite is still sharp) yielding a mixed state of Alice with 50-50 percent probabilities of having a "up" vs a "down" record of the electron state. Likewise to describe the electron alone Bob must trace over Alice's density matrix yielding again a mixed state of the electron in that 50-50 classical superposition of outcomes. And we're back to playing Lotto.

As I said earlier, this business of tracing over degrees of freedom is something WE do. It isn't a physical interaction. It's a modeling decision. We model the situation in which you have a system entangled with the rest of the universe as if there was a collapse to one of a number of macroscopically distinguishable states, and the mixed state represents the lack of information about exactly which state it is. But that's really bogus. The mixed state resulted from our decision to draw the boundary of the system we are interested in. There can't be physical consequences of a modeling choice.
 
  • Informative
Likes Mentz114
  • #65
jambaugh said:
[..]
I cannot access the photon any more since I am now outside its future light cone. I must discard any hope of recovering its part of our entangled information and trace it out. I alone now have a "mixed state".

( It grates on me to use this "my state" language because it is the very source of confusion by pretending one is making ontic descriptions using QM. But it is the language of the intended audience.)
I like what you've written because tracing out 'Me' has physical justification. But it cannot be a physical process because the entanglement is not broken. So it seems the tracing is to remove information that is inaccesible and not because of a change of any state.

I happen to believe what you say ( in so far as I understand it) and also that superposition does no physics and is only an accounting procedure. So I'd like to think that tracing out models actuality correctly.
 
  • #66
DarMM said:
Spekken's toy model gives an epistemic set of states that have interference at the observer + system level, while having a definite outcome at the system level. Superpostion of the total system isn't incompatible with definite outcomes for subsystems. If it was epistemic views would be finished.
To explain briefly in Spekkens toy model superposition arises from the structure of the space of probability distributions under the presence of an epistemic limit. When considered over "system + environment" that limit has a different form than over just "system", so we have a superposition for the former, but a definite outcome for the latter.
Actual epistemic views are more complicated than this, but it gives a good example of how superpositions are compatible with definite subsystem outcomes.

Spekkens's toy model assumes quantum mechanics is incomplete, ie. that there are hidden variables. Spekkens's work is motivated by the measurement problem, and aims to build on things like Bohmian Mechanics, and ask whether any other types of completions of quantum mechanics are possible. So I don't think Spekkens's use of "epistemic" is the same as @jambaugh's, since @jambaugh doesn't seem to be a fan of ontic states.
 
Last edited:
  • #67
Let me make my scenario even simpler. Imagine a world where there is basically nothing in the world except Alice, her measuring equipment, and her electron. That's the entire universe. This presumably could be described by an enormously complicated wave function. We can reason that if the electron starts off in a spin-up state, then the whole system will evolve into a state in which Alice is in the "having measured spin-up" state. If the electron starts off in the spin-down state, then the whole system will evolve into a state in which Alice is in the "having measured spin-down" state. So the linearity of quantum mechanics implies that if the electron starts in the state which is a superposition of spin-up and spin-down, with equal amplitudes, then the whole system will evolve into a state in which Alice is in a superposition of "having measured spin-up" and "having measured spin-down". That seems almost unquestionable to me.

(The reason I say "almost" is because measurement involves irreversible changes, and I'm not sure how to model irreversible changes using pure quantum mechanics.)

So if you want to insist that Alice's measurement of the electron's spin (along the z-axis, say) always produces the result of Alice having measured spin-up or the result of Alice having measured spin-down, then that means that we have to be able to interpret a superposition of macroscopically different states as being either in one state or the other.

But it's inconsistent to interpret it that way. A superposition of a spin-up electron and a spin-down electron is not either one or the other, it's an electron in a definite state (spin in some direction other than ##\pm \hat{z}##).
 
  • #68
atyy said:
Spekkens's toy model assumes quantum mechanics is incomplete, ie. that there are hidden variables. Spekkens's work is motivated by the measurement problem, and aims to build on things like Bohmian Mechanics, and ask whether any other types of completions of quantum mechanics are possible. So I don't think Spekkens's use of "epistemic" is the same as @jambaugh's, since @jambaugh doesn't seem to be a fan of ontic states.
Certainly, as I mentioned above there are two classes of epistemic views ##\psi##-statistical and ##\psi##-doxastic. Spekkens toy model is more in the tradition of the former kind.

QBism for example, a ##\psi##-doxastic view, would also say QM is incomplete, but not because there are hidden variables. Incomplete is not quite the same as "missing hidden variables".
 
Last edited:
  • #69
stevendaryl said:
So if you want to insist that Alice's measurement of the electron's spin (along the z-axis, say) always produces the result of Alice having measured spin-up or the result of Alice having measured spin-down, then that means that we have to be able to interpret a superposition of macroscopically different states as being either in one state or the other.
Okay let's look at Spekkens toy model. We have a system with an electron, a measuring device and then the lab environment. We'll model each with a qubit.

In the model a qubit has four ontic states ##1, 2, 3, 4##. However there is an epistemic limit so you can only resolve the ontic space to half the maximum limit. This means there are six maximum knowledge epistemic states:
$$
|\uparrow\rangle = \{1,2\} \\
|\downarrow\rangle = \{3,4\} \\
|+\rangle = \{1,3\} \\
|-\rangle = \{2,4\} \\
|i\rangle = \{2,3\} \\
|-i\rangle = \{1,4\} \\
$$
So for instance ##|\uparrow\rangle## is an epistemic state indicating the ontic state is ##1## or ##2##. Superposition of two epistemic states like
$$|\uparrow\rangle + |\downarrow\rangle = |+\rangle = \{1,3\}$$
can then be seen to not be "or" from Kolmogorov probability but a bilinear mapping from maximum knowledge epistemic states to maximum knowledge epistemic states where one ontic state from each is present in the pair of the superposition.

We also have states of non-maximal knowledge, mixed states, like:
$$\frac{1}{2}\mathbb{I} = \{1,2,3,4\}$$

Now let us consider an entangled state:
$$|\uparrow\uparrow\rangle + |\downarrow\downarrow\rangle = \{1.1, 2.2, 3.3, 4.4\}$$
where ##a.b## means the first particle is in ontic state ##a## and the second is in ontic state ##b##.

As you can see this means that the first particle is in one of the states:
$$\{1,2,3,4\}$$
which is a mixed state.

By gaining more knowledge of the entire system, i.e. that the two particles are always in the same ontic state, I know less about a single particle. I don't what state a single particle is in at all and thus have non-maximum knowledge of it, a mixed state.

Now consider the three particle state:
$$|\uparrow\uparrow\uparrow\rangle + |\downarrow\downarrow\downarrow\rangle = \{1.1.1, 1.2.2, 2.1.2, 2.2.1, 3.3.3, 3.4.4, 4.3.4, 4.4.3\}$$

Take the first particle to be the atomic system, the second to be the device and the third to be the lab environment. A superobserver outside the lab might use the above state. An observer within the lab, who doesn't track his lab environment (it's impossible otherwise he'd be a superobserver) might see his equipment read "up" in which case he knows that "atomic system + device" is in the state:
$$|\uparrow\uparrow\rangle = \{1.1, 1.2, 2.1, 2.2\}$$
which can easily be seen to be compatible with the above use of a superposition by the superobserver.

This is because states of maximum knowledge of the three particle system are consistent with states of maximum knowledge of subsystems that remove one of its "branches". This in turn is because one is reducing the systems tracked and thus one can increase knowledge of the subsystems.

This is essentially what happens in all epistemic models, but it's easier to see here in the toy model.

See Chapter 11 of Richard Healey's "The Quantum Revolution in philosophy" for a longer exposition on it in the ##\psi##-doxastic case.
 
Last edited:
  • Like
Likes dextercioby
  • #70
To follow on from the above, I recommend anybody to have a read of Spekkens paper:
https://arxiv.org/abs/quant-ph/0401052
It shows a bunch of "quantum" effects show up in a purely local classical theory with an epistemic limit, e.g.
  1. Superposition
  2. Non-commutativity of operators
  3. Uncertainty principle
  4. Interference
  5. Wigner's Friend ambiguities
  6. No cloning
  7. Superdense coding
  8. Indistinguishability of unknown states
  9. Collapse
  10. Entanglement Monogamy
  11. Teleportation
  12. Non-locality without entanglement (from https://arxiv.org/abs/quant-ph/9804053)
So it allows you to see what are "pure" quantum effects, i.e. Contextuality, Quantum Computing speedups and Non-classical correlations (e.g. CHSH violating correlations).
 
  • Like
Likes dextercioby and Demystifier

Similar threads

Back
Top