Is the collapse of the particle wave function?

In summary, the collapse of the wave function of the electron in the double slit experiment is not solely based on the act of observation, but can also be influenced by the physical interactions of the instrument used to measure the electron. This is supported by the delayed choice experiment. The term "collapse" is not used in a precise statement of quantum mechanics, and different interpretations such as the Bohmian or Many-Worlds approach may provide alternative explanations.
  • #1
Nick V
58
2
Is the collapse of the wave function of the electron in the double slit experiment based purely on the act of observation? Or could it be that the way the instrument used to measure the electron caused it to collapse by how it physically interacted with the electron? Keep in mind the delayed choice experiment.
 
Physics news on Phys.org
  • #2
Nick V said:
Is the collapse of the wave function of the electron in the double slit experiment based purely on the act of observation? Or could it be that the way the instrument used to measure the electron caused it to collapse by how it physically interacted with the electron? Keep in mind the delayed choice experiment.

QM says that the probability of the electron landing at a particular point on the screen is affected by any interactions (for example, with the slits, or with measuring devices behind the slits) it has on the way to the screen. There's nothing special about the subset of interactions that (for historical reasons) we call "observations"; if the interaction is such that we could in principle know which path the electron took, the effect of the interaction will be to eliminate the interference pattern.

That's pretty much a "no" answer to your first question and a "yes" to the second... but you should note that the word "collapse" does not appear in a precise statement of exactly what QM does and doesn't tell us.
 
  • Like
Likes vanhees71 and bhobba
  • #3
Nugatory said:
but you should note that the word "collapse" does not appear in a precise statement of exactly what QM does and doesn't tell us.

Are you allowing a Bohmian or Many-Worlds interpretation in this statement? The standard Copenhagen-style interpretation does include a mathematically precise statement that corresponds to what is colloquially called collapse, or more formally called state reduction or the conditional state or the a posteriori state.
 
  • #4
atyy said:
Are you allowing a Bohmian or Many-Worlds interpretation in this statement? The standard Copenhagen-style interpretation does include a mathematically precise statement that corresponds to what is colloquially called collapse, or more formally called state reduction or the conditional state or the a posteriori state.

That's a fair criticism, and I wouldn't even have made the comment if OP had said "state reduction" or similar.
 
  • #5
Nugatory said:
That's a fair criticism, and I wouldn't even have made the comment if OP had said "state reduction" or similar.

Well it's good to agree (I think) on the physics. I have come to realize that "collapse" is a loaded term, where Copenhagenists like me routinely mean "state reduction", while others think of something far more colourful.
 
  • #6
From a text of QM at Wiki: "The wave function in quantum mechanics evolves deterministically according to the Schrödinger equation as a linear superposition of different states (...)"
When we consider ALL the possible states we have a deterministic function, even if we don't know in advance what is the state of the system?
If we could know the state of the system and its evolution's equations, then we can write down deterministic equations with only those states? Without using probabilistic ones?
 
  • #7
afcsimoes said:
From a text of QM at Wiki: "The wave function in quantum mechanics evolves deterministically according to the Schrödinger equation as a linear superposition of different states (...)"
When we consider ALL the possible states we have a deterministic function, even if we don't know in advance what is the state of the system?
If we could know the state of the system and its evolution's equations, then we can write down deterministic equations with only those states? Without using probabilistic ones?

In the orthodox Copenhagen interpretation, time evolution is deterministic between measurements, and random at the instant of a measurement. Although this approach is consistent with all data to date, there problem with this approach is that the theory does not say when a measurement is made, which is something that the user of the theory must put in from outside.

This problem is solved in the Bohmian interpretation. In Bohmian mechanics, time evolution is determinstic, while the randomness lies in the initial conditions. However, there is not a unique Bohmian dynamics, and there is also no consensus yet about whether a Bohmian view is possible for all aspects of the standard model of particle physics.

A third approach called Many-Worlds is possibly completely deterministic. In this approach, all measurement outcomes occur. However, it is not yet known if this approach is correct in all details.
 
  • Like
Likes afcsimoes
  • #8
Nugatory said:
QM says that the probability of the electron landing at a particular point on the screen is affected by any interactions (for example, with the slits, or with measuring devices behind the slits) it has on the way to the screen. There's nothing special about the subset of interactions that (for historical reasons) we call "observations"; if the interaction is such that we could in principle know which path the electron took, the effect of the interaction will be to eliminate the interference pattern.

That's pretty much a "no" answer to your first question and a "yes" to the second... but you should note that the word "collapse" does not appear in a precise statement of exactly what QM does and doesn't tell us.
I think my question before is a little unclear but let me edit it...
When the electrons go through the double slit they create a wave like pattern. However, when there is a tracker to measure the electrons, they create a double band particle pattern. Could it be that the instrument used to measure the electrons is physically interacting with them causing them to collapse the wave like pattern, or is it just the act of observing that causes them to collapse the wave like pattern into a particle like pattern?
 
  • #9
Nick V said:
I think my question before is a little unclear but let me edit it...
When the electrons go through the double slit they create a wave like pattern. However, when there is a tracker to measure the electrons, they create a double band particle pattern. Could it be that the instrument used to measure the electrons is physically interacting with them causing them to collapse the wave like pattern, or is it just the act of observing that causes them to collapse the wave like pattern into a particle like pattern?

It is the interaction. But do remember that it is impossible to observe without interacting: all observations are interactions, but not all interactions are observations.
 
  • Like
Likes bhobba
  • #10
afcsimoes said:
When we consider ALL the possible states we have a deterministic function

Can you expand on what you mean because I can't follow it.

Its simple really. The state encodes within it the probabilities of the outcomes of observations. It evolves deterministically, but what it represents doesn't change.

Thanks
Bill
 
  • #11
Nugatory said:
It is the interaction. But do remember that it is impossible to observe without interacting: all observations are interactions, but not all interactions are observations.
Ok, but does the delay choice experiment play any part against the belief that the wave like nature of the electron in double slit collapses do to the instrument physically interacting with the electron, and that it is really just the act of observing that causes the wave like behavior to collapse?
Sorry if this is a stupid question.
 
  • #12
There's only one explanation for the delayed-choice experiments: quantum theory! It's not understandable with outdated concepts like wave-particle dualism. Delayed choice is just what it says: You prepare a lot of quantum systems in an appropriate state and make specific measurements on each of those. Then you can choose afterwards, just using the once and for all fixed measurement protocol, different subensembles showing either an interference pattern or imply which-way information as in the now pretty famous "quantum eraser" experiment by Walborn, et al.

See the explanation here:

https://www.physicsforums.com/threa...xperiment-affects-result.785711/#post-4935759
 
  • #13
vanhees71 said:

I guess we have discussed this many times, and I do not accept that the minimal interpretation does not have collapse or state reduction.

State reduction is undoubtedly part of the minimal interpretation, and is also known as the conditional state or the a posterior state. This is found in almost all standard textbooks, from Dirac, Landau & Lifshitz, Nieslen and Chuang, Weinberg, Sakurai & Napolitano, Shankar etc, and there really should be no controversy here.

Where there could be some controversy is whether the state reduction can be derived from unitary evolution and the Born rule alone. As far as I understand, it cannot, and requires an additional assumption. The main reason is that the Born rule only gives probabilities for a single time, whereas the conditional probability requires a rule for the joint probability of observations at different times. For spacelike observations, the Lorentz invariance can be used in place of the state reduction postulate, but the Lorentz invariance and spacelike separation is still an additional postulate.
 
  • #14
Again I disagree. State collapse is unnecessary and leads to conceptually to very problematic issues with relativistic causality. I'd ban the word "collapse" from any writings and talk about quantum theory (I hope I've not made myself guilty of this sin in my manuscripts yet, because I've also been exposed to this mantra of the otherwise brillant textbooks mentioned in the previous posting; I've to check :-)). All this we have discussed at length over the Christmas break :-).
 
  • Like
Likes atyy and bhobba
  • #15
Vanhees, would you mind clarifying for me? The bottom line is we only observe part of the WF. How can this be explained without either MWI or a collapse?
 
  • #16
We observe the wave function by doing experiments on ensembles of equally prepared systems, using the usual statistical technques to test a hypothesis, which in this case is provided by quantum theory. So what do you mean by the statement that we only observe part of the wave function? Whatever it is, why do you think, you need MWI or a collapse (which I consider as either esoterics that cannot be empirically tested or intrinsically inconsistent, respectively).
 
  • #17
Thanks for replying. I hope you don't mind if I rephrase my question. After I observe a single particle in a particular position, can the situation still be described, in theory, by a linear propagation of the original wf? If yes, the MWI follows. If not, why shouldn't this change be called a collapse?
 
  • #18
maline said:
Thanks for replying. I hope you don't mind if I rephrase my question. After I observe a single particle in a particular position, can the situation still be described, in theory, by a linear propagation of the original wf? If yes, the MWI follows. If not, why shouldn't this change be called a collapse?

The Ensemble interpretation is fully compatible with a number of different interpretations that say more about what going on so to speak eg Primary State Diffusion, Bohmian Mechanics, Nelson Stochastics, Many Worlds, and GRW collapse theory. The reason is its a minimalist interpretation making as few assumptions as possible - basically the assumption it makes is a frequentest interpretation of probability rather than a Baysian one which would lead to something like Copenhagen - it's what people mean when they say shut up and calculate - they mean an interpretation with very minimalist assumptions. Some have collapse and some do not. For definiteness I will pick Bohmian Mechanics. It does not have collapse - particles have well defined momentum, position etc at all times. What happens is when an observation is made decoherence occurs and you end up with the standard mixed state. However since it has definite properties the mixed state is a proper one and no collapse occurred. The ensemble interpretation is ambivalent to how the mixed state after decoherence becomes a proper mixed state - it simply assumes it is - somehow. That somehow may or may not involve collapse.

To understand the issue of proper and improper mixed states see the following:
http://philsci-archive.pitt.edu/5439/1/Decoherence_Essay_arXiv_version.pdf

Thanks
Bill
 
  • #19
bhobba said:
The Ensemble interpretation is fully compatible with a number of different interpretations that say more about what going on so to speak eg Primary State Diffusion, Bohmian Mechanics, Nelson Stochastics, Many Worlds, and GRW collapse theory. The reason is its a minimalist interpretation making as few assumptions as possible - basically the assumption it makes is a frequentest interpretation of probability rather than a Baysian one which would lead to something like Copenhagen - it's what people mean when they say shut up and calculate - they mean an interpretation with very minimalist assumptions. Some have collapse and some do not. For definiteness I will pick Bohmian Mechanics. It does not have collapse - particles have well defined momentum, position etc at all times. What happens is when an observation is made decoherence occurs and you end up with the standard mixed state. However since it has definite properties the mixed state is a proper one and no collapse occurred. The ensemble interpretation is ambivalent to how the mixed state after decoherence becomes a proper mixed state - it simply assumes it is - somehow. That somehow may or may not involve collapse.

To understand the issue of proper and improper mixed states see the following:
http://philsci-archive.pitt.edu/5439/1/Decoherence_Essay_arXiv_version.pdf

As far as I can tell, and as I have said many times, I don't disagree at all and I believe this is a correct Ensemble Interpretation. Although there are some difference in terminology, I cannot tell if it differs in any way in content from my default interpretation, which I usually call "Copenhagen" or "Copenhagen-style". When I say "collapse" I simply mean exactly what bhobba says by "The ensemble interpretation is ambivalent to how the mixed state after decoherence becomes a proper mixed state - it simply assumes it is - somehow."

So what is clear in bhobba's post is that there has to be something beyond unitary evolution and the Born rule in order to calculate the joint probability for sequential measurements. This clarity is lacking in Ballentine making his interpretation dubious at best, and as far as I can tell it is rejected by vanhees71.
 
  • #20
atyy said:
This clarity is lacking in Ballentine making his interpretation dubious at best, and as far as I can tell it is rejected by vanhees71.

I strongly disagree - but I will let Vanhees speak to his position.

Thanks
Bill
 
  • #21
bhobba said:
To understand the issue of proper and improper mixed states see the following:
http://philsci-archive.pitt.edu/5439/1/Decoherence_Essay_arXiv_version.pdf

As I said above, I agree with bhobba on physics, and at worst we only disagree on the interpretation of Ballentine. The linked article by Hansen is excellent, and as far as I can tell, I agree with it. Let me present Hansen's argument in a more mathematical way to make it clear that we have equations to discuss, and that the difference between proper and improper mixtures is not some vague interpretive thing.

First let's just assume unitary evolution and the Born rule, and no (the following are equivalent) collapse or state reduction or change from improper to proper mixtures. We have a quantum system and quantum ancilla. They start to interact at time=0, and finish interacting at time=t via unitary evolution. At time t, an observable A is measured on the ancilla and an observable B is measured on the system. Decoherence does not alter the picture for this argument, as it only restricts which observables A we can measure in a practical way. Since A and B are commuting observables that are measured at the same time, the Born rule gives P(A,B), and we can define the reduced density operator of the system by tracing over the ancilla, and we can define the conditional state or a posteriori state of the system using Bayes's rule. As long as we only calculate things derivable from P(A,B), the conditional state can be derived from unitary evolution and the Born rule alone, since it just amounts to writing Bayes's rule applied to P(A,B) in a different way. (See eg. http://arxiv.org/abs/0810.3536, Eq 6.7, 6.8 and Section 6.2.3.)

But if we use the conditional state or a posteriori state to calculate P(B,C), where C is a measurement on the system at a later time t+dt, then we are using the conditional state in a way that goes beyond applying Bayes's rule to P(A,B), and amounts to a postulate beyond unitary evolution and the Born rule, and this is collapse or state reduction or a change from an improper to proper mixture.

There is also very interesting discussion of the issue by Ozawa in http://arxiv.org/abs/quant-ph/9706027 and http://arxiv.org/abs/quant-ph/9711006. But I reference those papers only to note that as late as 1997, he writes that the consensus is that "partial trace does not derive state reduction", which is the same issue that bhobba brings up about improper and proper mixtures. Other texts that explicitly consider the issue and agree with bhobba are Haag https://www.amazon.com/dp/3540610499/?tag=pfamazon01-20 (1996, p301) and Haroche and Raimond https://www.amazon.com/dp/0198509146/?tag=pfamazon01-20 (2006, p82).
 
Last edited by a moderator:
  • #22
There is no general rule concerning what happens after the measurement of an observable. It depends on the experimental setup. If there is a measurement and you filter somehow the systems within the ensemble according to this measurement's result it's a form of state preparation, and you can calculate the probability of finding a value of an observable measured in a consecutive measurement. Then the first measurement is called a von Neumann filter measurement. Of course, usually you don't describe the filter process in full glory by using quantum theory. This is also an overkill, because measurement apparati are macroscopic systems and can be well described by classical or semiclassical means.

A good example is quantum optics. The quantum opticians on the one hand deal with single photons (or other many-photon Fock states like the entangled biphotons from parametric down conversion) and on the other with macroscopic optical devices as known from classical optics. The latter are usually described classically, e.g., birefringent crystals used for parametric down conversion via an index of refraction (linear response) and some non-linear "susceptibilities". The same holds for phase shifters, lenses, polarizers, beam splitters,... I don't think that a fully microscopic description exists and is possible in practice. Of course, it's fully sufficient to describe the measured effects, which are among the most fascinating and important findings about fundamental quantum physics made in recent decades!
 
  • #23
Okay, vanhees, I think I get it- please correct me if I misunderstood. You are saying that what we usually call a reduction can be thought of as simply a selection (and normalization) of those elements of the original wf that are relevant for predicting future observation probabilities. For the present observation, the Born rule can be defined without any reduction, as atyy explained. As far as what "actually happens" to the rest of the wf, you are taking the position that we have no reason to believe the wf "exists"; all we know is how it predicts probabilities, and we have no hope of ever interpreting this in terms of some underlying process because models that agree with QM, by definition, cannot make testable predictions of their own.
 
  • #24
maline said:
Okay, vanhees, I think I get it- please correct me if I misunderstood. You are saying that what we usually call a reduction can be thought of as simply a selection (and normalization) of those elements of the original wf that are relevant for predicting future observation probabilities. For the present observation, the Born rule can be defined without any reduction, as atyy explained. As far as what "actually happens" to the rest of the wf, you are taking the position that we have no reason to believe the wf "exists"; all we know is how it predicts probabilities, and we have no hope of ever interpreting this in terms of some underlying process because models that agree with QM, by definition, cannot make testable predictions of their own.

Just to be clear, I don't agree that in what is usually called a minimal interpretation that one can do without state reduction. Bhobba's "minimal interpretation" (I agree with bhobba, I put "minimal interpretation" in quotes just to clarify that his terminology is different from mine) explicitly mentions approaches like hidden variables or many-worlds as a possibility so that one can do without postulating state reduction, so that state reduction can be derived. However, there is no consensus that any of these can reproduce the standard quantum mechanical predictions for all known quantum mechanical predictions with state reduction or a change from an improper to proper mixture simply postulated as a phenomenological rule. What vanhees71 seems to be saying is we don't have to postulate hidden variables, many worlds, or state reduction or a change from an improper to proper mixture - that is strange. Certainly from unitary evolution and the Born rule alone, we can at least partly derive state reduction as a consistency requirement for measurements at spacelike separation - so state reduction is needed and standard in what is usually called a minimal interpretation.
 
  • #25
Why should I need something in addition to standard quantum theory? It's successful as it is! It predicts probabilities to an astonishing precision. There's no hint so far that there's something else than these probabilities. The wave function is the description of these probabilities. QT in this sense is just a special kind of description of a random process, no more no less.

Further I don't understand, what you mean by "part of a wave function" (I prefer to call what we discuss about the quantum state, because there are interesting things, where there is no wave-function description as for relativistic QT and photons particularly)? If I have an electron, described by a wave function (which makes sense in the non-relativistic approximation), I either detect it or not. In the latter case, I cannot say anything (not even, whether there was an electron at all). If I detect it and measure some of its observables, I just determine a fact that was either determined before through the state preparation or it was not, and then it's the outcome of a random experiment. The probabilities for the outcomes are predicted by the wave function describing the state, my electron was prepared in at the time of the measurement, and I can check this prediction in the usual way by repeating the preparation and measurement procedures independently for a many times and evaluate the findings with the usual mathematical tools of statistics. That's all. Nothing collapses. Or what collapses, if they draw the numbers of the weekly lottery?
 
  • #26
I've been reading through the thread with great interest. Please correct me if I'm wrong, but it seems to me that primary dispute here is regarding a fundamental conceptual difference as to what is being discussed. It's almost a semantic argument. Atyy seems to be taking the position that there is a "thing-in-itself" out there that is defined by the wave function (IS the wave function?), and therefore, in the absence of a Bohmian or Multiverse interpretation, the state reduction (collapse) concept is a necessary because it describes the "thing-in-itself" pre and post observation/measurement. In contrast, vanhees71 seems to be taking the position that QT says nothing about the "thing-in-itself" whatsoever, but is only a theory to describe the probability of an outcome to a specific type of measurement or observation. Is this even close to being accurate?
 
  • #27
That seems to be very accurate!
 
  • #28
Feeble Wonk said:
I've been reading through the thread with great interest. Please correct me if I'm wrong, but it seems to me that primary dispute here is regarding a fundamental conceptual difference as to what is being discussed. It's almost a semantic argument. Atyy seems to be taking the position that there is a "thing-in-itself" out there that is defined by the wave function (IS the wave function?), and therefore, in the absence of a Bohmian or Multiverse interpretation, the state reduction (collapse) concept is a necessary because it describes the "thing-in-itself" pre and post observation/measurement. In contrast, vanhees71 seems to be taking the position that QT says nothing about the "thing-in-itself" whatsoever, but is only a theory to describe the probability of an outcome to a specific type of measurement or observation. Is this even close to being accurate?

vanhees71 said:
That seems to be very accurate!

That is not accurate at all. The question is a mathematical statement. Collapse is the statement that after a measurement, we can use a classical measurement outcome to label a subensemble. In Ballentine, this is equation 9.28 and the statement preceding that this is the state of the sub-ensemble that is labelled by the previous measurement outcome. Ballentine does not call that state reduction, but that is only a matter of terminology since other people do call it state reduction or collapse. So yes, Ballentine does have collapse. So the question is can that be derived from unitary evolution and the Born rule alone?
 
  • #29
vanhees71 said:
That seems to be very accurate!
Your position would seem to be the ultimate in the "shut up and calculate" school of thought regarding quantum physics. Does that mean that you believe it is impossible to interpret the findings of QT with regard to what it implies about the fundamental nature of physical existence? Or is it simply your preference on how to approach the subject.
 
  • #30
Feeble Wonk said:
Your position would seem to be the ultimate in the "shut up and calculate" school of thought regarding quantum physics.

Let's be clear the difference between vanhees71 and me is not "shut up and calculate". In fact in a sense my view is more shut up and calculate. If one shuts up and calculates, there is absolutely nothing wrong with collapse or state reduction.

I am simply defending traditional shut up and calculate :)
 
  • #31
atyy said:
That is not accurate at all. The question is a mathematical statement. Collapse is the statement that after a measurement, we can use a classical measurement outcome to label a subensemble. In Ballentine, this is equation 9.28 and the statement preceding that this is the state of the sub-ensemble that is labelled by the previous measurement outcome. Ballentine does not call that state reduction, but that is only a matter of terminology since other people do call it state reduction or collapse. So yes, Ballentine does have collapse. So the question is can that be derived from unitary evolution and the Born rule alone?
The question is a physical statement. What Ballentine describes with his Eq. (9.28) is the Statistical operator after a von Neumann "filter measurement", which I'd consider a state-preparation procedure. For those who don't have available this excellent textbook, here's the statement

Suppose an ensemble of a quantum mechanical system is prepared such that it is discribed by the statistical operator ##\hat{\rho}##. Then a measurement of an observable ##R##. He takes an observable with a continuous spectrum. Let ##|r,\beta \rangle## denote the generalized eigenvectors, normalized to a ##\delta## distribution and ##\beta## labeling a possible degeneracy, which I take as a discrete set (you can also have a continuous set, but that doesn't change much). Then he considers a state-preparation procedure which filters out all systems out of the ensemble for which ##R## takes a value in an interval ##\Delta_a##. The remaining sub-ensemble is then described by a new statistical operator, which is constructed as follows: We first define the projection operator
$$\hat{M}_R(\Delta_a)=\int_{\Delta_a} \mathrm{d} r \sum_{\beta} |r,\beta \rangle \langle r,\beta|$$
and then define
$$\hat{\rho}'=\frac{1}{Z} \hat{M}_R(\Delta_a) \hat{\rho} \hat{M}_R(\Delta_a), \quad Z=\mathrm{Tr} \left [\hat{M}_R(\Delta_a) \hat{\rho} \hat{M}_R(\Delta_a)
\right]$$
as the statistical operator of the corresponding sub-ensemble.

This is just a definition what you call preparation through an idealized von Neumann filter measurement. An example, where you can describe this entirely with quantum theory is the Stern-Gerlach experiment. In this example there's then no collapse but just the ignorance of all particles with the spin component not in the desired state. The "collapse" in the Copenhagen doctrine is nothing but the choice of the subensemble by the experimenter through filtering out all undesired complementary subensembles. There's no spooky non-quantum action-at-a-distance magic called "collapse"!
 
  • #32
vanhees71 said:
The question is a physical statement. What Ballentine describes with his Eq. (9.28) is the Statistical operator after a von Neumann "filter measurement", which I'd consider a state-preparation procedure. For those who don't have available this excellent textbook, here's the statement

Suppose an ensemble of a quantum mechanical system is prepared such that it is discribed by the statistical operator ##\hat{\rho}##. Then a measurement of an observable ##R##. He takes an observable with a continuous spectrum. Let ##|r,\beta \rangle## denote the generalized eigenvectors, normalized to a ##\delta## distribution and ##\beta## labeling a possible degeneracy, which I take as a discrete set (you can also have a continuous set, but that doesn't change much). Then he considers a state-preparation procedure which filters out all systems out of the ensemble for which ##R## takes a value in an interval ##\Delta_a##. The remaining sub-ensemble is then described by a new statistical operator, which is constructed as follows: We first define the projection operator
$$\hat{M}_R(\Delta_a)=\int_{\Delta_a} \mathrm{d} r \sum_{\beta} |r,\beta \rangle \langle r,\beta|$$
and then define
$$\hat{\rho}'=\frac{1}{Z} \hat{M}_R(\Delta_a) \hat{\rho} \hat{M}_R(\Delta_a), \quad Z=\mathrm{Tr} \left [\hat{M}_R(\Delta_a) \hat{\rho} \hat{M}_R(\Delta_a)
\right]$$
as the statistical operator of the corresponding sub-ensemble.

This is just a definition what you call preparation through an idealized von Neumann filter measurement. An example, where you can describe this entirely with quantum theory is the Stern-Gerlach experiment. In this example there's then no collapse but just the ignorance of all particles with the spin component not in the desired state. The "collapse" in the Copenhagen doctrine is nothing but the choice of the subensemble by the experimenter through filtering out all undesired complementary subensembles. There's no spooky non-quantum action-at-a-distance magic called "collapse"!

The reason that this is not preparation only is that it is preparation based on the outcome of the preceding measurement. The new quantum state is prepared based on the previous outcome that ##R## takes a value in an interval ##\Delta_a##, so it is a conditional preparation, and is a conditional state. By saying that the new state is prepared by conditioning on the preceding outcome, one has in effect generalized the Born rule for the preceding outcome to the new state.
 
Last edited:
  • #33
Sure, it's a conditional preparation. That's what all preparations are. I think it's implied by the Born rule, but as I said, in the case of the Stern-Gerlach experiment, you can derive it from the dynamics. You just consider only a partial beam after one Stern Gerlach apparatus splitting up the incoming beam into partial beams with determined magnetic quantum number of its spin.
 
  • #34
atyy, my interest in this debate has receded a bit but in our recent discussion of the double slit you linked to a nice computer simulation which seems relevant to me.

A gaussian wave packet hits the double slit. There, a part of the wave packet is transmitted and propagated to the screen and a part is reflected. The reflected part never hits the screen. So for a position measurement at the screen, we get the same probabilities for two different states: (a) the superposition of the transmitted and the reflected part of the wave packet and (b) the transmitted part of the wave packet alone.

Would you say that QM needs a postulate for the case that we chose description (b) for calculations because we don't want to calculate things which aren't relevant for the measurement outcome?
 
Last edited:
  • #35
No, you just only consider that part of particles hitting the screen. There's no addition to standard QT needed whatsoever to understand this experiment.
 
  • Like
Likes bhobba
Back
Top