Evaluate this paper on the derivation of the Born rule

In summary, The paper discusses the Curie Wiess model of the quantum measurement process and how it can be used to derive the Born rule.
  • #1
Prathyush
212
16
I have encountered this paper "Curie Wiess model of the quantum measurement process". https://arxiv.org/abs/cond-mat/0203460

Another work by the same authors is "Understanding quantum measurement from the solution of dynamical models" https://arxiv.org/abs/1107.2138

I am still evaluating the papers. I find the general lessons implied to be interesting and probably compelling.

In the context of the model studied is this paper accurate? What do you think about the overarching viewpoint presented by the authors?
 
Physics news on Phys.org
  • #2
Prathyush said:
I have encountered this paper "Curie Wiess model of the quantum measurement process". https://arxiv.org/abs/cond-mat/0203460

Another work by the same authors is "Understanding quantum measurement from the solution of dynamical models" https://arxiv.org/abs/1107.2138

I am still evaluating the papers. I find the general lessons implied to be interesting and probably compelling.

In the context of the model studied is this paper accurate? What do you think about the overarching viewpoint presented by the authors?

I think it's great that they made such a careful analysis of a measurement process. But it seems to me a little circular to use the ensemble interpretation of QM to derive the Born rule.
 
  • #3
The difficulty that people generally perceive in the Born rule arises because of the historical association of the abstract idea of probability with relative frequency. The reason for the association is obvious because classically relative frequency is the natural practical and historical way we have to measure probability. So they assume that "probability" must be non-negative and, in particular, equate it to the normalized relative frequency.

But a moment's thought should tell us that any mathematical encoding that tells us how to compute the relative frequency can serve as a theoretical probability.

When it comes to QM, the scalar product of a "final" (i.e. outcome) state with (i.e. conditional on) the initial state is a natural encoding of probability because it is largest (i.e. most likely) when the final state vector is closest (in the usual scalar product of vectors sense) to the initial state vector.

Of course there remains a little mystery as to why the squared modulus is the correct way to compute the relative frequency, although it does seem to be the most straightforward.

But my point is that that little mystery is the entire content of the Born rule, once we have a Hilbert space because the association of theoretical probability with the scalar product is built into the association of states with Hilbert space vectors.
 
Last edited:
  • #4
mikeyork said:
But my point is that that little mystery is the entire content of the Born rule, once we have a Hilbert space because the association of theoretical probability with the scalar product is built into the association of states with Hilbert space vectors.
Certainly not.

If I use a Hilbert space setting to analyze for example the multipole expansion in classical electrodynamics, the mathematics is the same as in quantum mechanics, but there is no probability at all. The scalar product has a different meaning, and this this meaning does not come with the mathematical framework. The meaning is due to our interpretation of the mathematical framework, and this interpretation is different in classical electrodynamics and in quantum mechanics (and it is different in the various different interpretations of quantum mechanics).
 
  • #5
200 pages!
 
  • #6
tom.stoer said:
Certainly not.

If I use a Hilbert space setting to analyze for example the multipole expansion in classical electrodynamics, the mathematics is the same as in quantum mechanics, but there is no probability at all. The scalar product has a different meaning, and this this meaning does not come with the mathematical framework. The meaning is due to our interpretation of the mathematical framework, and this interpretation is different in classical electrodynamics and in quantum mechanics (and it is different in the various different interpretations of quantum mechanics).
I didn't say that the scalar product has that meaning in any Hilbert space. I said it comes from the association of physical states with state vectors.

Any function p serves as a probability encoding as long as there is some single-valued function f(p) that predicts the asymptotic relative frequency. In QM, the scalar product serves this purpose in a very natural way (with f(p) monotonic in |p|) because of the meaning we attach to state vectors. It is the same as saying that big changes are less likely than small changes -- a fairly obvious, perhaps even tautological, notion. The content of the Born rule lies with the simplicity of f(p) = |p|^2.
 
  • #7
stevendaryl said:
I think it's great that they made such a careful analysis of a measurement process. But it seems to me a little circular to use the ensemble interpretation of QM to derive the Born rule.

Why do you think the reasoning may be circular? I am in stage where I am still evaluating the arguments presented in the paper, assuming they are correct, here is what I think about the general context. We understand exactly how our measurement apparatus was constructed, it is reasonable to suppose that we use a statistical ensemble describe it. At a microscopic level we understand how degrees of freedom of the apparatus interact with system. We want to understand the physics that describes the macroscopic changes that the apparatus sees when it interacts with the system. The paper appears to precisely do this, and arrives with Born's rule as a consequence of this analysis. I see no circularity in this reasoning.

martinbn said:
200 pages!

The original paper is 5 pages. You may find "lectures on dynamical models for quantum measurements" useful https://arxiv.org/abs/1406.5178 Which is in lies in between at 45 pages. :)
 
  • #8
Prathyush said:
Why do you think the reasoning may be circular? I am in stage where I am still evaluating the arguments presented in the paper, assuming they are correct, here is what I think about the general context. We understand exactly how our measurement apparatus was constructed, it is reasonable to suppose that we use a statistical ensemble describe it. At a microscopic level we understand how degrees of freedom of the apparatus interact with system. We want to understand the physics that describes the macroscopic changes that the apparatus sees when it interacts with the system. The paper appears to precisely do this, and arrives with Born's rule as a consequence of this analysis. I see no circularity in this reasoning.

Well, describing ensembles using density matrices seems to me to be reliant on the Born rule. The point of a density matrix is that if [itex]\rho[/itex] is a density matrix, and [itex]A[/itex] is some operator, then the expectation value for [itex]A[/itex] is given by:

[itex]\langle A \rangle = tr(\rho A)[/itex]

Now write [itex]\rho[/itex] in terms of a complete orthonormal basis [itex]|\psi_n\rangle[/itex]:

[itex]\rho = \sum_n p_n |\psi_n\rangle \langle \psi_n | [/itex]

Then the above definition of [itex]\langle A \rangle[/itex] is equivalent to:

[itex]\langle A \rangle = \sum_n p_n \langle \psi_n | A | \psi_n \rangle[/itex]

That says that for a pure state [itex]|\psi_n\rangle[/itex],

[itex]\langle A \rangle = \langle \psi_n | A | \psi_n \rangle[/itex]

That definition for the expectation value of an operator when evaluated in a pure state seems to me equivalent to the Born rule.

Actually, now that I think about it, the above expression for expectation values is equivalent to the Born rule if we make the additional assumption that a measurement always produces an eigenvalue of the operator corresponding to the observable being measured. So maybe the point is that by considering the composite system (system being measured + measuring device), the fact that you always get an eigenvalue is derivable?
 
  • Like
Likes Leo1233783 and Jilang
  • #9
stevendaryl said:
Well, describing ensembles using density matrices seems to me to be reliant on the Born rule.
stevendaryl said:
That definition for the expectation value of an operator when evaluated in a pure state seems to me equivalent to the Born rule.

I think I understand what you are saying but I am not entirely sure. I will paraphrase you, to be sure.
When we define a state
[tex] |\psi> = c_1 |\psi_1> + c |\psi_2> [/tex]
There is an implicit use of the born rule because by definition a state vector is equivalent to the probabilities obtained when an ensemble of measurements are performed.
Yes that is true, but that is not what we are trying to investigate here. As soon as we introduce operators into quantum mechanics, in a sense one can say that Born's rule comes for free.

What we are(or I am) interested in investigating here is if can one understand the use of the Born's rule when we study the interaction between the experimental apparatus and the system.

stevendaryl said:
Actually, now that I think about it, the above expression for expectation values is equivalent to the Born rule if we make the additional assumption that a measurement always produces an eigenvalue of the operator corresponding to the observable being measured. So maybe the point is that by considering the composite system (system being measured + measuring device), the fact that you always get an eigenvalue is derivable?

Yes it is an expectation that measurement will produce an eigenvalue of the operator corresponding to be observable being measured. It has be shown by using a context. It may be a very difficult problem to show this in a sufficiently general context and would require a lot of mathematical sophistication.(even if the conclusion we arrived at is wrong it would require thorough investigation). What the authors have done is used a reasonable caricature of a real life situation and have show this. (assuming the derivation works through)
 
Last edited:
  • #10
It is very popular approach to represent quantum measurement as interaction with single device that can leave "measurement device" in different states depending on something. This paper in OP follows the same approach.
But I can not see any correspondence to real experimental setups. In all real setups that I know about there is some sort of manipulation device that does not produce any readings and detector that simply amplifies energy deposited in detector.
So I'm rather mystified why something sensible is expected from approach that is so detached from reality.
 
  • Like
Likes Leo1233783
  • #11
zonde said:
But I can not see any correspondence to real experimental setups. In all real setups that I know about there is some sort of manipulation device that does not produce any readings and detector that simply amplifies energy deposited in detector.
What you have written here is sufficiently vague, perhaps a clarification is needed. What is this manipulation device you are talking about? What are the principles behind the construction of an amplification apparatus? How do we understand amplification in terms of the behaviour of microscopic degrees of freedom.

What I am interested in doing is to understand the interaction apparatus and the system using formal mathematical tools. For this one has to make precise notions such as what it means to record information, amplification, macroscopic states etc. I am looking for a simple system which is sufficiently complex to display a measurement like phenomenon and provide an explanation of these concepts.

zonde said:
So I'm rather mystified why something sensible is expected from approach that is so detached from reality.

One can study other more complicated like cloud chambers or photo multiplier tubes but these systems are too complex for me to analyze. The essential details of a real life example can be captured using simple models. Once we can understand a caricature of a real life example by changing the details we can understand more complicated situations. To me this model looks like a reasonable caricature.
 
  • #12
Prathyush said:
What you have written here is sufficiently vague, perhaps a clarification is needed. What is this manipulation device you are talking about? What are the principles behind the construction of an amplification apparatus? How do we understand amplification in terms of the behaviour of microscopic degrees of freedom.
In that paper spin of particle is tested. So the manipulation device would be SG apparatus. If we are testing ion in a trap, manipulations are performed with radiation pulses of different wavelength. If we test beam of photons, we use filters and (polarization) beam splitters.
Amplification is done in detector by classical process. In threshold detector one electron triggers avalanche of other electrons over potential barrier.
Prathyush said:
What I am interested in doing is to understand the interaction apparatus and the system using formal mathematical tools. For this one has to make precise notions such as what it means to record information, amplification, macroscopic states etc. I am looking for a simple system which is sufficiently complex to display a measurement like phenomenon and provide an explanation of these concepts.
Mathematical tools are required when you want quantitative details of the model. But before you start working out quantitative details you need qualitative understanding of the model. I don't see how math can help you there except for getting some rough estimates.
Prathyush said:
One can study other more complicated like cloud chambers or photo multiplier tubes but these systems are too complex for me to analyze. The essential details of a real life example can be captured using simple models. Once we can understand a caricature of a real life example by changing the details we can understand more complicated situations. To me this model looks like a reasonable caricature.
What is caricature of photographic plate type device in this model? Photographic plate does not make different spots on it depending on electrons spin.
The caricature of detector is device that takes whatever particle as input and produces "click" in the output by classical process.
 
  • #13
I largely agree with the qualitative assessments you have made about the various apparatus(barring ion traps which I am unfamiliar with). I will examine details you have provided.

zonde said:
Amplification is done in detector by classical process. In threshold detector one electron triggers avalanche of other electrons over potential barrier.

I don't understand in what sense you are using the word "classical process"? You have used this term twice, and it requires clarification. Certainly whatever this process you are referring to is, It can be understood in terms of microscopic details using an appropriate statistical ensemble and Hamiltonian dynamics right?

zonde said:
What is caricature of photographic plate type device in this model? Photographic plate does not make different spots on it depending on electrons spin.
The caricature of detector is device that takes whatever particle as input and produces "click" in the output by classical process.

This model is in no way a caricature of the photographic plate in any direct way. I don't yet know what a good caricature of a photographic plate is, certainly analyzing Ag and Br atoms is too complex. Now that you asked this question this model is not a caricature of any measurement apparatus that I am aware of in any direct way.

However It is a caricature in the sense that it has some salient features shared by other measurement apparatus. Consider a cloud chamber for instance, we understand that upon the interaction of the charged particles with the water molecules a phase transition happens.(It turns from transparent to cloudy). This models also shares the property that the measurement apparatus is prepared in a metastable state and when it interacts with the state, it changes into one of its 2 ground states(ferromagnetic states). So the information about the state of system is amplified into something that is macroscopically observable. So in a very loose sense I say this is a caricature of the measurement phenomenon.(perhaps not any specific apparatus)
 
  • #14
I refereed the paper in question here.

stevendaryl said:
⟨A⟩=⟨ψn|A|ψn⟩⟨A⟩=⟨ψn|A|ψn⟩\langle A \rangle = \langle \psi_n | A | \psi_n \rangle

That definition for the expectation value of an operator when evaluated in a pure state seems to me equivalent to the Born rule.
In which way??

The formula you stated is just a simple mathematically valid statement about certain formulas in a Hilbert space.

Whereas Born's rule claims that the formula has a relation with theoretically ill-defined, complex physical processes (measurement) and philosophically loaded notions (probability). It is these two properties that make the incorporation of Born's rule into any foundation questionable.
 
  • #15
A. Neumaier said:
In which way??

The formula you stated is just a simple mathematically valid statement about certain formulas in a Hilbert space.

No, it's not. "Expectation value" for a measurable quantity means the limiting average of the quantity over many measurements (sometimes it means measurements of different members of an ensemble, or possible many measurements of the same system over time). That's not derivable from the mathematics of Hilbert space.
 
  • #16
stevendaryl said:
No, it's not. "Expectation value" for a measurable quantity means the limiting average of the quantity over many measurements (sometimes it means measurements of different members of an ensemble, or possible many measurements of the same system over time). That's not derivable from the mathematics of Hilbert space.
No. Expectation value means the value of a prescribed positive linear functional, nothing more. It is defined for arbitrary random variables, independent of the possibility or not of measuring them.

The use made of it in the statistical mechanics of a grand canonical ensemble, say, confoirms to this. One measures a single time a single number, for example the mass of all particles together, and gets a number very well predicted by ##\langle M\rangle## figuring in thermodynamics as derived from statistical mechanics.
 
  • #17
A. Neumaier said:
No. Expectation value means the value of a prescribed positive linear functional, nothing more.

That's quibbling over definitions. My point is that the identification of [itex]\langle \psi|A|\psi\rangle[/itex] with the average value of [itex]A[/itex] over many measurements is equivalent to the Born rule, or at least is very closely related. What you call the terms is not very interesting.

The use made of it in the statistical mechanics of a grand canonical ensemble, say, confoirms to this. One measures a single time a single number, for example the mass of all particles together, and gets a number very well predicted by [itex]\langle M \rangle[/itex] figuring in thermodynamics as derived from statistical mechanics.

That doesn't make any sense. If [itex]\langle M \rangle[/itex] is just a mathematical expression denoting the value of a positive linear functional, then it's not a prediction, at all. To relate [itex]\langle M \rangle[/itex] to measurements requires a physical hypothesis.
 
  • #18
stevendaryl said:
My point is that the identification of [itex]\langle \psi|A|\psi\rangle[/itex] with the average value of [itex]A[/itex] over many measurements is equivalent to the Born rule, or at least is very closely related.

But in statistical thermodynamics (which is the most successful application of the expectation calculus of statistical mechanics) one never makes this identification!

Instead one identifies the expectation value of the mass, say, with the single, macroscopic value of the mass! Each single measurement agrees in this way with the predictions of statistical mechanics! Probabilities arise only when a system is so small that uncertainties become significant, and many repetitions are needed to reduce the uncertainty by taking an average over many measurements. But microscopically this is then ##K^{-1}\sum_{k=1}^K\langle A_k\rangle## where ##A_k## is the observable in the ##k##th measurement, and not ##\langle A\rangle##!
 
  • #19
A. Neumaier said:
But in statistical thermodynamics (which is the most successful application of the expectation calculus of statistical mechanics) one never makes this identification!

Instead one identifies the expectation value of the mass, say, with the single, macroscopic value of the mass! Each single measurement agrees in this way with the predictions of statistical mechanics! Probabilities arise only when a system is so small that uncertainties become significant, and many repetitions are needed to reduce the uncertainty by taking an average over many measurements. But microscopically this is then ##K^{-1}\sum_{k=1}^K\langle A_k\rangle## where ##A_k## is the observable in the ##k##th measurement, and not ##\langle A\rangle##!

I'm not getting anything other than irritation from this discussion. I don't understand what point you're making.
 
  • #20
stevendaryl said:
I'm not getting anything other than irritation from this discussion. I don't understand what point you're making.
You make a silent but incorrect assumption about the interpretation of the statistical mechanics calculus, namely that ##\langle A\rangle:=\langle\psi|A\psi\rangle## is always interpreted as the average value of A over many measurements. But this interpretation is almost never valid. You can see this by comparing what is actually measured in thermodynamics, and how the thermodynamic formulas are derived from statistical mechanics. In each case you'll find that the extensive macroscopic variables (of a single thermodynamic system such as a single brick of gold) are encoded in statistical thermodynamics as expectation values of the corresponding microscopic operators. No repeated measurements, no averages anywhere!
 
  • #21
stevendaryl said:
I'm not getting anything other than irritation from this discussion. I don't understand what point you're making.

I will try and clarify what Arnold is saying(please correct me if I am wrong). I am not making any assumptions at this moment about its possible correctness. The sense in which we use Born's rule currently imply the existence of ideal measurement apparatus that work. However if one has to carefully analyze this situation in depth, i.e if we want to understand exactly how an ideal measurement is made by construction, we must refer to arrangements that must be described using macroscopic observables. And measurements made by these "ideal apparatus" ultimately refer to macroscopic changes of the apparatus. For instance what an electron gun or a cloud chamber is can be described using language that only refers to macroscopic observables.
 
Last edited:
  • #22
A. Neumaier said:
You make a silent but incorrect assumption about the interpretation of the statistical mechanics calculus, namely that ##\langle A\rangle:=\langle\psi|A\psi\rangle## is always interpreted as the average value of A over many measurements. But this interpretation is almost never valid. You can see this by comparing what is actually measured in thermodynamics, and how the thermodynamic formulas are derived from statistical mechanics. In each case you'll find that the extensive macroscopic variables (of a single thermodynamic system such as a single brick of gold) are encoded in statistical thermodynamics as expectation values of the corresponding microscopic operators. No repeated measurements, no averages anywhere!
That is a really interesting point you make. However, can you clarify how expectation in this statistical mechanics sense relates physically to expectation in QM in a way that does not imply averaging? I understand the mathematical point you make; it is the physical interpretation I have trouble with. The QM interpretation in the sense of averaging is clear. What sort of measurement of a quantum expectation would not involve averaging?
 
  • #23
A. Neumaier said:
You make a silent but incorrect assumption about the interpretation of the statistical mechanics calculus, namely that [tex]⟨A⟩:=⟨ψ|Aψ⟩\langle A\rangle:=\langle\psi|A\psi\rangle [/tex] is always interpreted as the average value of A over many measurements. But this interpretation is almost never valid. You can see this by comparing what is actually measured in thermodynamics, and how the thermodynamic formulas are derived from statistical mechanics. In each case you'll find that the extensive macroscopic variables (of a single thermodynamic system such as a single brick of gold) are encoded in statistical thermodynamics as expectation values of the corresponding microscopic operators. No repeated measurements, no averages anywhere!

Now I don't quite understand why you are talking about statistical mechanics, and how it can address stevendaryl's question about the use of Born rule in quantum mechanics. Certainly when we consider the stern gerlach experiment [tex]<\sigma_z> = p_{up}+p_{down} [/tex] is a sum over repeated measurements.
 
Last edited:
  • #24
Well, what's measured on macroscopic systems usually are indeed very "coarse-grained observables", i.e., e.g., the position and momentum of a macroscopic body is something like the center-of-mass (relativistically center-of-energy) position and the macroscopic momentum the total momentum of the many-body system. Very often (particularly in situations when the system is close to an equilibrium state) in such cases the measured values are close to the expectation value, because the quantum (and even the thermal!) fluctuations of such macroscopic ("collective") observables (i.e., the corresponding standard deviations) are small compared to the macroscopically relevant scales. That's why macroscopic systems very often behave "classically" concerning the macroscopic observables.

I haven't looked on the 200p detailed treatise and I cannot say, whether the authors really prove Born's rule from only the dynamical laws of QT. It's hard to conceive, how this should be done without using the probability interpretation of the quantum formalism, which basically is Born's rule. It's quite probable that the authors somehow use Born's rule to prove it, but as I said, I cannot claim that, because I haven't the time to study their thesis carefully.
 
  • Like
Likes dlgoff
  • #25
Prathyush said:
I don't quite understand why you are talking about statistical mechanics
Because the methods used are those of statistical mechanics. All is about the behavior of a macroscopic system (which can be analyzed only through statistical mechanics) interacting with a microscopic one. And because st
stevendaryl said:
But it seems to me a little circular to use the ensemble interpretation of QM to derive the Born rule.
stevendaryl said:
describing ensembles using density matrices seems to me to be reliant on the Born rule.
is based on a misinterpretation of how statistical mechanics relates to observation. I tried to correct this misinterpretation, but it is so deeply ingrained in stevendaryl's thinking that the information I provided only confuses him.
Prathyush said:
Certainly when we consider the stern gerlach experiment [...]
is a sum over repeated measurements.​
Yes, but s single spin is a tiny system, whereas statistical mechanics is about the behavior of macroscopic systems, in our case of the measurement device. Measuring a macroscopic system is not governed by Born's rule but by the identification of the expectation value with the measured value (within the intrinsic uncertainty). Born's rule is valid only for complete measurements of very small quantum systems such as a Stern-Gerlach spin.
 
  • Like
Likes bhobba
  • #26
mikeyork said:
What sort of measurement of a quantum expectation would not involve averaging?
The sort of measurement done in thermodynamics and in ordinary life. One measures a macroscopic system once and gets (if the measurement is done carefully) an accurate value for the measured variable that agrees with the expectation value calculated from statistical mechanics. No repetition is necessary, no averaging is involved. Repeated measurement gives essentially the same value, within the intrinsic (small) uncertainty.
 
  • #27
A. Neumaier said:
Because the methods used are those of statistical mechanics. All is about the behavior of a macroscopic system (which can be analyzed only through statistical mechanics) interacting with a microscopic one. And because st

is based on a misinterpretation of how statistical mechanics relates to observation. I tried to correct this misinterpretation, but it is so deeply ingrained in stevendaryl's thinking that the information I provided only confuses him.

Yes, but s single spin is a tiny system, whereas statistical mechanics is about the behavior of macroscopic systems, in our case of the measurement device. Measuring a macroscopic system is not governed by Born's rule but by the identification of the expectation value with the measured value (within the intrinsic uncertainty). Born's rule is valid only for complete measurements of very small quantum systems such as a Stern-Gerlach spin.
I'm puzzled how you come to this conclusion. In many-body physics we describe the system's state as any other quantum system by a statistical operator ##\hat{\rho}## and define the expectation value of an observable, represented by a self-adjoint operator ##\hat{A}## by
$$\langle A \rangle_{\rho}=\mathrm{Tr}(\hat{\rho} \hat{A}).$$
This is Born's rule.

What you usually measure for macroscopic systems in a "classical setup" are not arbitrary expectation values of this kind, but relevant "pretty coarse-grained" observables. E.g., for an ideal gas you consider some volume and check quantities like density, pressure, temperature etc. (assuming you have a state in or close to thermal equilibrium). These are quantities describing a large number of particles in a "collective way". The density, e.g., is just counting particle numbers in a volume element containing many particles and then dividing by this volume. The volume is microscopically large, i.e., it' contains many particles, but on the other hand it's macroscopically small, i.e., the properties of the coarse-grained variable don't change much on the scale of the volume. Experience shows that such a "separation of scales" for such coarse-grained variables occurs quite often for macroscopic systems and that makes the success of the thermodynamical approach. You can derive the macroscopic equations like the (Vlasov-)Boltzmann(-Uehling-Uhlenbeck) from this approach (formalizing the "coarse graining procdedure" either by applying the gradient-expansion or some projection method) or further viscous or even ideal hydrodynamics.

There is some community of quantum physicists who likes to try to derive Born's rule from the other postulates, among them Weinberg with his newest textbook, as an attempt to solve an apparent "measurement problem". As far as I understand what they mean by this "problem" is, why a measurement leads to a certain outcome although according to quantum theory, if the system is not prepared such that the measured observable is determined, there are only probabilities to find a certain value, given by Born's rule. The idea behind these attempts seems to be to be able to derive how the certain outcome of the observable's value occurs when measured with these probabilities from the dynamical postulates alone. The problem is of course that you get a circular argument, and Weinberg finally comes to the conclusion that he cannot satisfactorily derive Born's rule from the other postulates. That doesn't mean that not some other clever physicist comes along and finds a convincing set of postulates for QT, from which you can derive Born's rule. Whether or not you find this line of argument convincing or not is subjective. I don't see any merit in such an attempt, except if you find something really new. The paradigmatic example where such a methodology opened a whole new universe (for general relativists quite literally) of mathematical thinking is the attempt to prove the axiom about parallels of Euclidean geometry from the other postulates. It lead to the discovery of non-Euclidean geometries and a plethora of new "geometrical ideas", including the group-theoretical approach a la Klein, which is so vital for modern physics through Noether's theorems and all that.

In the minimal interpretation there is no measurement problem. For the Stern-Gerlach experiment, i.e., measuring a spin component (usually chosen as ##\sigma_z##) of a spin-1/2 particle, it's pretty simple to explain everything quantum theoretically, because the system is so simple that you can do so. The experiment consists more or less just of an appropriately chosen static magnetic field, which has a large homogeneous component in ##z## direction and also some inhomogeneity (also along the ##z## direction, but of course one has necessarily another in another direction due to Gauss's Law ##\vec{\nabla} \cdot \vec{B}=0##; one simple model for such a field is ##\vec{B}=(B_0 + \beta z)\vec{e}_z-\beta y \vec{e}_y##. Then it's quite easy to show that if ##\beta |\langle y \rangle| \ll B_0## that the particle beam splits into two quite well separated partial beams which are almost 100% sorting the two possible states ##\sigma_z = \pm 1##. That's how you measure (or even prepare!) nearly pure ##\sigma_z## eigenstates by entangling position and ##\sigma_z##. There's nothing mysterious about this, but of course this argument uses the full standard QT postulates, including Born's rule.
 
  • Like
Likes dextercioby
  • #28
A. Neumaier said:
I refereed the paper in question here.
Thank you for the review. Just some nit-picking:
A. Neumaier said:
200 page treatise (from 2011, available as an arXiv preprint only) containing all the details:
A.E. Allahverdyan, R. Balian and T.M. Nieuwenhuizen,
Understanding quantum measurement from the solution of dynamical models,
https://arxiv.org/abs/1107.2138
(201 pages)
This work was published in Phys. Rep. 525 (2013) 1-166
 
  • #29
I largely agree with the analysis done by Vanhees71 in post #27.

vanhees71 said:
The problem is of course that you get a circular argument, and Weinberg finally comes to the conclusion that he cannot satisfactorily derive Born's rule from the other postulates. That doesn't mean that not some other clever physicist comes along and finds a convincing set of postulates for QT, from which you can derive Born's rule.

I don't think any new postulate is required, I think it would only require a careful analysis of what we mean by measurement. As I suggested (admittedly with uncertainty ) on post #21, in the construction of any experiment, we refer to situations that can be described only using macroscopic observables. Born's rule should naturally follow from a clarification on what an ideal experimental apparatus is.

A. Neumaier said:
is based on a misinterpretation of how statistical mechanics relates to observation. I tried to correct this misinterpretation, but it is so deeply ingrained in stevendaryl's thinking that the information I provided only confuses him.
It is possible that Born's rule is ingrained deeply stevendaryl's, but it is also possible that you have not made a sufficiently clear and compelling argument. I too find what you are saying to be in need of clarification.

A. Neumaier said:
Yes, but s single spin is a tiny system, whereas statistical mechanics is about the behavior of macroscopic systems, in our case of the measurement device. Measuring a macroscopic system is not governed by Born's rule but by the identification of the expectation value with the measured value (within the intrinsic uncertainty). Born's rule is valid only for complete measurements of very small quantum systems such as a Stern-Gerlach spin.

Perhaps the central point is "Measuring a macroscopic system is not governed by Born's rule but by the identification of the expectation value with the measured value (within the intrinsic uncertainty)." which is in direct contrast with the main point of Vanhess71's post. I certainly have to think about this. But It appears that Arnold has a point because when we observe a macroscopic system we certainly do not destroy the entanglements and superposition between various particles, So it cannot be treated as an average of individual measurements over the different particles.
 
Last edited:
  • #30
My point was that for a macroscopic system we don't measure properties of single particles but quite "coarse-grained variables". This we do (perhaps more implicitly) already in classical mechanics. For the purpose of describing, within Newtonian mechanics, the motion of the planets around the Sun we can (in a good approximation) simplify them to "point particles", i.e., instead of describing the entire bodies as complicated systems we simplify them to simply be represented by their center of mass mutually attracted by other centers of mass via Newton's Law of gravity. For the purpose at hand that's a quite satisfactory treatment. The same holds at other layers of approximations going from the "most fundamental theory" (which is also a time-dependent notion; for us nowadays it means the Standard Model of elementary particles and general relativity) to more "effective" descriptions.

The point is that for the observables of macroscopic bodies in almost all cases (particularly of everyday life) the quantum expectation values of these macroscopic bodies are a sufficient description, because the fluctuations (both quantum and "thermal") around these expectation values are small compared to the accuracy which is needed to have a sufficiently detailed description of these macroscopic bodies. This means that the expectation values are as good as the single measurements within the needed accuracy. This doesn't make Born's rule superfluous. To the contrary, you need it to define how to evaluate the expectation values in the first place. Nobody thinks about this using quantum many-body theory, because it's just how the theory is formulated, i.e., Born's rule is an integral part of it, without which you cannot relate the formalism to real-world observations.

As I said before, some people find this still weird after over 90 years of QT. Maybe someone finds another way to formulate quantum theory (or a better more comprehensive other theory, containg QT as an approximation) and then derive Born's rule from these more convincing basis, but as long as we don't have such a convincing new formulation, it's safe to stick with the standard postulates and apply QT in terms of the minimal interpretation to the analysis of real-world experiments.
 
  • #31
vanhees71 said:
My point was that for a macroscopic system we don't measure properties of single particles but quite "coarse-grained variables".

Sorry about misrepresenting your post. I would be interested to see how Arnold has to respond.

This is clearly a question that I haven't thought about in depth, this discussion was extremely fruitful to me because it brought these issues into the forefront.

This is a crude and preliminary analysis, please treat it that way. I think the important point here is to understand in what situations a coarse grained macroscopic description is applicable. The born's rule does not apply when we talk about measuring a macroscopic system, in the sense that when we extract coarse grained information about the macroscopic system, we don't redefine(or recreate ) a new state for the system. Formally we write $$<A> = tr(\rho A)$$ however it does not have the same meaning when we talk about the same equation for a microscopic system. Which is formally defined as a sum over multiple observations.

This sharp differences in meaning, must lie in the fact that our interaction with a macroscopic system does not appreciably change its state, and we can use extrinsic variables for its description. Any precise observation of our macroscopic system will entail a different experimental apparatus, its meaning of an observation must be changed appropriately and a macroscopic description must be changed into a microscopic one.

It requires more analysis, I don't see right now how to proceed from here. More precisely I think this inherent stability when we talk about a macroscopic system could be formally analysed.
 
  • #32
I disagree. If you take the expectation value of, say the total energy or momentum of a macroscopic system, it's just that, the average of these quantities, and it's done with Born's rule. You contradict yourself, because what you wrote down is Born's rule. Given the state ##\hat{\rho}## (positive semidefinite trace-1 self-adjoint operator) then ##\langle A \rangle=\mathrm{Tr}(\hat{\rho} \hat{A})## is the expectation value of ##A##, and that's Born's rule.
 
  • #33
vanhees71 said:
I disagree. If you take the expectation value of, say the total energy or momentum of a macroscopic system, it's just that, the average of these quantities, and it's done with Born's rule. You contradict yourself, because what you wrote down is Born's rule

The point I am making is the following, the sense in which we apply Born's rule, to a macroscopic system is in sharp contrast with the sense in which we apply Born's rule to a macroscopic system. I have explained in post#31 is why this different procedures appear to be in sharp contrast.

Just to be clear when we use Born's rule for a microscopic system, we also insist that the state of system be redefined to highlight the new information that was obtained. However no such thing is done for a macroscopic system. It appears that there are 2 completely independent ways in which we are using Born's rule. This fact needs further clarification.
 
  • #34
I completely disagree with this idea that there are different meanings of the very clear formalism of QT. Born's rule is Born's rule, no matter whether you apply it to a single electron or to a many-body system. I don't know what you mean by the "state of the system is redefined". Do you mean something like a collapse hypothesis? I don't believe in this religion. In almost all measurements the collapse hypothesis is invalid, and where it is valid (a socalled von Neumann filter measurement) it's indeed a preparation procedure (as in my above most simple example of a Stern-Gerlach apparatus).
 
  • #35
vanhees71 said:
I completely disagree with this idea that there are different meanings of the very clear formalism of QT. Born's rule is Born's rule, no matter whether you apply it to a single electron or to a many-body system. I don't know what you mean by the "state of the system is redefined". Do you mean something like a collapse hypothesis? I don't believe in this religion. In almost all measurements the collapse hypothesis is invalid, and where it is valid (a socalled von Neumann filter measurement) it's indeed a preparation procedure (as in my above most simple example of a Stern-Gerlach apparatus).

Now you need to clarify what you mean? Basically yes I am talking about the collapse. When we measure the spin of a system, for any futher measurements we must set newly measured state as the initial state. That is the essence of the born's rule as we learn in textbooks.

Now what don't you believe about this?
 
Last edited:
Back
Top