Why MWI cannot explain the Born rule

In summary, the argument suggests that the minimal set of assumptions defining MWI cannot explain the Born rule. This can be seen by finding a counterexample of a system that satisfies these assumptions but does not have the probabilistic interpretation. The suggestion to simulate a virtual quantum world on a classical computer and consider the internal virtual observers also leads to the conclusion that the Born rule cannot be derived from the minimal set of assumptions defining MWI.
  • #106
Fredrik said:
The thing is, decoherence uses more than those two axioms. It uses the Born rule implicitly, by taking the Hilbert space to be a tensor product, and by computing the "reduced" density matrix as a partial trace of the state operator of the universe.

Without the possibility to do decoherence calculations, the only way to define the worlds is to say that given a basis (any basis) for the Hilbert space of the universe, each basis vector represents a world. To go beyond that, we need the Born rule, and a way to express the Hilbert space as a tensor product. Those things make decoherence a meaningful concept.

I have previously said that decoherence defines the worlds. I no longer think that that's the most appropriate way to define the worlds. What decoherence does is to single out a basis that defines interesting worlds. If my understanding of decoherence ideas is accurate (it might not be), any other basis defines worlds where the subsystems can't contain stable records of the states of other subsystems (such as a memory in the brain of a physicist). If well-defined memory states is an essential part of what consciousness is, the worlds identified by decoherence are the only ones that can contain conscious observers.
Fredrik, I don't agree with you that definition of subsystems in terms of tensor products is equivalent to the Born rule. After all, the former says nothing about probability per se.
And of course, one can calculate reduced density matrices without the Born rule.
 
Physics news on Phys.org
  • #107
Demystifier said:
Fredrik, I don't agree with you that definition of subsystems in terms of tensor products is equivalent to the Born rule. After all, the former says nothing about probability per se.
Consider two systems that aren't interacting with each other. If system 1 is in state [itex]|\psi\rangle[/itex] when we measure A, the probability of result a is

[tex]P(a)=|\langle a|\psi\rangle|^2[/tex]

If system 2 is in state [itex]|\phi\rangle[/itex] when we measure B, the probability of result b is

[tex]P(b)=|\langle b|\phi\rangle|^2[/tex]

According to the standard rules for probabilities, the probability of getting both of these results is

[tex]P(a,b)=P(a)P(b)=|\langle a|\psi\rangle|^2|\langle b|\phi\rangle|^2=|\langle a|\otimes\langle b|\ |\psi\rangle\otimes|\phi\rangle|^2[/tex]

This means that if we use the tensor product space to represent the states of the combined system, the Born rule will hold for that space too. Can you really look at this and think that we didn't choose to use tensor product to make sure that the probabilities assigned by the Born rule satisfy P(a,b)=P(a)P(b) when the systems aren't interacting?

Demystifier said:
And of course, one can calculate reduced density matrices without the Born rule.
You might be able to calculate them, but can you really justify the use of reduced density matrices to represent states of subsystems without using the Born rule? I'm pretty sure the answer is no.

I'll try to return to both of these things with more complete answers later, but feel to investigate it yourself. My feelings won't be hurt if you post a proof of some enlightening stuff before I do. :smile:
 
  • #108
An arbitrary state can be written as

[tex]\rho=\sum_i w_i|s_i\rangle\langle s_i|[/tex]

If we write [itex]|s_i\rangle=|f_i\rangle\otimes|g_i\rangle[/itex], the state can be expressed as

[tex]\rho=\sum_i w_i |f_i\rangle\langle f_i|\otimes|g_i\rangle\langle g_i|[/tex]

The easiest way to define the reduced density matrix, which I'll call [itex]\rho'[/itex], is to use a basis [itex]|\psi_\mu\rangle[/itex] for the first Hilbert space, and a basis [itex]|\phi_\alpha\rangle[/itex] for the second Hilbert space, which together define a basis [itex]|\phi_\mu\phi_\alpha\rangle=|\psi_\mu\rangle\otimes|\phi_\alpha\rangle[/itex] for the tensor product space. We define the operator [itex]\psi'[/itex] by saying that its matrix elements in the [itex]|\psi_\mu\rangle[/itex] basis are

[tex]\rho'_{\mu\nu}=\sum_\alpha\rho_{\mu\alpha,\nu\alpha}=\sum_\alpha\langle\psi_\mu\phi_\alpha|\rho|\psi_\nu\phi_\alpha\rangle=\sum_i w_i\langle\psi_\mu|f_i\rangle\langle f_i|\psi_\nu\rangle\sum_\alpha|\langle\phi_\alpha|g_i\rangle|^2[/tex]

The sum over [itex]\alpha[/itex] is =1, so

[tex]\rho'_{\mu\nu}=\langle\psi_\mu|\bigg(\sum_i w_i|f_i\rangle\langle f_i|\bigg)|\psi_\nu\rangle[/tex]

so we have

[tex]\rho'=\sum_i w_i|f_i\rangle\langle f_i|[/tex]

This is just an ordinary (mixed) state operator for the physical system associated with the first Hilbert space, so if we use a definition of QM that takes state operators to be the "states" of the theory, we don't have to justify the interpretation of the reduced density matrix as a representation of a state of a subsystem.

I'm not sure if we should be talking about the original Born rule [itex]P(a)=|\langle a|\psi\rangle|^2[/itex] or its generalization to mixed states [itex]\langle A\rangle_\rho=\mbox{Tr}(\rho A)[/itex].

If we're using a definition of QM that takes unit rays to be the "states" of the theory, then the use of state operators in general needs to be justified. This is done by first noting that the average result in a series of measurements of A on identically prepared systems is

[tex]\langle A\rangle=\sum_a P(a)a=\sum_a a|\langle a|\psi\rangle|^2=\sum_a\langle a|\psi\rangle\langle\psi|A|a\rangle=\mbox{Tr}(\rho A)[/tex] ...and also [tex]=\langle\psi|\Big(\sum_a|a\rangle\langle a|\Big)A|\psi\rangle=\langle\psi|A|\psi\rangle[/tex]

and then noting that the average result on an ensemble with a fraction [itex]w_i[/itex] of the members prepared in state [itex]|\psi_i\rangle[/itex] is

[tex]\sum_i w_i\langle\psi_i|A|\psi_i\rangle=\sum_n\sum_i w_i\langle\psi_i|A|n\rangle\langle n|\psi_i\rangle=\sum_n\langle n|\Big(\sum_i w_i|\psi_i\langle\psi_i|\Big)A|n\rangle=\mbox{Tr}(\rho A)[/tex]

We're using the original Born rule in the first step, so if our axioms talk about state vectors rather than state operators, we need the original Born rule to justify that expectation values can be written as [itex]\langle\psi|A|\psi\rangle[/itex], which then gives us the generalized Born rule.

When we're dealing with state operators and the generalized Born rule, the rule P(a,b)=P(a)P(b) for non-interacting systems is replaced by

[tex]\langle AB\rangle=\langle A\rangle\langle B\rangle[/tex]

where "AB" is still undefined if we haven't decided to use the tensor product yet. "AB" is supposed to be the mathematical representation of the operationally defined "measure B first, then A". If we use the tensor product, the above holds true with AB defined as [itex](A\otimes I)(I\otimes B)=A\otimes B[/itex]. Can you look at this and not think that the reason we're using tensor products is that it ensures that this result holds for non-interacting systems?

I guess that answers the question of which version of the Born rule we should be talking about. The answer is that it doesn't matter. If we're talking about the original Born rule, the argument in my previous post shows that it's intimately connected to tensor products, and if we're talking about the generalized Born rule, the argument in the previous paragraph shows the same thing. (I'm leaving the proof of [itex]\mbox{Tr}(A\otimes B)=\mbox{Tr}(A)\mbox{Tr}(B)[/itex] as an excercise).
 
Last edited:
  • #109
Fredrik said:
Consider two systems that aren't interacting with each other. If system 1 is in state [itex]|\psi\rangle[/itex] when we measure A, the probability of result a is

[tex]P(a)=|\langle a|\psi\rangle|^2[/tex]
...
Here you assume the Born rule. Therefore, your further steps (which I don't quote) cannot be qualified as a derivation of the Born rule.

Fredrik said:
You might be able to calculate them, but can you really justify the use of reduced density matrices to represent states of subsystems without using the Born rule? I'm pretty sure the answer is no.
That's an interesting question. I will think about it.
 
  • #110
Demystifier said:
Here you assume the Born rule. Therefore, your further steps (which I don't quote) cannot be qualified as a derivation of the Born rule.
That's right (and also obvious). I just proved that if we use the tensor product and the Born rule, we get P(a,b)=P(a)P(b) for non-interacting systems. (It would be a disaster to get P(a,b)≠P(a)P(b). If QM works for nuclei and electrons separately, it wouldn't work for atoms. But of course it wouldn't work for nuclei either...) The point is that quantum mechanics for individual systems, which by definition includes the Born rule, more or less forces us to us to use the tensor product to represent the states of the composite system.

I wonder if it's possible to prove that the Born rule is the only probability measure that gives us P(a,b)=P(a)P(b). That would be a derivation of the Born rule from tensor products, but it was never my goal to find a derivation. I'm just saying that the two are clearly not completely independent.

As I pointed out in #108, when we use axioms that start with state operators instead of state vectors, we should require <AB>=<A><B> instead of P(a,b)=P(a)P(b), but the conclusion is the same.

Edit: It was probably a mistake to think that we should require <AB>=<A><B> when the axioms talk about state operators and P(a,b)=P(a)P(b) when the axioms talk about state vectors. I think both identities must hold in both cases.

Demystifier said:
That's an interesting question. I will think about it.
I think #108 answers it. If we start with axioms that take state operators to be the mathematical representation of states, no justification is required. It's just an axiom. If we start with axioms that start with state vectors, the justification is given by the stuff I said about expectation values.
 
Last edited:
  • #111
Fredrik, I mostly agree with your last post. In fact, sooner or later you and me allways arrive at an agreement, which is why I like to discuss with you.

But let me clarify one thing regarding the question whether many worlds (without the Born rule) can be derived from the Schrodinger equation. You are right that description of a subsystem by a reduced density matrix can hardly be justified without assuming the Born rule. Nevertheless, my point is that essential physical aspects of decoherence
(or at least of something closely related to decoherence) can be understood without introducing the reduced density matrix. Indeed, this stuff has been known much before the word "decoherence˝ was introduced in quantum mechanics. For example, see the Quantum Mechanics textbook written by Bohm in 1951, the chapter on the theory of quantum measurements. (This book is written in a Copenhagen style, a year before Bohm introduced his hidden-variable interpretation of QM.)

Here is the basic idea. Consider the wave function in the configuration space describing a collection of many interacting particles. Before the interaction, the total wave function is typically a product of wave functions of the non-interacting subsystems. However, after the interaction, the total wave function becomes a superposition of such products. The crucial property of this superposition is that each term of the superposition is a many-particle wave function which DOES NOT OVERLAP with any other term in the superposition. (More precisely, the overlap is negligible, due to a large number of degrees of freedom - particles.) While this lack of overlap is technically not the same as decoherence, it is closely related to it. Indeed, if you calculate the reduced density matrix from this wave function, you will obtain a decohered density matrix. But you don't need to calculate the reduced density matrix at all. This wave function itself is sufficient to understand how "many worlds˝ emerge. Since this wave function consists of many (almost) non-overlaping chanels, each chanel may be thought of as another "world˝. This is how Scrodinger equation of many degrees of freedom predicts the existence of "many worlds˝, without any additional assumptions.
 
  • #112
Demystifier said:
You are right that description of a subsystem by a reduced density matrix can hardly be justified without assuming the Born rule.
I'll take a stab at sketching it anyways. My apologies if I've missed something important.


If we start with the premise:
  • We have some "full" Hilbert space
  • We have some collection of observables
  • Each observable A is represented as an operator Af on the full Hilbert space
  • The expectation of A on a density matrix p is Tr(p Af)

And we can find:
  • A "reduced" Hilbert space
  • Each of our observables A can be represented as an operator Ar on the reduced Hilbert space
  • A map that turns a density matrix p into a reduced density matrix pr

Then all we really need to justify the reduced density matrices is the identity
Tr(p Af) = Tr(pr Ar)​
right?
 
  • #113
Demystifier said:
he crucial property of this superposition is that each term of the superposition is a many-particle wave function which DOES NOT OVERLAP with any other term in the superposition.
...
This wave function itself is sufficient to understand how "many worlds˝ emerge.
...
This is how Scrodinger equation of many degrees of freedom predicts the existence of "many worlds˝, without any additional assumptions.
I have sort of been flip-flopping back and forth between thinking that this is right and thinking that this is wrong. I'm leaning towards wrong. Consider what I said here:
Fredrik said:
Suppose e.g. that you bet $1000 that the spin will be "up", and then you perform the measurement. The state of the system+environment will change like this:

(|↓>+|↑>)|:rolleyes:> → |↑>|:smile:> + |↓>|>

Yes, there will be other terms, which has your memory in a superposition of "smile" and "yuck", but what decoherence does is to make the coefficients in front of them go to zero very rapidly. Now each of the remaining terms is interpreted as a "world" in which a particular result happened, and "you" (a different you in each world) remember that it happened.

Edit: This was actually a mistake. What I should have done is to define |S>=|↓>+|↑> and then said that the density matrix changes as described by

|S>|:rolleyes:><:rolleyes:|<S| → |↑>|:smile:><:smile:|<↑| + |↓>|><|<↓|

This is a mixed state, not a superposition.
I'm not 100% sure what you mean by "overlap", but I think you're probably talking about final state vectors like the one in the post I just quoted. We would have an "overlap" if there had also been a |↑>|> term on the right. (Technically, there always is, but we're talking about situations in which the coefficient in front of it is really small).

The process described by the first → in the quote is the development of correlations between subsystems. This is what happens when a silver atom goes through a Stern-Gerlach magnet, before we determine its position by detecting it in one location or the other. The process described by the second → in the quote is a measurement. This is what decoherence does. It turns pure states into mixed states.

I think we need the second process to define the "interesting" worlds, and I don't think we need either of them to define "worlds".
Fredrik said:
Without the possibility to do decoherence calculations, the only way to define the worlds is to say that given a basis (any basis) for the Hilbert space of the universe, each basis vector represents a world.
Edit: The first process defines some set of worlds. It's too small to be all the worlds, and too big to be all interesting worlds (with "interesting" defined as worlds in which the environment can contain stable records of the system's state), but it's certainly a set of worlds.
 
Last edited:
  • #114
Hurkyl said:
[*] The expectation of A on a density matrix p is Tr(p Af)[/list]
Should we consider this a definition of a mathematical term, or a statement about what to expect when we perform measurements? If it's the latter, then this is the Born rule (the generalized version that works for mixed states too), which is precisely what the text you quoted (Demystifier's post) said not to use.

No time to think about the rest now. I need to get some sleep.
 
  • #115
RUTA said:
It's very different because you have access to information from the space that surrounds you. You have no such access to "other" universes, by definition (if you have access, they're not "other," they're part of this one).

This argument is not a very sound one. MWI does not imply "other universes that we do not have access to", on the opposite it says that there is one universe that is defined by the universal wavefunction.

So I don't see why I can't say, that analogous to my position in space, my 'position in the probability' is not in any way preferred or unfairly sampled, and that this 'bifurcated history tree' you've mentioned is not symmetric.
 
  • #116
dmtr said:
This argument is not a very sound one. MWI does not imply "other universes that we do not have access to", on the opposite it says that there is one universe that is defined by the universal wavefunction.

So I don't see why I can't say, that analogous to my position in space, my 'position in the probability' is not in any way preferred or unfairly sampled, and that this 'bifurcated history tree' you've mentioned is not symmetric.

See Adrian Kent's Perimeter presentation, "Theory Confirmation in One World and its Failure in Many," http://pirsa.org/index.php?p=speaker&name=Adrian_Kent.
 
  • #117
RUTA said:
See Adrian Kent's Perimeter presentation, "Theory Confirmation in One World and its Failure in Many," http://pirsa.org/index.php?p=speaker&name=Adrian_Kent.

Well, my advice to these inhabitants would be "use the symmetry and the number of simulation copies to derive the probabilities for future events". Following this advice will help these inhabitants to predict the future better.
 
Last edited:
  • #118
Fredrik said:
Should we consider this a definition of a mathematical term
Certainly, but that does not exclude the possibility that it has something to say about what to expect when we perform measurements.

If it's the latter, then this is the Born rule (the generalized version that works for mixed states too), which is precisely what the text you quoted (Demystifier's post) said not to use.
I assumed that we wouldn't be worrying about reduced density matrices if we hadn't already made up our minds about density matrices.

Or, to put it differently, my argument justifies the use of reduced density matrices for studying quantitative and qualitative properties of density matrices, and is independent of whatever issues we have of the relation between density matrices and reality.
 
  • #119
I have been meaning to ask a question related to some of the issues of this thread namely; Does decoherence require a an environment (decomposition into subsystems)?

I tend to view decoherence as an approximate feature defined as practical inability to observe a relative phase between two states. This feature does not seem to require an environment. Let's consider the prototypical case of the Schrödinger cat. The cat is a macroscopic object the state of which is determined by it's microscopic constituents. Let us assume that we can define subspaces of microscopic states where the cat is alive or dead with certainty (i.e. eigenstates of an "aliveordead operator") and the Hilbert space of the cat is a direct sum of these two subspaces. When we talk about the cat being in the state |alive> we are actually talking about a class of microscopic states belonging to the alive subspace, and similarly with the state |dead>. Now if we consider a superposition of states a|alive>+b|dead>, the relative phase [itex]\varphi=\text{arg}(b/a)[/itex] is determined by the precise microscopic states. Due to the internal dynamics within each subspace the phase becomes a chaotic variable, which for all practical purposes would be impossible to observe. Note that the unobservability of the relative phase means that we could not (in practice) distinguish a pure state from a mixed state.

Regarding the MWI: Even if we accept that a macroscopic system, at least in a coarse grained view, evolves into a mixed state we still would need to somehow relate the diagonal parts of the density matrix to the probabilities of the observer experiencing a particular result. And I simply don't see how this can be done without invoking something equivalent to the born rule.
 
  • #120
Fredrik said:
The process described by the second → in the quote is a measurement. This is what decoherence does. It turns pure states into mixed states.
I don't think that it is correct.

First, no physical process turns pure states into mixed states, provided that the whole system is taken into account. Such a process would contradict unitarity. Of course, if you consider a SUBsystem, then then such a process is possible.

Second, a measurement can be described even by describing the whole system. It may be impossible in practice due to a large number of the degrees of freedom, but it is possible in principle. Therefore, a measurement can be described in terms of pure states as well, at least in principle.
 
  • #121
Demystifier said:
Second, a measurement can be described even by describing the whole system. It may be impossible in practice due to a large number of the degrees of freedom, but it is possible in principle.
I am under the impression that is actually a significant point of contention between interpretations -- e.g. that Copenhagen says it's impossible in principle.

(unless, of course, you switch to a new and better physical theory)
 
  • #122
Fredrik said:
I wonder if it's possible to prove that the Born rule is the only probability measure that gives us P(a,b)=P(a)P(b).

I wonder if you like this, more than I do?

Ariel Caticha's
Consistency, Amplitudes and Probabilities in Quantum Theory "Quantum theory is formulated as the only consistent way to manipulate probability amplitudes. The crucial ingredient is a consistency constraint: if there are two different ways to compute an amplitude the two answers must agree. This constraint is expressed in the form of functional equations the solution of which leads to the usual sum and product rules for amplitudes. A consequence is that the Schrödinger equation must be linear: non-linear variants of quantum mechanics are inconsistent. The physical interpretation of the theory is given in terms of a single natural rule. This rule, which does not itself involve probabilities, is used to obtain a proof of Born’s statistical postulate. Thus, consistency leads to indeterminism."
-- http://arxiv.org/PS_cache/quant-ph/pdf/9804/9804012v2.pdf

In effect I think he is trying to generalize Cox, Jaynes "derivation" of the rules of koglomorov's probability, not from the traditional axioms, but from some assumptions of consistency of reasoning around information.

One key assumption is that he assumes as part of the microstate of information beeing represented/quantified by a REAL numbers (or degree of beliefs). Then by a series of arguments he shows that the only consistent logical system och ang or operators etc that fulfills this (and some hidden mor or less natural assumptions) is indistinguishable from koglomorov probability axioms.

The generalisation is to isntead, assume (again key assumption) that instead the state of information is represented/quantified by a COMPLEX number. Then his idea is that similarly quantum logic emerges as the only consistent system. He argument for the Born rule is I think effectively the same as the ones trying to "count" distinguishable microstates and arrive at some kind of "frequency".

I think both of these ideas as intersting but my objection to both of those are the key assumptions of choose real or complex numbers as representation.

Ariel points out this himself:
"A second, simpler question is why do we seek a representation in terms of complex numbers? Again, no answer here; this is an unexplained feature of quantum theory. It seems that a single complex number is sufficient to convey the physically relevant information about a setup."

So I think there mere starting point of a normed linear vector space over complex numbers is quite non-trivial as well.

For me, the physical insight of the meaning of real numbers in a framework where only a finite amount of information is encodable is still mysterious. Before I understand this, I can not claim to be content with any higher level derivation which contains this as unmotivated baggage.

I'm curious if Fredrik, finds Ariels paper more satsifactory than I do?

/Fredrik
 
  • #123
Fredrik said:
That's right (and also obvious). I just proved that if we use the tensor product and the Born rule, we get P(a,b)=P(a)P(b) for non-interacting systems. (It would be a disaster to get P(a,b)≠P(a)P(b). If QM works for nuclei and electrons separately, it wouldn't work for atoms. But of course it wouldn't work for nuclei either...) The point is that quantum mechanics for individual systems, which by definition includes the Born rule, more or less forces us to us to use the tensor product to represent the states of the composite system.

Do you picture that the notion of non-interacting systes still makes sense when you incorporate say gravity? how do you prevent the gravitational interaction?

/Fredrik
 
  • #124
Fra said:
I wonder if you like this, more than I do?
...
I'm curious if Fredrik, finds Ariels paper more satsifactory than I do?
My impression (after spending about half an hour on it) is that he seems to have found a way to understand probability amplitudes on a slightly deeper level. It looks good, but not very significant.

Fra said:
Do you picture that the notion of non-interacting systes still makes sense when you incorporate say gravity? how do you prevent the gravitational interaction?
I'm not going to try to unify QM with GR today. Maybe tomorrow. :smile:
 
  • #125
dmtr said:
Well, my advice to these inhabitants would be "use the symmetry and the number of simulation copies to derive the probabilities for future events". Following this advice will help these inhabitants to predict the future better.

You're still missing the point. How do you know what probability to use in a simulation if empirically all you ever have access to is one branch of the outcome? He's giving you a God's eye view so you can see that it's impossible for the individual inhabitants to do science. We're an individual inhabitant, not God.
 
  • #126
Demystifier said:
I don't think that it is correct.

First, no physical process turns pure states into mixed states, provided that the whole system is taken into account. Such a process would contradict unitarity.
I meant approximately. I was just too lazy to explain that (and I thought people would understand what I meant after I had just explained it for the first type of process).
 
Last edited:
  • #127
Fredrik said:
I'm not going to try to unify QM with GR today. Maybe tomorrow. :smile:

I sure won't do that neither today nor tomorrow either :)

But I guess my attitude is that I see indications that the deeper understanding of the foundations of QM, and a possible deeper motivation for QM and it's structure, might be additionally complicated by us trying to keep gravity out of it. Although on the surface foundational QG for sure looks more hairy than foundational QM, trying to find a connection as deep as possible might also enlighten us with regards to normal QM, and the structure of the standard model of particle physics in the quest for a GUT. Maybe gravity is simply the "missing link". Maybe it's a mistake to think that trying to think of gravity already from start will make the quest harder, rather than easier?

/Fredrik
 
  • #128
Fredrik said:
I'm not going to try to unify QM with GR today. Maybe tomorrow. :smile:

I'm hoping tomorrow ... or the day after. I have the equation, I just need a solution to get the party started :smile:
 
  • #129
Fra said:
Maybe it's a mistake to think that trying to think of gravity already from start will make the quest harder, rather than easier?
That's certainly possible. The difference between non-relativistic QM and special relativistic QM is a different group of symmetries for the theory (or equivalently, a different algebra of observables). Is there a difference between special relativistic QM and general relativistic QM? Probably. I think that's what LQG is trying to answer. I also think that almost everything that's been written about attempts to interpret QM or its underlying mathematical structure as a description of what actually happens, have completely ignored those differences.
 
  • #130
jensa said:
I have been meaning to ask a question related to some of the issues of this thread namely; Does decoherence require a an environment (decomposition into subsystems)?
It think the answer is yes. I don't know decoherence well, but the impression I got is that the crucial step is the calculation of a reduced density matrix, as discussed above. This isn't possible without a decomposition into subsystems. If you want a better answer, try searching arxiv.org for articles by Zurek, or buy Schlosshauer's book. (I haven't read it yet, but it's in my shopping cart).
 
  • #131
Fredrik said:
It think the answer is yes. I don't know decoherence well, but the impression I got is that the crucial step is the calculation of a reduced density matrix, as discussed above. This isn't possible without a decomposition into subsystems. If you want a better answer, try searching arxiv.org for articles by Zurek, or buy Schlosshauer's book. (I haven't read it yet, but it's in my shopping cart).

Thank you for the response Fredrik,

I don't think that decomposition into subsystems is sufficient (maybe not even necessary) to produce decoherence. Consider two sets of interacting two-level systems. The Hilbert space of this composite system we describe as a tensor product of the individual Hilbert spaces. If we are only interested in the properties (observables) of one of the subsystems we may trace out the other one to produce a reduced density matrix. The reduced density matrix will generally be a mixed state but the coherence factors (off diagonal elements of the density matrix) need not vanish irreversibly. Most likely you would observe an oscillatory behaviour at a certain frequency. To observe irreversible behaviour you need to also assume that the system you are tracing out contains a large (infinite) number of degrees of freedom so that different frequencies add up to produce a decay on average.

It seems to me that the macroscopic nature of the environment (many degrees of freedom) is more important than the decomposition into subsystems in order to observe irreversible loss of coherence.

I hope I don't project too much of my own ignorance onto the people on this board but loss of coherence seems to be similar to the classical increase of entropy in that everybody believes it to occur but very few can actually show it (and explain what causes it). In the case of increase of classical entropy there seems to be a number of ways how to justify it. Personally I prefer the coarse grained explanation that the increase of our ignorance (entropy) comes from our mapping from the microscopic configurations towards macroscopic observables. I.e. it can be shown that when we map the many-particle phase space onto a space of macroscopic observables, and replace the distribution function [itex]\rho(X,t)[/itex] where [itex]X[/itex] is the coordinate in phase space, by the distribution function [itex]\tilde{\rho}(A,t)[/itex] where [itex]A[/itex] is a coordinate in the space of macroscopic observables, the entropy defined in terms of the macroscopic distribution function [itex]\tilde{\rho}[/itex] always increases.

I believe the origin of decoherence has a similar source, namely that it is the mapping from microscopically distinguishable states onto macroscopically distinguishable observables that produces a practical impossibility to observe interference effects of macroscopic objects. As in my example with Schrödingers cat; our ignorance about the microscopic states along with the extremely short time scales of microscopic processes causes the interference effects to become "averaged out".

Of course, when we are talking about the decoherence of truly microscopic systems one needs to consider the entanglement with a macroscopic environment. Now the observables we are interested in are of course the observables associated with the original microscopic system (so they are not macroscopic observables). But still there exists a many-to-one mapping corresponding to our ignorance about the many degrees of freedom of the environment.

I wish I could be more thorough with my explanations but I have a lot of work to do. Btw, how do you guys manage to spend so much time on the boards and learn a lot of new stuff?

PS. I have Schlosshauers book and will try to read it when I get some time.
 
  • #132
RUTA said:
I'm hoping tomorrow ... or the day after. I have the equation, I just need a solution to get the party started :smile:

What was your general idea? (There was probably a past thread about this, but I can't remember)

/Fredrik
 
  • #133
Hurkyl said:
I am under the impression that is actually a significant point of contention between interpretations -- e.g. that Copenhagen says it's impossible in principle.
I guess you have in mind the idea that not everything can be described by QM, but that you also need a classical world. I would say that this is only one of several different variants of the Copenhagen interpretation. And I think that this particular variant is quite obsolete.
 
  • #134
Some of you mentioned the Schlosshauer book. Let me just say that I recommend it to everyone. Today no one can say that he understands QM well without being familiar with basics of decoherence. Actually, this book contains more than just basics, but it is worth reading.
 
  • #135
Demystifier said:
I would say that this is only one of several different variants of the Copenhagen interpretation. And I think that this particular variant is quite obsolete.
My impression is that CI always has the property that collapse is real (as opposed to a mathematical technique or a change-of-frame-type thing). I know of three variants:
  1. Quantum and classical mechanics tell us what's really going on1, but QM becomes inaccurate above a certain scale and CM becomes inaccurate below a certain scale. We can effectively use the two in concert by invoking a Heisenberg cut.
  2. QM doesn't tell us anything about what's really going on -- it is a theory of our ignorance2 of deeper reality
  3. QM will work all the way up, once we figure out what nonlinear terms should be inserted into Schrödinger's equation to make collapse occur

(note this last one still asserts that unitary evolution of states is wrong on macroscopic scales)

Is there another variant you had in mind?


1: Meaning, roughly, that the elements of the theory correspond to elements of reality
2: I don't mean this pejoratively -- I mean it as in "ignorance probabilities"
 
  • #136
Hurkyl said:
Demystifier said:
Second, a measurement can be described even by describing the whole system. It may be impossible in practice due to a large number of the degrees of freedom, but it is possible in principle.

I am under the impression that is actually a significant point of contention between interpretations -- e.g. that Copenhagen says it's impossible in principle.

(unless, of course, you switch to a new and better physical theory)

There was a recent discussion about the "definition of CI". But to leave aside that classification of views here, the reason why I think it's impossible even in principle is this:

As I see it, a given observer doesn't in general "choose" to observe this or that. The "complete" picture (as I see it) is that a given observer always "observes" (ineracts with) it's own environment.

But no given observer, can relate to, and decode all possible the degrees of freedom in the environment. So each observer sees a truncated world.

Now, I take a similar to rovelli's RQM view here, that the only way to level two "truncated worlds" against each other, is by means of the observers interacting. The view of each observer, will then be revealed by the way that act upon each other.

Like a game of poker. Each player has "a vision" of the future of the game - if my opponent folds, I have good grounds to think it's because he thinks he has slim chance to win - his vision is revealed to me. The only way to find out the opponents "visions" are to play the game.

The problem in QM, is that even though it's true that a second observer (possibly a massive one) COULD in some approximation observe the measurement process in the sense of "environment as the observer", this massive amount of information could never be conveyed back to the original observer - even if a perfect communication channel was established - simply because they are not comparable in complexity.

The analogy to say SR or GR would be perfect if we could establish the transformations that restore observer invariance here, and view these transformations in the realist sense.

But one problem is that there is no way that there can exist one-2-one transformations between structures that can't not encode the same amount of information. The transformations themselves must create and destroy information unless we have an *equilibrium situation* where the environment contains copies of the same repeating and thus "redundant" information, then we find the special case that a truncated system might contain the same information as a larger system.

But if we consider the general non-equilibrium case, I think the information preserving idea simply won't do.

I think that to understand QM, would be to find the general case, and then see why and how QM structure as we know it, does emerge as a uniqe special case. Like GR vs SR.

/Fredrik
 
  • #137
There have been several discussions about the CI recently. There was a thread started by Demystifier, and this thread started by me. There are links to interesting papers in #7 and #20.

My impression is that the original CI is essentially the same as what we've been calling "the ensemble interpretation" in this forum. There's a formal difference in their definitions, but I don't see how that formal difference is an actual difference. See #33 in the thread I linked to above. Hurkyl, your #2 appears to be the same thing, but if QM doesn't tell us what actually happens, then there's nothing that suggests that "collapse" is a physical process, as you suggested at the start.
 
  • #138
Fra said:
What was your general idea?

A discrete path integral over graphs. The difference matrix K and source vector J are constructed from boundary operators in the spacetime chain complex of the graph so that Kx = J where x is the vector of nodes, links or plaquettes in the resulting scalar, vector or tensor field theory, respectively (this follows from the boundary of a boundary principle, BBP). This restricts K and J in the discrete action of the transition amplitude, Z. Note that Z is not a function of the field Q, i.e., Q is the integration variable in computing Z, yet quantum and classical field theories are all about Q. How is that?

When you ask for the probability that the kth node, link or plaquette has the value Qo you obtain Z(Qk=Qo)/Z (Z is a partition function since we're using a Euclidean path integral), which is the discrete counterpart to QFT. If you ask for the most probable value of Qo, you find you must solve KQo = J, i.e., the discrete counterpart to CFT. Notice that by requiring the graphical basis satisfies the BBP, your discrete CFT automatically satisfies the BBP, which is the basis for the local conservation of momentum and energy (divergence-free stress-energy tensor). These are standard calculations, we're just proposing a different take on them that leads to a discrete CFT at odds with GR.

Anyway, we're trying to solve the tensor version of KQo = J for some simple situation and compare the result to GR to "get the party started." We expect differences from GR since our version of discrete tensor CFT is linear and constitutively non-local while GR is non-linear and local. Essentially, our discrete tensor CFT is Regge calculus where the nodes of the simplices are clusters of graphical elements (think of clusters of cubes joined by line segments) so there are no vacuum solutions (it's constitutively non-local aka inseparable -- no empty spacetime, just spacetimematter) and the lengths of the line segments joining the clusters are simply equal to the average values of the Q's on the shared cube faces (plaquettes) between clusters (average value = most probable value since we've a Gaussian distribution function). This resolves the problem resulting from the fact that violations of Bell's inequality imply causal and/or constitutive non-locality while GR is local on both counts, in favor of QM (GR must be revised).

We have a paper under review at Foundations of Physics. If that gets accepted and we find our classical solution, then maybe I'll start a thread :smile: We can't discuss this anymore here, it's out of context.
 
  • #139
I have now read enough of this paper to see that the fact that we should be using the tensor product to represent a composite system can be derived without direct reference to the Born rule. It's based on the quantum logic approach to QM, which associates a mathematical structure with the set of statements of the form "if you measure observable A, you will get a result in the set B with probability 1". So it has some connection to probabilities, but it's not as strong as I expected. Maybe the stuff I mentioned above about how we need to use the tensor product to ensure that the Born rule satisfies P(a,b)=P(a)P(b) for non-interacting systems is in there somewhere, and I just can't see it.
 
  • #140
What about http://arxiv.org/abs/0903.5082 which tries to derive the Born rule from

(i) States are represented by vectors in Hilbert space
(ii) Evolutions are unitary
(iii) Immediate repetition of a measurement yields the same outcome

"To derive it we cannot use reduced density matrices, Eqs. (1,2). Tracing out is averaging [25, 29, 30] - it relies on pk = |psik|2, Born's rule we want to derive."
 
Last edited:
Back
Top