Tensor products and subsystems

  • Thread starter Fredrik
  • Start date
  • Tags
    Tensor
In summary, the conversation discusses different interpretations and theorems regarding the construction and decomposition of Hilbert spaces, specifically in relation to subsystems in quantum mechanics. The SSC Theorem and the "external correlations" theorem are mentioned as possibly relevant. The article "Physical justification for using the tensor product to describe two quantum systems as one joint system" by Aerts and Daubechies is also referenced as a possible source of information. Mermin's "Ithaca" interpretation is briefly discussed, which suggests that an additional assumption on top of QM is necessary to avoid the many-worlds interpretation.
  • #1
Fredrik
Staff Emeritus
Science Advisor
Gold Member
10,877
423
I'm not sure what forum to put this in. It's a math question, but it's only of interest to physics people.

Given two Hilbert spaces [itex]H_1[/itex] and [itex]H_2[/itex], we can construct their tensor product [itex]H=H_1\otimes H_2[/itex]. This is another Hilbert space.

What I'm wondering is if there are any theorems about what sort of decompositions we can make if we're given a Hilbert space [itex]H[/itex], and want to express it as a tensor product of two "smaller" Hilbert spaces. Can we pick an arbitrary subspace and call it [itex]H_1[/itex], and then construct [itex]H_2[/itex] from [itex]H[/itex] and [itex]H_1[/itex]?

I'm interested in subsystems in QM, and not just ensembles of identically prepared systems. I want to know e.g. if we can always decompose the Hilbert space of the universe (in the many-worlds interpretation) into "this guy" [itex]\otimes[/itex] "everything else".

Maybe I'm phrasing the question wrong. Maybe I should focus on the observables instead of the states. I don't know. If you do, let me know.
 
Physics news on Phys.org
  • #2
Fredrik,

there is a theorem in quantum logic, which justifies the use of the tensor product of Hilbert spaces to describe a compound system

T. Matolcsi, "Tensor product of Hilbert lattices and free orthodistributive product of orthomodular lattices", Acta Sci. Math. (Szeged) 37 (1975), 263.

D. Aerts and I. Daubechies, "Physical justification for using the tensor product to describe two quantum systems as one joint system", Helv. Phys. Acta, 51 (1978), 661.

This theorem does not answer your question exactly, but possibly it can give you a clue.
 
  • #3
Fredrik said:
What I'm wondering is if there are any theorems about what sort of decompositions we can
make if we're given a Hilbert space [itex]H[/itex], and want to express it as a tensor product
of two "smaller" Hilbert spaces. Can we pick an arbitrary subspace and call it [itex]H_1[/itex],
and then construct [itex]H_2[/itex] from [itex]H[/itex] and [itex]H_1[/itex]?

One needs to think in terms of correlations between subsystems.
Two theorems may be relevant here:

A) The SSC Theorem: "Subsystem correlations determine the state".
Roughly stated: [...] the density matrix of a composite system determines all the correlations
among the subsystems that make it up and, conversely, the correlations among all the
subsystems completely determine the density matrix for the composite system they make up.

B) "The external correlations of a system are necessarily trivial if and only
if its state is pure".

The above are taken from Mermin's paper quant-ph/9801057 (see the two appendices).

That paper may also provide food for thought in other ways. :-)
 
  • #4
Thanks guys. I'm going to sleep now, but I'll take at look at the Mermin paper tomorrow. (I don't have an easy way to access the articles Meopemuk referenced, and I suspect they'll just be more elaborate versions of the argument that we use tensor products because that's what the Born rule needs).
 
  • #5
Fredrik said:
Can we pick an arbitrary subspace and call it [itex]H_1[/itex], and then construct [itex]H_2[/itex] from [itex]H[/itex] and [itex]H_1[/itex]?

Generally speaking, no. An n-dimensional Hilbert space, where n is prime, can't be decomposed into a tensor product at all. If n=ab, then there is an enormous number of possible decompositions into an a-dimensional and a b-dimensional space. You just pick an arbitrary basis in the original space and form 'a' groups of 'b' basis vectors.

With regard to physical systems, you need to think about lattice discretizations. Simplest case: consider a set of n nodes, each of which may or may not be occupied by a single particle. The Hilbert space is 2^n dimensional. You can decompose the set into a+b nodes, with 2^a and 2^b dimensional Hilbert spaces whose tensor product gives the original space.
 
  • #6
I haven't read the whole article yet, but I'm going to. It's a pretty good article. Mermin is a good enough writer to make sure that even some of the things he's wrong about are worth reading. :smile:

My first impression was that everything he's saying about his "Ithaca" interpretation is exactly the view of the many-worlds interpretation that I've arrived at through my discussions here. A few pages later he made it clear that what distinguishes "Ithaca" from MWI is that he assumes that QM just can't be applied to consciousness! I find that very strange, but I'll keep reading. If I just ignore the stuff about consciousness, the rest of it seems to be the sort of things I've been wanting to read about the MWI for a while now.
 
  • #7
Fredrik said:
My first impression was that everything he's saying about his "Ithaca" interpretation is exactly the view of the many-worlds interpretation that I've arrived at through my discussions here. A few pages later he made it clear that what distinguishes "Ithaca" from MWI is that he assumes that QM just can't be applied to consciousness! I find that very strange, but I'll keep reading. If I just ignore the stuff about consciousness, the rest of it seems to be the sort of things I've been wanting to read about the MWI for a while now.
Hm, my impression is that the Mermin interpretation is a variant of the information interpretation, not a variant of MWI.
 
  • #8
I'm not familiar with the information interpretation. (Not even sure if I've heard of it). Mermin seems to agree that we need an additonal assumption on top of QM to avoid having to deal with many worlds, and his assumption is that consciousness is weird enough to invalidate the conclusion that realism implies many worlds.
 
  • #9
strangerep said:
One needs to think in terms of correlations between subsystems.
Two theorems may be relevant here:

A) The SSC Theorem: "Subsystem correlations determine the state".
Roughly stated: [...] the density matrix of a composite system determines all the correlations
among the subsystems that make it up and, conversely, the correlations among all the
subsystems completely determine the density matrix for the composite system they make up.

B) "The external correlations of a system are necessarily trivial if and only
if its state is pure".

The above are taken from Mermin's paper quant-ph/9801057 (see the two appendices).

That paper may also provide food for thought in other ways. :-)
I didn't read the appendices until today. They're not quite what I asked for, but they matched my needs pretty well anyway. I think the answer to my question is actually quite obvious. (I don't know why I didn't get this right away). QM doesn't answer questions like that, but each specific theory of matter in the framework of QM does.
 
  • #10
I have a question about appendix A in Mermin's article, but someone who knows this stuff well might be able to answer without reading it. Why do expressions of the form [itex]\mbox{Tr }A\otimes B[/itex] represent correlations between subsystems? :confused:

What Mermin does in appendix A is to prove that any matrix element of the state operator (I don't want to call it "density matrix" when it's an operator) can be expressed as a sum of terms of the form [itex]\mbox{Tr }A\otimes B[/itex]. He claims that the interpretation of this is that the correlations are completely determined by the state operator and that the state operator is completely determined by the correlations.
 
  • #11
Fredrik said:
I have a question about appendix A in Mermin's article, but someone who knows this stuff well might be able to answer without reading it. Why do expressions of the form [itex]\mbox{Tr }A\otimes B[/itex] represent correlations between subsystems? :confused:
Think about an ordinary Hilbert space for a moment. Suppose you have two state
operators [itex]\rho_1, \rho_2[/itex]. How do you compute the correlation between these states?
[itex]Tr(\rho_1 \rho_2)[/itex], right?
(If that's not clear, think about the case when both are pure...)

The same principle applies in a the tensor product Hilbert space (operators therein are in
general a sum of the tensor products of operators in the component subspaces).

What Mermin does in appendix A is to prove that any matrix element of the state operator
(I don't want to call it "density matrix" when it's an operator)
"Density matrix" and "state operator" are pretty much interchangeable terms
(though I prefer the latter because it's less confusing in infinite dimensions).
 
  • #12
strangerep said:
Think about an ordinary Hilbert space for a moment. Suppose you have two state
operators [itex]\rho_1, \rho_2[/itex]. How do you compute the correlation between these states?
[itex]Tr(\rho_1 \rho_2)[/itex], right?
(If that's not clear, think about the case when both are pure...)

The same principle applies in a the tensor product Hilbert space (operators therein are in
general a sum of the tensor products of operators in the component subspaces).
I understand what you're saying about pure states, but nothing beyond that. When they're both pure, we have

[tex]\mbox{Tr}(|\alpha\rangle\langle\alpha|\beta\rangle\langle\beta|)=\sum_n\langle n|\alpha\rangle\langle\alpha|\beta\rangle\langle\beta|n\rangle=\langle\beta|\Big(\sum_n|n\rangle\langle n|\Big)|\alpha\rangle\langle\alpha|\beta\rangle=|\langle\alpha|\beta\rangle|^2[/tex]

This is 1 when they're the same, and 0 when they're orthogonal, so it's a measure of how similar they are. The corresponding result when they're not pure is

[tex]\sum_i\sum_j a_i b_j |\langle\alpha_i|\beta_j\rangle|^2[/tex]

I don't see how this measures anything significant. If we e.g. take the alphas to be orthogonal to each other, and the two state operators to be the same, we get

[tex]\sum_i|a_i|^2[/tex]

So we don't even get the same value each time we consider two identical ensembles.

The type of correlation I think Mermin is talking about is the kind we encounter in a Stern-Gerlach experiment. When the silver atom has passed through the magnetic field, its state vector is |↑>|left>+|↓>|right>. This is a correlation between position states and spin states. If the state vector had contained the terms |↑>|right> and |↓>|left> with coefficients of the same magnitude as the ones in front of the other two terms, the states of the subsystems (spin-z and position-x) had been uncorrelated. A good measure of the degree of correlation should be large when the coefficients in front of the first two terms are large and the other two coefficients are small, or vice versa.

What makes this even weirder is that if A and B are observables, [itex]\mbox{Tr}(A\otimes B)[/itex] is independent of the state and therefore can't be a measure of correlation, and that if A and B are state operators, the result is always 1, because [itex]\mbox{Tr}(A\otimes B)=\mbox{Tr }A\cdot \mbox{Tr }B[/itex].

Perhaps I'm a bit slower than usual. I'm writing this one just before going to bed. Same thing with the question in my previous post.

Edit: I didn't mean to suggest that we're looking for a single number that tells us how correlated the subsystems are. We're looking for a set of numbers that in some way represent correlations of the type I have described, and together represent all of the information that's contained in the state operator.

strangerep said:
"Density matrix" and "state operator" are pretty much interchangeable terms
"State operator" feels like a natural thing to call it, but I honestly don't know if I've ever seen anyone use that term. Ballentine calls it "statistical operator". Sakurai calls it "density.." uhh I don't remember if he calls it "density operator" or "density matrix". Wikipedia calls it "density operator", but the article is titled "density matrix".
 
Last edited:
  • #13
D'oh. I felt that something was very wrong when I wrote my previous post, but I didn't know what. Now I see that Mermin says that the correlations are represented by quantities of the form [itex]Tr(\rho A\otimes B)[/itex], not [itex]Tr(A\otimes B)[/itex]. I haven't thought about what that means yet. I'm just posting to say that I've spotted this mistake.
 
  • #14
Fredrik said:
I felt that something was very wrong when I wrote my previous post, [...]
I'm sorry. My previous post was unhelpful. (I foolishly thought I could dash off an
answer without re-reading those appendices.)

Oh well... this is why I enjoy talking to you. You have a way of making me realize when
I don't understand things thoroughly enough. :-)

I'll try to post something more helpful later when I've re-studied some things properly.

In the meantime, you might also like to read Mermin's earlier paper quant-ph/9609013.
I get the feeling his later paper is partly an attempted response to questions about
this earlier paper about things that weren't clear to others. But he's partly only
succeeded in saying a lot more words without a proportionate increase in clarity (imho).
All this stuff about subsystem correlations is deceptively difficult. It looks easy enough
on a casual reading, but it's really quite subtle.

I suspect I'm yet to emerge from the "mist of incomprehension" that Herbut mentions
in his paper (0811.3674 [quant-ph]) in which he builds on Mermin's interpretation.
Actually, you might want to take a look at Herbut's paper as well. Among other things,
it reveals indirectly how hard it is to understand the depth of Mermin's ideas properly.
 
  • #15
The calculation in appendix A is pretty straightforward. If [itex]\{|\psi_\mu\rangle\}[/itex] is a basis for a Hilbert space, then [itex]\{|\psi_\mu\rangle\langle\psi_\nu|\}[/itex] is a basis for the algebra of operators. This follows immediately from

[tex]X|\alpha\rangle=\sum_{\mu,\nu}|\psi_\mu\rangle\langle\psi_\mu|X|\psi_\nu\rangle\langle\psi_\nu|\alpha\rangle[/tex]

The identity

[tex](|\alpha\rangle\langle\beta|)^\dagger=|\beta\rangle\langle\alpha|[/tex]

implies that the [itex]|\psi_\mu\rangle\langle\psi_\nu|[/itex] operators aren't hermitian in general. Mermin wants to work with hermitian operators, so he uses a standard trick to express the basis operators in terms of hermitian operators. Define [itex]K_{\mu\nu}=|\psi_\mu\rangle\langle\psi_\nu|[/itex] and write

[tex]K_{\mu\nu}=\frac{K_{\mu\nu}+K_{\mu\nu}^\dagger}{2}+\frac{K_{\mu\nu}-K_{\mu\nu}^\dagger}{2}=\frac{K_{\mu\nu}+K_{\nu\mu}}{2}+i\frac{K_{\mu\nu}-K_{\nu\mu}}{2i}=M^r_{\mu\nu}+iM^i_{\mu\nu}[/tex]

Mermin considers a system that consists of two subsystems. He writes the basis vectors as [itex]|\psi_\mu,\phi_\alpha\rangle=|\psi_\mu\rangle\otimes|\phi_\nu\rangle[/itex]. He calls the state operator W, and uses the above to express an arbitrary matrix element of W in terms of the hermitian M operators for the two subsystems. (He uses the letter N for the "M operators" of the second subsystem).

[tex]\langle\psi_\nu,\phi_\beta|W|\psi_\mu,\phi_\alpha\rangle=\sum_{\rho,\gamma}\langle\psi_\nu,\phi_\beta|\psi_\rho,\psi_\gamma\rangle\langle\psi_\rho,\psi_\gamma|W|\psi_\mu,\phi_\alpha\rangle=\sum_{\rho,\gamma}\langle\psi_\rho,\psi_\gamma|W|\psi_\mu,\phi_\alpha\rangle\langle\psi_\nu,\phi_\beta|\psi_\rho,\psi_\gamma\rangle[/tex]

[tex]=\mbox{Tr}\big(W|\psi_\mu,\phi_\alpha\rangle\langle\psi_\nu,\phi_\beta|\big)=\mbox{Tr}\Big(W\Big(|\psi_\mu\rangle\langle\psi_\nu|\otimes|\phi_\alpha\rangle\langle\phi_\beta|\Big)\Big)[/tex]

[tex]=\mbox{Tr}\Big(W\Big((M^r_{\mu\nu}+iM^i_{\mu\nu})\otimes(N^r_{\alpha\beta}+iN^i_{\alpha\beta})\Big)\Big)[/tex]

[tex]=\mbox{Tr}\big(W\big(M^r_{\mu\nu}\otimes N^r_{\alpha\beta}\big)\big) +i\mbox{Tr}\big(W\big(M^r_{\mu\nu}\otimes N^i_{\alpha\beta}\big)\big) +i\mbox{Tr}\big(W\big(M^i_{\mu\nu}\otimes N^r_{\alpha\beta}\big)\big) -\mbox{Tr}\big(W\big(M^i_{\mu\nu}\otimes N^i_{\alpha\beta}\big)\big)[/tex]

So the calculation isn't too hard. Now I just need to understand why we're doing it. :smile:

I'm including links to the three papers you've mentioned, just to make it easier to access them from here:

http://arxiv.org/abs/quant-ph/9801057
http://arxiv.org/abs/quant-ph/9609013
http://arxiv.org/abs/0811.3674

I've read the easy parts of the original Ithaca paper now, but I haven't tried to understand the details of the proofs. Herbut's paper looks much more difficult.
 
Last edited:
  • #16
Fredrik said:
So the calculation isn't too hard. Now I just need to understand why we're doing it. :smile:

Not sure what your current worry is. What this proof shows or the meaning of [itex] Tr(\rho A\otimes B) [/itex]?
 
  • #17
yossell said:
Not sure what your current worry is. What this proof shows or the meaning of [itex] Tr(\rho A\otimes B) [/itex]?
Aren't those two options the same? I'd like to know why the terms in the final result represent correlations between subsystems.
 
  • #18
Fredrik said:
I'd like to know why the terms in the final result represent correlations between subsystems.
OK, I'll try to say something more helpful this time.

First, a quick review...

The "correlation" [itex]\rho_{X,Y}[/itex] between two observables X,Y in a given state is defined as
[tex]
\rho_{X,Y} ~:=~ \frac{cov(X,Y)}{\sigma_X ~ \sigma_Y} ~,
[/tex]
where the terms in the denominator are the standard deviations
of the respective observables, and the numerator is their covariance:

[tex]
cov(X,Y)_\Psi ~:=~ \langle \; (X - \bar{X}_\Psi) \; (Y - \bar{Y}_\Psi) \; \rangle_\Psi ~,
[/tex]

where the barred terms denote the expectations of the respective observables
in the state [itex]\Psi[/itex].

The "correlation" defined above can be viewed as a "normalized" covariance.

Since the barred terms are just scalars, a few lines of linear algebra shows that

[tex]
cov(X,Y)_\Psi ~=~ \langle X Y \rangle_\Psi ~-~ \bar{X}_\Psi \, \bar{Y}_\Psi ~.
[/tex]

So far, this is just equivalent to the well-known result that the covariance
between two observables is zero if and only if their expectation factorizes
as shown above.

Now we apply this to the case of a tensor product system, where X is an
operator acting nontrivially only on the first subsystem, and Y acts
nontrivially only on the second. Let W be the tensor product state.

[tex]\bar{X}_W[/tex] clearly involves only the first subsystem, so
[tex](X - \bar{X}_W)[/tex] is also an observable acting nontrivially
only on the first subsystem. Similarly for Y and the second subsystem.

Likewise, the standard deviation [tex](\sigma_X)_W[/tex] only involves
quantities from the first subsystem nontrivially, so we can divide
[tex](X - \bar{X}_W)[/tex] by [tex](\sigma_X)_W[/tex] and get yet another
observable that acts nontrivially only on the first subsystem.
(And similarly for Y and the second subsystem.)

So let's consider the observable

[tex]
A ~:=~ \frac{X - \bar{X}}{\sigma_X}
[/tex]

and similarly B in terms of Y.

(Of course, I'm ignoring the special case where the standard
deviation is zero, but in that case we'd just use covariances alone.)

We can now see that correlations between A and B can be
expressed simply as

[tex]
\langle A B \rangle_W
[/tex]

provided we note the abuse of notation by which "A" and "B" have been
extended to mean

[tex]
A ~\equiv~ A \otimes 1 ~~;~~~~~ B ~\equiv~ 1 \otimes B
[/tex]

To tie this in more closely to Mermin's SSC theorem, we note that the
theorem says "...values of [tex]Tr(W\, A\otimes B)[/itex] for an appropriate
set of observable pairs A,B ..." (my italics).

Is this adequate to explain how/why the far RHS of Mermin's eq(30)
represents "the values of the subsystem correlations between all the
A's and B's", and hence that they "are enough to determine all the
matrix elements of W in a complete set of states for the total system [...]" ?
 
  • #19
Thank you strangerep. That was helpful. I didn't know those definitions. I see that the definition

[tex]A=\frac{X-\langle X\rangle}{\Delta X}[/tex]

implies

[tex]\langle A\rangle=\frac{\langle X\rangle-\langle X\rangle}{\Delta X}=0[/tex]

[tex]\langle A^2\rangle=\left\langle\left(\frac{X-\langle X\rangle}{\Delta X}\right)^2\right\rangle=\left\langle\frac{X^2+\langle X\rangle^2-2X\langle X\rangle}{(\Delta X)^2}\right\rangle=\frac{\langle X^2\rangle-\langle X\rangle^2}{(\Delta X)^2}=1[/tex]

and therefore

[tex](\Delta A)^2=\langle A^2\rangle-\langle A\rangle^2=1[/tex]

and similarly for B. So the correlation of A and B is equal to the covariance of A and B, which is equal to

[tex]\langle AB\rangle-\langle A\rangle\langle B\rangle=\langle AB\rangle
=\mbox{Tr}(WAB)=\mbox{Tr}(W(A\otimes I)(I\otimes B))=\mbox{Tr}(W(A\otimes B))[/tex]

where W is the state operator.

I also see that this can be related to what I said before about the state |↑>|left>+|↓>|right>. Consider the pure state

a|↑>|left> + b|↓>|right> + c|↑>|right> + d|↓>|left>

What I said there can be translated to "the states of the subsystems are correlated when the magnitudes of a and b are large and the magnitudes of c and d are small". Your definitions agree with that pretty well. When the state is [itex]|\psi\rangle[/itex]=|↑>|left>+|↓>|right>, we have

[tex]\langle\psi|S_z|\psi\rangle\langle\psi|X|\psi\rangle=\Big(\langle\uparrow|S_z|\uparrow\rangle+\langle\downarrow|S_z|\downarrow\rangle\Big)\Big(\langle L|X|L\rangle+\langle R|X|R\rangle\Big)[/tex]

[tex]=\langle\uparrow|S_z|\uparrow\rangle\langle L|X|L\rangle+\langle\uparrow|S_z|\uparrow\rangle\langle R|X|R\rangle+\langle\downarrow|S_z|\downarrow\rangle\langle L|X|L\rangle+\langle\downarrow|S_z|\downarrow\rangle\langle R|X|R\rangle[/tex]

where I have abbreviated "left" and "right" to L and R, and X is the "position" operator that only has two eigenstates, "left" and "right". Your definition compares this to

[tex]\langle\psi|S_zX|\psi\rangle=\langle\uparrow|S_z|\uparrow\rangle\langle L|X|L\rangle+\langle\downarrow|S_z|\downarrow\rangle\langle R|X|R\rangle[/tex]

Your definition says that the correlation of Sz and X in the state [itex]|\psi\rangle[/itex] is just the difference between these two quantities, which is

[tex]\langle\uparrow|S_z|\uparrow\rangle\langle R|X|R\rangle+\langle\downarrow|S_z|\downarrow\rangle\langle L|X|L\rangle=\langle\phi|S_z|\phi\rangle\langle\phi|X|\phi\rangle[/tex]

where [itex]|\phi\rangle=|\uparrow\rangle\otimes|R\rangle+|\downarrow\rangle\otimes |L\rangle[/itex]. So the correlation is (at least in this case) the difference between two expressions of the form [itex]\langle\chi|S_z|\chi\rangle\langle\chi|X|\chi\rangle[/itex], where [itex]|\chi\rangle[/itex] is first chosen to consist of the two "desirable terms" from a|↑>|left> + b|↓>|right> + c|↑>|right> + d|↓>|left> and then of the two "undesirable" terms.

I have to leave the computer now...will add a few more things later.
 
Last edited:
  • #20
Fredrik said:
I'm not familiar with the information interpretation. (Not even sure if I've heard of it). Mermin seems to agree that we need an additonal assumption on top of QM to avoid having to deal with many worlds, and his assumption is that consciousness is weird enough to invalidate the conclusion that realism implies many worlds.
The main difference between MWI and the information interpretation (II) is that MWI claims that the wave function is an objectively real entity, while II claims that the wave function is not an objectively real entity, but a mathematical function that describes our information about reality. An extreme II (which Mermin seems to adopt in his "correlations without correlata" mantra) asserts that objective reality does not even exist, i.e., that there is only information and nothing else. Another prominent supporter of such an extreme II is Zeilinger.
 
  • #21
Demystifier said:
The main difference between MWI and the information interpretation (II) is that MWI claims that the wave function is an objectively real entity, while II claims that the wave function is not an objectively real entity, but a mathematical function that describes our information about reality. An extreme II (which Mermin seems to adopt in his "correlations without correlata" mantra) asserts that objective reality does not even exist, i.e., that there is only information and nothing else. Another prominent supporter of such an extreme II is Zeilinger.

I do not know Zeillingers reasoning in detail, but this angle sounds more reasonable to my taste.

One step further and we are quite close to a subjetive information interpretations which is close to the many mind interpretations http://en.wikipedia.org/wiki/Many-minds_interpretation

"The many-minds interpretation of quantum mechanics extends the many-worlds interpretation by proposing that the distinction between worlds should be made at the level of the mind of an individual observer."

Of course, that absolute ridicilous part of many minds if you seriously think that the human MIND has anything to do with it.

But if we replace many human minds, with just many physical observers, and add the constraint that any physical observer encodes only a finite amount of information then we get something that makes sense and that has nothing to do with human brain.

My view is very close to this, and here objective reality is emergent as a result of interacting observers. The mutual selection resulting from the interacting makes objective reality the information that is shared by the local neighbourhood of observers. An observer has to comply to this "consensus" in order to stay stable - but this does not ban other local agreements that are only weakly interacting with the first group.

/Fredrik
 
  • #22
Fredrik said:
I have to leave the computer now...will add a few more things later.
I found a way to understand the definitions better, at least when the Hilbert spaces are finite dimensional. Suppose that [itex]\{|\psi_\mu\rangle\}_{\mu=1}^N[/itex] is a basis for [itex]H_1[/itex] that consists of eigenvectors of A and that [itex]\{|\phi_\alpha\rangle\}_{\alpha=1}^M[/itex] is a basis for [itex]H_2[/itex] that consists of eigenvectors of B. An arbitrary state [itex]|f\rangle\in H_1\otimes H_2[/itex] can be written as

[tex]|f\rangle=\sum_{\mu,\alpha}c_{\mu\alpha}|\psi_\mu\rangle\otimes|\phi_\alpha\rangle[/tex]

We want to define a measure of how correlated the basis vectors are, and we want states that can be expressed in the form

[tex]\sum_{\mu}c_{\mu}|\psi_\mu\rangle\otimes|\phi_\mu\rangle[/tex]

or

[tex]\sum_{\alpha}c_{\alpha}|\psi_\alpha\rangle\otimes|\phi_\alpha\rangle[/tex]

to be considered "completely", or at least "very", correlated, and states with [itex]|c_{\mu\alpha}|^2=\frac{1}{NM}[/itex] for all [itex]\mu,\alpha[/itex] to be considered completely uncorrelated.

Now let [itex]X=X_1\otimes X_2[/itex] be an arbitrary operator. We have

[tex]\langle f|X|f\rangle=\langle f|\sum_{\mu,\alpha}c_{\mu\alpha}X_1|\psi_\mu\rangle\otimes X_2|\phi_\alpha\rangle=\sum_{\nu,\beta}\sum_{\mu,\alpha}c_{\nu\beta}^*c_{\mu\alpha}\langle\psi_\nu|X_1|\psi_\mu\rangle\langle\phi_\beta|X_2|\phi_\alpha\rangle[/tex]

In the special case when the [itex]|\psi_\mu\rangle[/itex] are eigenstates of [itex]X_1[/itex] and the [itex]|\phi_\alpha\rangle[/itex] are eigenstates of [itex]X_2[/itex], this simplifies to

[tex]\sum_{\mu,\alpha}|c_{\mu\alpha}|^2\langle\psi_\mu|X_1|\psi_\mu\rangle\langle\phi_\alpha|X_2|\phi_\alpha\rangle[/tex]

and when we're dealing with a completely uncorrelated state, this simplifies to

[tex]\frac{1}{NM}\sum_{\mu,\alpha}\langle\psi_\mu|X_1|\psi_\mu\rangle\langle\phi_\alpha|X_2|\phi_\alpha\rangle=\frac{\mbox{Tr }X_1\ \mbox{Tr }X_2}{NM}[/tex]

So in particular, we have

[tex]\langle f|A|f\rangle=\langle f|A\otimes I|f\rangle=\frac{\mbox{Tr }A\ \mbox{Tr }I}{NM}=\frac{\mbox{Tr }A}{N}[/tex]

[tex]\langle f|B|f\rangle=\langle f|I\otimes B|f\rangle=\frac{\mbox{Tr }I\ \mbox{Tr }B}{NM}=\frac{\mbox{Tr }B}{M}[/tex]

and

[tex]\langle f|AB|f\rangle=\langle f|A\otimes B|f\rangle=\frac{\mbox{Tr }A\ \mbox{Tr }B}{NM}=\langle f|A|f\rangle\langle f|B|f\rangle[/tex]

in perfect agreement with your definitions of "covariance" and "correlation", since this implies that both of them would be =0 for a state that I consider completely uncorrelated.

I have to go to bed, so I can't continue this now, but I hope that I'll also be able to show something interesting about maximally correlated states, i.e. states with [itex]c_{\mu\alpha}=c_\mu\delta_{\mu\alpha}[/itex].
 

FAQ: Tensor products and subsystems

What is a tensor product?

A tensor product is a mathematical operation that combines two vector spaces to create a new vector space. It is often used to represent the relationship between two systems or subsystems in physics and engineering.

How is a tensor product different from a regular product?

A tensor product takes into account the multidimensional nature of vector spaces, while a regular product only considers one-dimensional quantities. Additionally, the result of a tensor product is a tensor, which is a multidimensional array, while the result of a regular product is a scalar or vector.

What are the applications of tensor products?

Tensor products have many applications in physics, engineering, and computer science. They are used to represent physical quantities, such as force and velocity, in multiple dimensions. They are also used in machine learning and image processing to analyze and manipulate multidimensional data.

How are tensor products related to subsystems?

Tensor products are often used to describe the relationship between subsystems in a larger system. For example, in quantum mechanics, the state of a composite system can be represented as a tensor product of the states of its subsystems. This allows for the analysis of the behavior of the composite system based on the behavior of its subsystems.

Can tensor products be performed on any type of vector space?

Yes, tensor products can be performed on any type of vector space, including finite-dimensional and infinite-dimensional spaces. However, the resulting tensor may have different properties depending on the type of vector space involved. For example, tensor products of infinite-dimensional spaces may have infinite dimensions themselves, making them more challenging to work with.

Similar threads

Replies
1
Views
2K
Replies
12
Views
2K
Replies
3
Views
940
Replies
1
Views
896
Replies
61
Views
3K
Replies
3
Views
1K
Replies
7
Views
1K
Replies
1
Views
1K
Replies
4
Views
866
Back
Top