On the use of Hilbert Spaces to represent states

In summary, there are a lot of coherent definitions of Hilbert spaces that are suitable for representing states in QM. These definitions include that the space should be linear, have an inner product, and be complete. These properties are necessary for the Superposition Principle and to allow for continuous transformations between pure states. The mathematical structure of the set of observables also leads to the use of Hilbert spaces. The axioms of QM can be derived from these mathematical properties and imply the use of a Hilbert space. The use of complex numbers is required for interference effects and for allowing continuous evolution of physical systems.
  • #36
the_pulp said:
[...] an argument to support the linearity hypothesis (if it is the case that there indeed is an argument out there).
In this context, it is eye-opening to study Ballentine section 7.1 in which he derives the well-known half-integral angular momentum spectrum from nothing more than the SO(3) generators, their commutation relations, and the assumption that they are represented as Hermitian operators on an abstract Hilbert space.

AFAIK, there is no other method of deriving the angular momentum spectrum that does not involve use of such a linear Hilbert space.

Though this is not a conclusive "it-can't-be-anything-else" argument, it does set quite a high bar that alternative approaches must clear.
 
Physics news on Phys.org
  • #37
strangerep said:
In this context, it is eye-opening to study Ballentine section 7.1 in which he derives the well-known half-integral angular momentum spectrum from nothing more than the SO(3) generators, their commutation relations, and the assumption that they are represented as Hermitian operators on an abstract Hilbert space.
Maybe I'm misinterpreting the intent of your statement, but it doesn't sound right.

In the proof idea as I know it, you don't presuppose a representation of SU(2) at all.

Instead, the argument proceeds (roughly) by looking at all representations, and invoking the fact that any element of the spectrum appears as an eigenvalue in some representation.

_____________________________________________On an unrelated note, while I mentioned I find the use of Hilbert spaces (and the linearity therein) as merely a matter of mathematical technique, I do find the fact the observable algebra is an algebra to be curious. Why do expressions like PX and P+X make sense??
 
  • #38
Fredrik said:
What atyy is getting at is that Gleason's theorem doesn't tell us why we should make that kind of probability assignments in the first place.

The way I see it is you have two choices - either a deterministic or statistical theory. The deterministic case however is contained in the statistical one - it simply means the probabilities are 0 or 1. What Gleasons Theorem shows is you can't assign 0 and 1 only to the subspaces of a vector space thus ruling out the deterministic case. The Kochen-Specker Theorem does as well but it really is a simple corollary to Gleasons Theroem.

The issue with Glesasons theorem is its hidden assumption of non contextuality - you assume the probably assignment does not depend on the way the rest of the vector space is partitioned off by a resolution of the identity. However this fits nicely in with my approach to QM which is based on invariance.

Thanks
Bill
 
Last edited:
  • #39
bhobba said:
The way I see it is you have two choices - either a deterministic or statistical theory. The deterministic case however is contained in the statistical one - it simply means the probabilities are 0 or 1. What Gleasons Theorem shows is you can't assign 0 and 1 only to the subspaces of a vector space thus ruling out the deterministic case. The Kochen-Specker Theorem does as well but it really is a simple corollary to Gleasons Theroem.

The issue with Glesasons theorem is its hidden assumption of non contextuality - you assume the probably assignment does not depend on the way the rest of the vector space is partitioned off by a resolution of the identity. However this fits nicely in with my approach to QM which is based on invariance.

Thanks
Bill

Does one have to assume that measurements are projective operations?
 
  • #40
atyy said:
Does one have to assume that measurements are projective operations?

No. The idea is up to a multiplicative constant (the superposition principle implies this - superimposing a state with itself gives the same state) you know the outcome of any observation is an element of the vector space so you are assigning probabilities to the projection operators in a resolution of the identity (the space generated by multiplying any element by a constant is a subspace so automatically defines a projection operator). Gleasons theorem shows the usual trace formula for probabilities is the only one that can be defined. As mentioned there is a hidden assumption - namely the probability does not depend on what resolution of the identity a projection operator is part of - which is pretty much what a vector space is all about anyway (ie the elements do not depend on the whatever basis you chose) but is nonetheless an assumption.

You then associate each projection operator with a real number so a hermitian operator defines a resolution of the identity as the eigenvectors with the eigenvalues being the value associated with each projection operator.

Thanks
Bill
 
Last edited:
  • #41
Hurkyl said:
Maybe I'm misinterpreting the intent of your statement, but it doesn't sound right.
Do you have a copy of Ballentine at hand? I believe what I said does indeed correspond to what he does.

In the proof idea as I know it, you don't presuppose a representation of SU(2) at all.

Instead, the argument proceeds (roughly) by looking at all representations, and invoking the fact that any element of the spectrum appears as an eigenvalue in some representation.
I haven't seen it done with that emphasis. Can you give me a (readable) reference?

On an unrelated note, while I mentioned I find the use of Hilbert spaces (and the linearity therein) as merely a matter of mathematical technique, I do find the fact the observable algebra is an algebra to be curious. Why do expressions like PX and P+X make sense??
Maybe because the symmetry/dynamical groups tend to be Lie groups, and physical representations are classified via Casimirs thereof? (Also, continuity and differentiability, etc, imply certain properties for the generators.)

BTW, if the same question is posed in the classical regime, we have things like ##J = X \times P##, and also ##H = P^2 + X^2##, etc. Is this curious or boring? :-)
 
  • #43
strangerep said:
No. POVM's are a more general approach:

http://en.wikipedia.org/wiki/POVM

Indeed it is and the modern proof of Gleasons Theorem makes use of them instead of resolutions of the identity - but POVM's are derivable from projections of higher dimensional resolutions of the identity via Newmarks Theorem.

Thanks
Bill
 
  • #44
Helloooo! I opened this thread because I was asking about linearity!

And here I arrive to Linearity. Is there any other axiom that you know that will sound more physical to me (perhaps you ask me how would you know what sounds more physical to me? I just think you know) and through which linearity can be demostrated?

(Nevertheless, pretty interesting the talk about Gleasons Theorems)
 
  • #45
A question relating linearity and measurements:

Compare the classical and Schroedinger wave equations. Both have linear solution spaces. However, a solution of the classical equation is not considered a sate, because it contains only information about position, not velocity, and both can be measured. However, a solution of the Schroedinger equation is considered a state. Is this because the quantum notion of position or velocity is different from the classical one?
 
  • #46
atyy said:
A question relating linearity and measurements:

Compare the classical and Schroedinger wave equations. Both have linear solution spaces. However, a solution of the classical equation is not considered a sate, because it contains only information about position, not velocity, and both can be measured. However, a solution of the Schroedinger equation is considered a state. Is this because the quantum notion of position or velocity is different from the classical one?

I don't think so. The classical wave equation is second order w.r.t. time, the solution plus its time derivative is the state. The state should be something that determines the future evolution of the system and in the classical case the solution of the equation (at a given time) is not enough for initial conditions.
 
  • #47
strangerep said:
I haven't seen it done with that emphasis. Can you give me a (readable) reference?
Unfortunately, it's not something I've actually worked through. Here is wikipedia's page on the topic. The result they state near the bottom is what I remember -- the fact you derive is not a theorem about the behavior of SU(2) in a particular representation: instead it is a complete classification of all irreducible finite-dimensional representations.

Mulling it over, I think the statement you are referring to (I don't have the text) is effectively equivalent, just phrased differently.



BTW, if the same question is posed in the classical regime, we have things like ##J = X \times P##, and also ##H = P^2 + X^2##, etc. Is this curious or boring? :-)
Boring. On any particular state, P and X are definite numbers, and so P^2+ X^2 makes sense in the classical regime: "Measure position and momentum, square them, then add them".

I can dramatically point out the issue with spin states. If X and Y are spin about the X and Y axis, if we try to think of the spectrum of an operator as being the possible outcomes of a hypothetical measurement, and we try to interpret X+Y in the same way as I mentioned classically, we run into the problem that the interpretation says [itex]1 + 1 = \sqrt{2}[/itex]! (or more accurately, if we add either of [itex]\pm 1[/itex] and either of [itex]\pm 1[/itex], we get either of [itex]\pm \sqrt{2}[/itex])
 
  • #48
Hurkyl said:
Mulling it over, I think the statement you are referring to (I don't have the text) is effectively equivalent, just phrased differently.
After a night's sleep, I now get what you were previously saying. A more careful revision of what I said should include some caveats about superselection sectors for different Casimir values.
I can dramatically point out the issue with spin states. If X and Y are spin about the X and Y axis, if we try to think of the spectrum of an operator as being the possible outcomes of a hypothetical measurement, and we try to interpret X+Y in the same way as I mentioned classically, we run into the problem that the interpretation says [itex]1 + 1 = \sqrt{2}[/itex]! (or more accurately, if we add either of [itex]\pm 1[/itex] and either of [itex]\pm 1[/itex], we get either of [itex]\pm \sqrt{2}[/itex])
I wonder whether one hits the same issue if the quantum case were constructed using POVMs corresponding to spin-coherent states. It's been a while since I looked at these and it's too early in the morning right now... :-)
 
  • #49
bhobba said:
[...] POVM's are derivable from projections of higher dimensional resolutions of the identity via Newmarks Theorem.
I guess you meant "Neumark", or "Naimark"? For other readers, this is also mentioned in the Wiki page on POVMs previously linked. More specifically:
http://en.wikipedia.org/wiki/Neumark's_dilation_theorem

BTW, how does this work in the case of unbounded operators and continuous spectra? The construction of a "higher dimensional" Hilbert space seems like it might be a bit tricky in that context, and the Wiki page seems to deal only with bounded operators.
 
Last edited:
  • #50
the_pulp said:
Helloooo! I opened this thread because I was asking about linearity!
Sorry if you've received the impression that your thread was being hijacked. The posts were indeed mostly relevant to your topic, but clearly that was not obvious. [Although... perhaps the stuff on Naimark's theorem should be moved to a different thread.]

Did you get (or study) the point I was trying to make in my earlier reference (post #36) to Ballentine's derivation of quantum angular momentum spectra?
 
Last edited:
  • #51
Sorry if you've received the impression that your thread was being hijacked. The posts were indeed mostly relevant to your topic, but clearly that was not obvious. [Although... perhaps the stuff on Naimark's theorem should be moved to a different thread.]

Did you get (or study) the point I was trying to make in my earlier reference (post #36) to Ballentine's derivation of quantum angular momentum spectra?

No, I did not study it (I thought it was related to another thing, but it seems that it wasnt!). I am going to do it today at night. In the meantime if you can give me a short summary of what is the relation between that and my doubt, very welcome.
 
  • #52
the_pulp said:
No, I did not study [Ballentine] (I thought it was related to another thing, but it seems that it wasnt!). I am going to do it today at night. In the meantime if you can give me a short summary of what is the relation between that and my doubt, very welcome.
Ballentine ch1 explains how we have physically meaningful observable quantities, and that we wish to construct probability measures over them.

Doing this in a linear space rather than a nonlinear space is simply easier. There's no point using a technically less-convenient space if we don't need to. Similarly, a metric-topological space has lots more nice properties than more general topological spaces, so we don't use the latter unless the physics can't be modeled satisfactorily in the former. Rigged Hilbert space is a step towards a more general structure, but we don't need to use that for simple cases.

My point about the quantum angular momentum spectrum was about how a basic Hilbert space framework and Hermitian operators already gives an important physical result: the half-integral angular momentum spectrum.

Not sure what else I can usefully say here. Maybe after reading the earlier chapters of Ballentine, you should re-articulate whatever issues remain unresolved...
 
  • #53
the_pulp said:
Helloooo! I opened this thread because I was asking about linearity!

Another possible way I think one could motivate linearity is to note that elementary particles in quantum theory are considered point particles so they have no size (this also happens in the linear classical mechanics and electrodynamics theories and their idealized point particles) . Of course if that assumption was not physically exact I suppose it would introduce some non-linearity in the picture and QM as we know it would be just a linear approximation for a point.

We do know that we have to recur to QFT perturbative methods to have good approximations to certain experimental results (I'm thinking about the Lamb shift or the anomalous magnetic moment of the electron that are not so well approximated by either relativistic or NRQM).
 
Last edited:
  • #54
the_pulp said:
Yes, that was sort of my route. I was not very sure about:
1) The linearity of the states (I think that when someone says to me that the state space should be linear because of the superposition principle it sounds like "the state space should be linear because the state space should be linear").
2) The use of complex numbers

And, related to the born rule, I was not using Gleasons theorems because I was thinking about Saunder's paper, but it is more or less the same (I will read your paper). Now, Ballentine and the link you sent me about complex numbers I think I've got the ideas a little bit more ordered.

Thanks!

there is nonlinear quantum theory.
 
  • #55
yoda jedi said:
there is nonlinear quantum theory.

Not in any of the various QM textbooks I've consulted, do you have some reference?
 
  • #56

Similar threads

Replies
61
Views
4K
Replies
5
Views
1K
Replies
44
Views
4K
Replies
26
Views
8K
Replies
8
Views
2K
Replies
19
Views
5K
Replies
7
Views
1K
Replies
2
Views
1K
Back
Top