Understanding Projectors in Quantum Mechanics: A Mathematical Approach

In summary: Sum_n P_{e_n} = P_{\sum_n e_n} ...This is where things get a little fuzzy. It looks like you are saying that for every pair (\psi, \psi_n) in H,Sum_n P_{e_n} = P_{\sum_n e_n}
  • #106
Response to post #97

Sammywu said:
1) I did not know | q > < q | is not a projector. I have to think about that.
It would more properly be called a "projector density".

In terms of the projector PI onto an interval I = (q, q+∆q) , ∆q > 0 , it would be defined as

lim∆q → 0+ PI/∆q .

As you can see, this is a "density", and as such when it is squared, the same thing does not come back ... instead, something infinite comes back.
____________
2). I did hesiate to write [tex] Q | \psi > = q | \psi > [/tex] in the same reasons you mentioned, but in both Leon's Ebook and another place I did see their mentioning about the Q's defintion is [tex] Q | \psi > = q | \psi > [/tex]. Just as I mentioned, the only reason I could see this "make sense" is by either
[tex] Q | \psi > = \int q |q > < q> dq [/tex] or
[tex] Q | \psi > = q \psi ( q ) [/tex] in the form of wavefunvtions.
I think what you saw in Leon's book and any other place is

[1] Qψ(q) = qψ(q) .

This is not the same as Q|ψ> = q|ψ>.

Strictly speaking, we should not be using the same "Q" in both cases. They are different 'objects'. The first Q acts on a "function space" (so, now I will use Q to denote it). The second Q acts on a "ket space". Although the two spaces are "isomorphic", they are nevertheless formally distinct.

Now, what is the difference then between [1], which I now write as

[1] Qψ(q) = qψ(q) ,

and Q|ψ> = q|ψ> ? Well ... let's write the Q|ψ> = q|ψ> in "q-space". But(!) wait ... we've already used the "q" ... so, let's write it in "q'-space". We then have

<q'|Q|ψ> = <q'|(Q|ψ>) = <q'|(q|ψ>) = q<q'|ψ> .

The LHS we write as <q'|Q|ψ> = Q'ψ(q'), and the RHS we write as q<q'|ψ> = qψ(q'). Since LHS = RHS, we then have

[2] Q'ψ(q') = qψ(q') .

Do you see the difference between [1] and [2]? ... In [1], Q takes the function ψ(q) to a new function qψ(q), for which the two 'instances' of q are the same variable. But in [2], Q' takes ψ(q') to qψ(q'), for which q' is the variable and q is a constant ... and the fact that q is a CONSTANT here is what makes [2] an "eigenvalue" equation. The q in [1] is not a constant – equation [1] is not an "eigenvalue" equation. Equation [1] is what defines the action of Q in "q-space"!
____________
3). I think your defining that [tex] \psi ( q ) = < q \ \psi > [/tex] actually will make many calculations I did in showing in general
[tex] < \psi \prime | \psi > = \int \overline{\psi \prime ( q ) } \psi ( q ) dq [/tex]
much more straighforward thamn my cumbersome calculations.

But one problem is then what is < q | q >. The discrete answer will be it needs to be one. Which of course will lead to some kind of conflicts in a general requirement of
[tex] \int \overline{\psi ( q ) } \psi ( q ) dq = 1 [/tex] .

This is probably related to the eigenprojector [tex] P_{I_n} [/tex] you mentioned.
<q|q> = ∞. That is what it means to say that these kets do not "belong" to the Hilbert space – they do not have finite norm ... and, on account of that, we say that these kets are "generalized". Of course, when the two q's are distinct, say q ≠ q', then <q|q'> = 0. But we can say more than just "∞" and "0" ... we can say more precisely

<q|q'> = δ(q - q') , for all q,q' Є R .

Now, going back to what I said about a "discrete" position observable:
To make the position observable "discrete" we want to have a small "interval" In corresponding to each "point" qn, say

qn = n∙∆q , and In = (qn-∆q/2 , qn+∆q/2] .

Then, our "discrete" position observable, call it Q∆q, will be a degenerate observable. It will have eigenvalues qn and corresponding eigenprojectors (not kets!) PI_n. That is,

Q∆q = ∑n qnPI_n .
Each eigenvalue qn of Q∆q is infinitely-degenerate. If you think about it, then you will realize that each degenerate eigensubspace En , corresponding to qn , is nothing but the set of square-integrable functions on the interval (qn- ∆q/2 , qn+ ∆q/2). The 'magic' of it all is that ... in the limit ∆q → 0+ ... each ∞-dimensional eigensubspace En 'collapses'(!) into SOMETHING which can be characterized by a single object |q> whose interpretation is that of a "vector density", called a "generalized vector", and which has an infinite norm ... but nevertheless ... stands in the relation Q|q> = q|q>. ... abracadabra ... and so, you get a "generalized ket".
____________
4). Actually, I noticed my deduction has a contradition unresolved.

There is an issue to be resolved in my eigenfunction for "position" [tex] q_0 [/tex] as
[tex] lim_n { \rightarrow \infty } \int \delta_n(q - q_0) | q> dq [/tex]
. The problem here is whether the norm of \delta_n( q - q_0 ) shall be one or its direct integration shall be one.
If the norm shall be one, then it shall be altered to be its square root then.
Its direct integration shall be one. Look at the definition:

δn(q) ≡
n , q Є In
0 , otherwise ,

where In = (-[2n]-1 , [2n]-1) .

Its norm, however, is

sqrt{ ∫ |δn(q)|2 dq } = √n .

In the limit, this is ∞ ... as required.
 
Last edited:
Physics news on Phys.org
  • #107
Response to post #105

I am tentatively bypassing a response to your posts #98-102 in order to address your current concerns.
______________
You know that after you pointed out the objects as "vector density" and "infinitesmal vector", I have made some corrections on my late posts.
I will keep this in mind when I get to them.
______________
Actually, that was what initially confussed me. In my mind, I have this [tex] \psi_q [/tex] as the generalized eigen vector of eigen value q for the position operator Q. I have it confussed with |q>, which is the "vector density" for it.
Now, let's make sure we have gotten this straight. The object "|q>" is the "generalized eigenket for Q" ... and(!) it is also what I have referred to as a "vector density".

All "generalized eigenkets" are "vector densities".

The object which you are now referring to as "|ψq>" is what I have referred to as "belonging to a class of generalized vector", and it satisfies the relation d|ψq> = |q>dq, giving d|ψq> the interpretation of an "infinitesimal vector".

But ... the object "|ψq>" itself is neither(!) a "generalized eigenket" nor a "vector density"!
______________
Now I organized them, so I realized [tex] d \psi_q [/tex] is the "infinitismal vector" as "d|q)" in your writing.

The "|q>" as the "vector density" is [tex] d \psi_q/dq \prime [/tex] in my thought and "d|q)/dq" in your writing.

Now in that, this falls into place:

[tex] \psi = \int \psi(q) |q> dq = \int < q | \psi > < q | dq = \int | q > < q | \psi > dq [/tex]
for any [tex] \psi [/tex] .

In particular,
[tex] \psi_q \prime = \int | q > < q | \psi_q \prime > dq [/tex]
for the generlized eigenvector of eigenvalue [tex] q \prime [/tex] .
Where is the ket notation "| >" on your "ψ" and "ψq'"? And remember what I said above:

The object "|ψq>" is not(!) a "generalized eigenket" (... of Q).
______________
In order for the above EQ. to be true, My first impression was
[tex] < q | \psi_q \prime > = \delta ( q \prime - q ) dq [/tex]
, but actually it turns out wrong, it will give an object more like [tex] |q \prime> [/tex] in the sense of
[tex] \int \delta ( q \prime - q ) O(q) dq = O ( q \prime ) [/tex]
The condition which makes

q'> = ∫ |q><q|ψq'> dq

true is simply

∫ |q><q| dq = 1 ,

where the "1" here is the identity operator on the Hilbert space. Thus, there are three important properties to note about the |q>-family.

[1] Q|q> = q|q> ... "eigenkets" ,

[2] <q|q'> = δ(q - q') ... "generalized" (ortho-'normal') ,

[3] ∫ |q><q| dq = 1 ... "complete" .
______________
Actually, if I use
[tex] d \psi_q \prime /dq \prime \prime = | q \prime > [/tex]
, I can get
[tex] \psi_q \prime = \int | q \prime > dq \prime \prime [/tex]
.
This notation does not make sense. What you really mean to say is d|ψq> = |q>dq, and therefore a suitable definition for |ψq> is

q> = ∫-∞q |q'> dq' .

If we introduce the "step function"

Θ(x) ≡
1 , x > 0
0 , x < 0 ,

then we can write

q> = ∫ Θ(q - q') |q'> dq' .

As you can see, the "representation" of |ψq_o> in "q-space" is just
Θ(qo - q).
______________
I can put this into the previous formula
[tex] \psi_q \prime = \int | q > < q | \int | q \prime > dq \prime \prime > dq = [/tex]
[tex] \int \int | q > < q | q \prime > dq dq \prime \prime [/tex]
If
[tex] < q | q \prime > = \delta ( q - q \prime ) [/tex]
then this does make back to
[tex] \psi_q \prime = \int | q \prime > dq \prime \prime [/tex]
And it must "take you BACK", because ∫ |q><q| dq = 1.
______________
This is basically my self-verification on the relationship between
[tex] <q|q \prime> [/tex]
and
[tex] < q \prime | \psi_q > [/tex].

Because in my mind,
[tex] < \psi_q | \psi_q > =1 [/tex]
so
[tex] < q | q > = < d \psi_q \div dq \prime | d \psi_q \div dq \prime > =1 [/tex]
, but apparently it does not come out so.
Indeed, <qo|qo> = δ(0) = ∞. ... What about <ψq_oq_o> ?

Well, the "q-space" representation of |ψq_o> is just Θ(qo - q). So,

q_oq_o> = ∫ |Θ(qo - q)|2 dq

= ∫ Θ(qo - q) dq

= ∫-∞q_o dq

= ∞ .

This "infinite norm" is the reason why I originally said that |ψq_o> is in a class of "generalized vector".

However, if, instead, we define an object

q',q> = ∫q'q |q"> dq" , for q' < q ,

then this object would be an ORDINARY vector of the Hilbert space, and moreover, we would also have ∂|ψq',q>/∂q = |q>. But as I have already said:
From the perspective of any calculation I have ever performed in quantum mechanics, the "|q>" notation of Dirac is superior.
... And now I would like to suggest the following as well:

The only additional thing which bringing such considerations into a calculation of any kind can offer is a headache!

On the other hand, I do appreciate that 'playing around' with these objects can offer some measure of clarification of what is going on, and moreover, that this is what you are in fact accomplishing through such exercises.
______________
II) ... For example, if we use the "momentum" representation to represent this "position" eigenvector, you can always multiply a phase facor such as [tex] e^{iwt} [/tex] to it.
No. You didn't write what you meant. The correct statement is:

The most general relationship between a pair of families of eigenkets |q> of Q and |p> of P is

<q|p> = ei[θ(p) - Φ(q)] 1/sqrt{2π} eipq/h_bar .
______________

P.S. On Sunday, regardless, of whether or not I am able to respond to any of your other posts, I will post something on the next "Postulate".
 
Last edited:
  • #108
Eye,

I roughly got you, still reading it, glad that you clarify many points here .

I definitely agree that you can just bypass #98-102.

About my [tex] \psi_q [/tex] and |q>, I tried some more in clarifing what's going on:

I). Using an self-adjoint operator as example, starting from a discrete case to a continuous case:
[tex] A = \sum a_n | \psi_n > < \psi_n | [/tex]

Let's take
[tex] \triangle_n A = a_n | \psi_n > < \psi_n | [/tex]

[tex] \triangle_n a = a_n - a _{n-1} [/tex]

Now A can be written:
[tex] A = \sum_n ( \triangle_n A / \triangle_n a ) \triangle_n a = [/tex]
[tex] \sum_n ( a_n | \psi_n > < \psi_n | / \triangle_n a ) \triangle_n a [/tex]

Converting it to continuous spectrum, that means
[tex] da = d_n a = \triangle_n a \rightarrow 0 [/tex]
for all n, and of course [tex] n \rightarrow \infty [/tex]
so
[tex] A = \int d ( a | \psi_a > < \psi_a | / da ) da [/tex]

Compare that to
[tex] A = \int a | a > < a | da [/tex]

So, we know
[tex] d ( | \psi_a > < \psi_a | ) / da = | a> < a | [/tex]

II) Look from a perspective of any nomal vector or ket, let
[tex] | \psi > = \sum_n | \psi_n > < \psi_n | \psi > [/tex]

Let's take
[tex] \triangle_n \psi = | \psi_n > < \psi_n | \psi > [/tex]
and again
[tex] \triangle_n a = a_n - a _{n-1} [/tex] .

Now,
[tex] | \psi > = \sum_n ( \triangle_n \psi / \triangle_n a ) \triangle_n a = [/tex]
[tex] \sum_n ( | \psi_n > < \psi_n | \psi > / \triangle_n a ) \triangle_n a [/tex]

Converting it to continuous spectrum, that means
[tex] da = d_n a = \triangle_n a \rightarrow 0 [/tex]
for all n, and of course [tex] n \rightarrow \infty [/tex]
so
[tex] | \psi > = \int ( d ( | \psi_a > < \psi_a | \psi > ) / da ) da [/tex]

Compare that to
[tex] | \psi > = \int | a > < a | \psi > da [/tex]

Again we see
[tex] d ( | \psi_a > < \psi_a | ) / da = | a> < a | [/tex]

III) If I want to make this formula in general:

[tex] | \psi > = \int | a > \psi(a) da = [/tex]
[tex] \int | a > < a| \psi > da = [/tex]

so, I expect to write
[tex] | \psi_q > = \int | a > \psi_q(a) da = [/tex]
[tex] \int | a > \delta ( a - q) da = [/tex]
[tex] \int | a > < a| q > da = [/tex]

and also
[tex] | \psi_q > = \int | a > < a| \psi_q > da [/tex]

while
[tex] | q > = \int | a > < a| q > da = [/tex]
[tex] \int | a > < a| \psi_q > da = | \psi_q > [/tex]

.

Take the conclusion of I) , II) and III), I found all I need is set
[tex] | \psi_q > = \ q> [/tex]
and
[tex] d ( | \psi_q > < \psi_q | ) / dq = | q> < q | [/tex]
.

Replacing [tex] | \psi_q > [/tex] with |q> into the second EQ. I got

[tex] d ( | \psi_q > < \psi_q | ) / dq = d ( | q > < \psi_q | ) / dq= | q> < q | [/tex]

All I need to do is
[tex] < q | = d < \psi_q | /dq [/tex]

This can also well explain
[tex] < q| q > = \infty [/tex]
and
[tex] | q > < q | [/tex] is not a projector.


Thanks
 
Last edited:
  • #109
Eye,

My previous post regarding the object [tex] \psi_q [/tex] seems to make sense, but I just got into troube when trying to verify that with the inner products or norms.

So, I guess it's not working any way.

You can just disregard that and just move on.

To me, I really appreciate what you showed me. At least I have some ideas why this is not working, and what was brought into make it work.

Thanks
 
  • #110
Response to posts #98-102

Post #98

Overall, this section is handled very well.

What you say about the operator L which you construe as a "change of basis" is correct: L shall be unitary.

Concerning the Gram-Schmidt orthonormaliztion procedure which you outline in I.2), as you point out, there is a need for a countable set which "spans" the entire Hilbert space. You are quite right in identifying "separability" as the characteristic which ensures the existence of such a set. The formal definition is as follows:

Let H be a vector space with an "inner product" ( , ) , "complete" in the "induced norm" ║ ║ ≡ √( , ) . Then, H is said to be separable iff:

H has a countable dense subset.
___

This property is then equivalent to:

H has a countable orthonormal basis.
______________

Post #99

Again, overall, this section is handled quite well.

... And, yes, the operator L which transforms from one continuous "generalized" basis |q> to another one |p> given by

[1] |p> = L|q>

shall be unitary. Thus, we also have

[2] |q> = L|p> .

If we write [1] and [2] in terms of "kernels", these two relations become:

[1'] |p> = ∫ L(p,q) |q> dq ,

[2'] |q> = ∫ M(q,p) |p> dp ,

where M(q,p) = L(p,q)*. I am pointing this out, because there is an ambiguity with your notation in the "kernel" where you have "switched" the order of p and q:
[tex] | q > = \overline{L^T} | p > = \int \overline{L(q,p)} | p > dp [/tex]
Regarding what you say next:
I.2) From an inner product to a basis:

The process of this part is almost exactly the same as the discrete one.
I need to figure out how separability contribute to ensure its countability.
The families |q> and |p> are not countable. By assumption these are "generalized vectors" whose parameters q and p vary continuously over R such that

<q|q'> = δ(q - q') , ∫ |q><q| dq = 1 , and similarly for |p> .
______________

Post #100
When dealing with [tex] ( \psi , q) = < q | \psi > [/tex], there shall be an extra care because |q> could be representing two different things here.

Inside the integral
[tex] \int \psi(q) | q > dq [/tex]
, it's a " vector density".

And we aslo use it to denote the eigenvector or eigenket of "position", in this case it's a normal vector not a "vector density".
No! The object "|q>" is the same in both cases! This appears to be an essential point of confusion in some of your other posts. To repeat what I said in post #107, there are three important properties to note about the |q>-family (and others like it):

Q|q> = q|q> ... "eigenkets" ,

<q|q'> = δ(q - q') ... "generalized" (ortho-'normal') ,

∫ |q><q| dq = 1 ... "complete" .

Also:

All "generalized eigenkets" are "vector densities".
______________

Note that in your posts #98 and #99, you have missed my point concerning L. The operator L which I originally defined in post #96 was an arbitrary linear operator:
Now, let L be a linear operator. Let L act on one of the basis vectors bj; the result is another vector in the space which itself is a linear combination of the bi's. That is, for each bj, we have

[1] Lbj = Lijbi .
This is not (in general) a transformation from "one basis to another". L does not even have to have an inverse. Furthermore, notice that the summation is on the first index in Lij – this is not a typo! The summation needs to be defined that way in order for Lij to obey the "component transformation rule":

(Lv)i = Lijvj .

That this relation results from [1] above was shown explicitly in post #96.

... So, to answer the question I asked there, the 'connection' is:

For any orthonormal basis |bi> the "components" relative to this basis are given by

<bi|v> ... vector ,

<bi|L|bj> ... operator .

Note that with Dirac notation, all of this is, in a certain sense, 'trivialized' by the relation (I am now writing the summation explicitly)

j |bj><bj| = 1 (the identity on H) ,

because we can merely "insert" this relation into the appropriate spot so that

<bi|L|v> = <bi|L (∑j |bj><bj|) |v>

= ∑j <bi|L|bj> <bj||v> .
_____

Similarly, in the continuous case, we have, for the "components" relative to any "family" |a>,

<a|v> ... vector ,

<a|L|a'> ... operator ,

and this is "verified" by

<a|L|v> = <a|L ( ∫ |a'><a'| da' ) |v>

= ∫ <a|L|a'> <a'|v> da' .

_____

In your posts #101 and #102, you came close to this idea (... except your operators there were special).
______________
 
  • #111
Another Postulate

Recall the previous postulates:

P0: To a quantum system S there corresponds an associated Hilbert space HS.

P1: A pure state of S is represented a ray (i.e. a one-dimensional subspace) of HS.

P2: To a physical quantity A measurable on (the quantum system) S, there corresponds a self-adjoint linear operator A acting in HS. Such an operator is said to be an "observable".
____

And now here is the next postulate:

P3: The only possible result of a measurement of a physical quantity A is one of the eigenvalues a of the corresponding observable A. In the following, let the quantum system be in a pure state represented by |ψ>:

(i) If the eigenvalue a belongs to a discrete part of the spectrum of A with corresponding eigenprojector Pa , then the probability of obtaining the result a is given by

P(a) = <ψ|Pa|ψ> .

(ii) If the eigenvalue a belongs to a continuous part of the spectrum of A with corresponding generalized eigenket |a>, then the probability of obtaining a result in the infinitesimal interval (a, a + da) is given by

p(a) da = |<a|ψ>|2 da .
___________

Notes:

N.3.1) It is possible for an observable A to have a "mixed" spectrum – i.e. parts which are discrete as well as parts which are continuous. The discrete part of the spectrum is referred to as "the point spectrum of A", denoted Sp(A). The continuous part is referred to as "the continuous spectrum of A", denoted Sc(A). The (overall) spectrum of A is given by
S(A) = Sp(A) U Sc(A).

N.3.2) In (ii) of P3, an implicit assumption is being made, namely, that the continuous part of the spectrum of A is nondegenerate. In the examples which we encounter in quantum mechanics, whenever such an assumption does not hold, we will find that the Hilbert space in question admits a decomposition into a "tensor product" of Hilbert spaces, and that the assumption will then hold with regard to one of the Hilbert spaces in that decomposition.

N.3.3) Let A be an observable whose spectrum may have a continuous (but nondegenerate) part. For a Є Sp(A), let Pa be the corresponding eigenprojector. For a Є Sc(A), let |a> be the corresponding generalized eigenket. Then,

A = ∑a Є S_p(A) a Pa + ∫S_c(A) a |a><a| da ,

and

a Є S_p(A) Pa + ∫S_c(A) |a><a| da = 1 ,

where 1 is the identity on HS . This last relation is called the "closure relation" – it expresses the idea that A has a complete set of eigenkets ("generalized" or otherwise). In case the set Sp(A) or Sc(A) is empty, then the associated sum or integral is understood to be zero.
___________

Exercises:

E.3.1) Regarding N.3.1, give one or two specific examples.

In the following exercises, use Dirac notation.

E.3.2) Let A be an observable whose spectrum may have a continuous (but nondegenerate) part, and let the quantum system be in a pure state |ψ>. Use the postulate P3 and the "closure relation" of N.3.3 to verify that

a Є S_p(A) P(a) + ∫S_c(A) p(a)da = 1 .

E.3.3) Let an Є Sp(A) be a nondegenerate eigenvalue with corresponding eigenket |an>. Let the quantum system be in the pure state |ψ>.

(a) Verify that P(an) = |<an|ψ>|2.

(b) Define ρ = |ψ><ψ|. Verify that P(an) = <an|ρ|an>.

E.3.4) Let an Є Sp(A) be an eigenvalue (possibly degenerate) and let |ank>, k = 1, ... , g(n) , be an orthonormal basis of the eigensubspace corresponding to an. Let the quantum system be in the pure state |ψ>.

(a) Verify that P(an) = ∑k=1g(n) |<ank|ψ>|2.

(b) Define ρ = |ψ><ψ|. Verify that P(an) = ∑k=1g(n) <ank|ρ|ank>.

E.3.5) Verify the analogous expression for (b) of E.3.3, in the case of a (nondegenerate) eigenvalue a Є Sc(A).

E.3.6) Let A be an observable whose spectrum may have a continuous (but nondegenerate) part, and let the quantum system be in a pure state |ψ>. Define "the expectation value of A", denoted <A>, by

<A> ≡ ∑a Є S_p(A) a P(a) + ∫S_c(A) a p(a) da ,

where P(a) is the probability of obtaining a Є Sp(A), and p(a) is the probability density of obtaining a Є Sc(A).

(a) Explain why this definition is correct.

(b) From the definition of <A> and the postulate P3, show that <A> = <ψ|A|ψ>.

(c) Define ρ = |ψ><ψ|. In the notation of E.3.4 for the discrete part, verify that

<A> = ∑nk=1g(n) <ank|ρA|ank> + ∫S_c(A) <a|ρA|a>da .
 
Last edited:
  • #112
E 3.1)

The one I can think of that mixes discrete and continuous spetrum seems to be the simple one in that the discrete eigenvalues are basically the degenerate case of [tex] [ q , q + \triangle q ] [/tex].

By basically taking a function that maps the underlining continuous interval to discrete values, such as
[tex] f ( x ) = 2 \ when \ x \in ( 1 , 3 ) [/tex]
[tex] f ( x ) = 6 \ when \ x \in ( 5 , 7 ) [/tex]
[tex] f ( x ) = x \ when \ x \ not \ \in ( 1 , 3 ) \cup ( 5 , 7 ) [/tex]
, this function f of the continuous observables Q will have a mixed set of continuous spectrum and discrete spectrum.

[tex] S_{c(A)} = R - \ ( 1, 3) - \ ( 5 , 7 ) [/tex]

[tex] P_2 = \int_1^3 | q > < q | dq [/tex]

[tex] P_6 = \int_5^7 | q > < q | dq [/tex]

[tex] A = f(Q) = \int_{S_{c(A)}} q | q > < q | dq + 2 P_2 + 6 P_6 [/tex]

Is there any other more interesting ones?
 
Last edited:
  • #113
Good. Your response shows that you have understood the concept.

Here are two other examples:

(i) the Hamiltonian for a particle moving in one dimension subject to a finite square-well potential;

(ii) the Hamiltonian for a particle subject to a Coulomb potential.

In each of these cases, the eigenstates of the Hamiltonian will involve bound and unbound states of the particle. The bound states are "quantized" – i.e. discrete – whereas, the unbound states are continuous. This is true in general.
____

By the way, here is another basic exercise which I forgot to include:

E.3.1) (b) Regarding N.3.2, give one or two specific examples.

(But ... if you haven't learned "tensor products" yet, this question will have to wait.)
 
  • #114
I will do E 3.3) first.

(a)
Using
[tex] P_{a_{n}} | \psi > = | a_n > < a_n | \psi > [/tex]
for nondegerate eigenvalue a_n because the projector will map | \psi > to its subcomponent of | a_n > , we get
[tex] P ( a_n ) = < \psi | P_{a_{n}} | \psi > = [/tex]
[tex] < \psi ( | a_n > < a_n | \psi > ) = [/tex]
[tex] < \psi | a_n > < a_n | \psi > = [/tex]
[tex] | < a_n | \psi > |^2 [/tex]
, knowing that
[tex] < \psi | a_n > = \overline { <a_n | \psi > } [/tex]
.

(b)
Using
[tex] \rho | a_n > = | \psi > < \psi | a_n > [/tex]
, we get
[tex] < a_n | \rho | a_n > = [/tex]
[tex] <a_n ( | \psi > <\psi | a_n > ) = [/tex]
[tex] < a_n | \psi > < \psi | a_n > = [/tex]
[tex] | < a_n | \psi > |^2 = P ( a_n ) [/tex]
 
  • #115
E 3.4) is very similar to E 3.3).

(a)
Using
[tex] P_{a_{n}} | \psi > = \sum_{k=1}^{g(n)} | a_n^k > < a_n^k | \psi > [/tex]
, we get
[tex] P ( a_n ) = < \psi | P_{a_{n}} | \psi > = [/tex]
[tex] < \psi | ( \sum_{k=1}^{g(n)} | a_n^k > < a_n^k | \psi > ) = [/tex]
[tex] \sum_{k=1}^{g(n)} < \psi | a_n^k > < a_n^k | \psi > = [/tex]
[tex] \sum_{k=1}^{g(n)} | < a_n^k | \psi > |^2 [/tex]
, knowing that
[tex] < \psi | a_n^k > = \overline { <a_n^k | \psi > } [/tex]
.

(b)
Using
[tex] \rho | a_n^k > = | \psi > < \psi | a_n^k > [/tex]
, we get
[tex] \sum_{k=1}^{g(n)} < a_n^k | \rho | a_n^k > = [/tex]
[tex] \sum_{k=1}^{g(n)} <a_n^k | ( | \psi > <\psi | a_n^k > ) = [/tex]
[tex] \sum_{k=1}^{g(n)} < a_n^k | \psi > < \psi | a_n^k > = [/tex]
[tex] \sum_{k=1}^{g(n)} | < a_n^k | \psi > |^2 = P ( a_n ) [/tex]
 
  • #116
E 3.2)

[tex] \forall \psi \in H \ , [/tex]
[tex] < \psi | \ ( \sum_{S_{P(a)}} P_{a_{n}} + \int_{S_{c(a)}} | a > < a | da ) \ | \psi > = [/tex]
[tex] < \psi | 1 | \psi > = [/tex]
[tex] < \psi | \psi > = 1 [/tex]

[tex] < \psi | \ ( \ \sum_{S_{P(a)}} P_{a_{n}} + \int_{S_{c(a)}} | a > < a |da \ ) \ | \psi > = [/tex]
[tex] < \psi | \ ( \ \sum_{S_{P(a)}} P_{a_{n}} \ ) \ | \psi > + < \psi | \ ( \ \int_{S_{c(a)}} | a > < a |da \ ) \ | \psi > = [/tex]
[tex] \sum_{S_{P(a)}} < \psi | P_{a_{n}} | \psi > + < \psi | \ ( \ \int_{S_{c(a)}} | a > < a | da \ ) \ | \psi > = [/tex]
[tex] \sum_{S_{P(a)}} P( a_n ) + < \psi | ( \ \int_{S_{c(a)}} | a > < a | da \ ) | \psi > = [/tex]

Now all I need to prove is
[tex] < \psi | ( \int_{S_{c(a)}} | a > < a | da ) | \psi > = [/tex]
[tex] \int_{S_{c(a)}} P(a) da [/tex]

[tex] < \psi | ( \ \int_{S_{c(a)}} | a > < a | da \ ) | \psi > = [/tex]
[tex] < \psi | ( \ \int_{S_{c(a)}} | a > < a | \psi > da \ ) > = [/tex]
( By N 3.3 Completeness , it will be translated into below :)
[tex] < (( \ \sum_{S_{P(a)}} P_{a_{n}} | \psi > \ ) + ( \ \int_{S_{c(a)}} | a \prime > < a \prime | \psi > da \prime \ ) )| ( \int_{S_{c(a)}} | a > < a | \psi > da ) > = [/tex]
( By orthogonality of eigenkets, we can reduce it to )
[tex] < ( \ \int_{S_{c(a)}} | a \prime > < a \prime | \psi > da \prime \ ) | ( \int_{S_{c(a)}} | a > < a | \psi > da ) > = [/tex]
[tex] \int_{S_{c(a)}} \overline{< a \prime | \psi >} < ( | a \prime > ) | ( \int_{S_{c(a)}} | a > < a | \psi > da ) > da \prime = [/tex]
[tex] \int_{S_{c(a)}} \overline{< a \prime | \psi >} \int_{S_{c(a)}} < a \prime | a > < a | \psi > da da \prime = [/tex]
[tex] \int_{S_{c(a)}} \overline{< a \prime | \psi >} ( \int_{S_{c(a)}} \delta ( a - a \prime ) < a | \psi > da ) da \prime = [/tex]
[tex] \int_{S_{c(a)}} \overline{< a \prime | \psi >} < a \prime | \psi > da \prime = [/tex]
[tex] \int_{S_{c(a)}} | < a \prime | \psi > |^2 da \prime = [/tex]
[tex] \int_{S_{c(a)}} P(a) da [/tex]
 
  • #117
E 3.5)
By P3,
[tex] P(a) = | < a | \psi > |^2 da =
< a | \psi > \overline{< a | \psi >} da [/tex]
.

[tex] < a | \rho | a > = < a | ( \psi > < \psi | a > ) = [/tex]
[tex] < a | \psi > < \psi | a > [/tex]

We already know what is [tex] < a | \psi > [/tex] , but unclear about what [tex] < \psi | a > [/tex] could be.

By comparing the two EQs above, the only way we can equate them is
[tex] < \psi | a > = \overline{< a | \psi >} da [/tex]
.

In a way this make sense, because a is a "generalized vector" and " vector density".

So , [tex] < \psi | a > [/tex] shall be an infinitesmal amount proportional to da and also it shall be propotional to [tex] \overline{< a | \psi >} [/tex] by the general rule of inner product.

I think maybe I can use similar mechanism of [tex] \triangle a [/tex] to get a better analogous proof for it. Later I will give it a try.
 
  • #118
Just try to summarize what I learn here before continue on E 3.5)

1) Let me start with
[tex] \sum_{S_{P(a)}} P_{a_{n}} + \int_{S_{c(a)}} | a > < a | da = I [/tex]
, so
[tex] \int_{S_{c(a)}} | a > < a | da = I - \sum_{S_{P(a)}} P_{a_{n}} [/tex]
.

2). Take derivative of a to it at the continuous part, then
[tex] \frac{d I}{ da } da = | a > < a | da [/tex]
or in other words,
[tex] \int_{S_{c(a)}} \frac{d I}{ da } da = \int_{S_{c(a)}} | a > < a | da [/tex]

3). Further,
[tex] \frac{d I}{ da } da | \psi > = | a > < a | \psi > da [/tex]

4) Consider
[tex] P ( a ) = [/tex]
[tex] < \psi | \frac{d I}{ da } da | \psi > = [/tex]
[tex] < \psi | \frac{d I}{ da } | \psi > da [/tex]
 
  • #119
E 3.5)

By taking the generlized EQ. you posted at #110,

[tex] < \psi | L | \varphi > = \int < \psi | L | a> < a | \varphi > da [/tex]
or
[tex] < a \prime | L \varphi > = \int < a \prime | L | a> < a | \varphi > da [/tex]
, this is pretty easy.

Taking
[tex] < \psi | I | \varphi > = \int < \psi | a> < a | \varphi > da = [/tex]
[tex] \int \overline{< \psi | a>} < a | \varphi > da [/tex]
, we know
[tex] < \psi | a > = \overline{< a | \psi >} [/tex]
.

Actually, it was my fault, somehow I missed the da in the LHS, the actual EQ. in your post is :
[tex] P(a) da = | < a | \psi > |^2 da [/tex]
.

Verifying this, I take
[tex] D ( a \prime ) = \int_{ - \infty}^{ a \prime } \overline{< a | \psi >} < a | \psi > da [/tex]
, then
[tex] P(a) = dD /da = \overline{< a | \psi >} < a | \psi > [/tex]
, in that P(a) is a "probability density" and then
[tex] P(a) da = | < a | \psi > |^2 da [/tex] can be an "infinitismal probability".

Other than E 3.5), this EQ also shows other facts you have posted:

In order for this to be true,
[tex] < q | \psi > = < q | I | q > = \int < q | q \prime > < q \prime | \psi > dq \prime [/tex]
, we need
[tex] < q | q \prime > = \delta ( q \prime - q ) [/tex]
.

Also,
[tex] < q | q > = < q | I | q > = \int < q | q \prime > < q \prime | q > dq \prime = [/tex]
[tex] \int \delta ( q - q \prime) ( q \prime - q ) dq \prime = \infty [/tex]
.
 
  • #120
E 3.6 )

A). If the probability of an obeservable showing a is 1, of course we will expect it always shows a.

If the observable has multiple possible outcomes, we really can not say which one it will definiitely show, we basically make a math. average of them and saying this is its average value, which is defined as " expectation value".

So, in a discrete case, it's of course:
[tex] \sum a P(a) [/tex]

In a continuous case, we will take the approximate average as
[tex] \sum a \ \ Probability([a - \triangle a, a + \triangle a ]) [/tex]
when [tex] \triangle a \rightarrow 0 [/tex] , we get
[tex] \int a P(a) da [/tex]

If we have both discrete and continuous, we will take the average as
[tex] \sum_{S{p(a)}} a P( a | S_{p(a)} ) P( S_{p(a)} ) + \int_{S{c(a)}} a P( a | S_{c(a)} ) P( S_{c(a)} ) da = [/tex]
[tex] \sum_{S{p(a)}} a P( a ) + \int_{S{c(a)}} a P( a ) da [/tex]

b)
[tex] < \psi | A | \psi > = [/tex]
[tex] < \psi | \ \sum_{S{p(a)}} a P_a + \int_{S{c(a)}} a |a> <a| da \ | \psi > = [/tex]
[tex] < \psi | \ \sum_{S{p(a)}} a P_a \ | \psi > \ + \ < \psi | \int_{S{c(a)}} a |a> <a| da \ | \psi > = [/tex]
[tex] \sum_{S{p(a)}} a < \psi | P_a | \psi > \ + \ \int_{S{c(a)}} a < \psi |a> <a| \psi > da = [/tex]
[tex] \sum_{S{p(a)}} a P(a) + \ \int_{S{c(a)}} a P(a) da = [/tex]
[tex] < A > [/tex]
 
  • #121
[tex] < A > = [/tex]
[tex] \sum_{S{p(a)}} a P(a) \ + \ \int_{S{c(a)}} a P(a) da = [/tex]

( By E 3.4 and E 3.5 , we derive below: )

[tex] \sum_n \sum_{k=1}^{g(n)} a < a_n^k | \rho | a_n^k > \ + \ \int_{S{c(a)}} a < a | \rho | a > da = [/tex]

( Using the definition of eigenkets as,
[tex] A a_n^k = a a_n^k [/tex]
[tex] A | a > = a | a > [/tex]
we can derive below: )

[tex] \sum_n \sum_{k=1}^{g(n)} < a_n^k | \rho A | a_n^k > \ + \ \int_{S{c(a)}} < a | \rho A | a > da [/tex]
 
  • #122
E 3.1) a)

Even though my knowledge of "tensor product" of vector spaces might not be enough, I did give a thought on how to handle this exercise.

1). How do we know the spectrum of Q is degenerate?

Unless, we have another set of eigenbasis and self-adjoint operator, said E, and we found we have problems reconciling them with Q. For example,
[tex] < E | \psi > = \int < E | q > < q | \psi > dq [/tex]
does not hold steady or meet the expectation for a state [tex] | \psi > [/tex] .

2). We will then believe that we need to expand the Hilbert space by adding another one in. Why don't we choose "direct sum" instead of "tensor product"?

If we we choose "direct sum", the basis will be extended as
[tex] v_1, v_2, ... v_n, w_1, ... w_n [/tex]
; we are basically just adding more eigenvalues and eigenvectors in doing so.

We need something like
[tex] < x | \psi > = \int < xy | \psi > dy [/tex]
or
[tex] P (xy) dxdy = | < xy | \psi > |^2 dxdy [/tex]
.

So we need to define a "product" of vector spaces that fit our needs.

I may continue later.

================================================

For E 3.1) a), I have "roughly" read a chapter about the unbounded and bounded solutions of square well potential problem.

I found the difference that led to the discrete bounded solution is basically that the solutions outside of the well are in the forms of [tex] A e^{-p_1 x/h} [/tex] at the LHS and [tex] D e^{p_1 x/h} [/tex] at the RHS of the well, because the condition of E < 0 took out the imaginary part.

It's very interesting.
 
  • #123
E 3.1) B)

3) In looking what we already have,

[tex] | \psi_x > = | x> < x | \psi_x > [/tex]
,
[tex] P(y) dy = < \psi | y> < y | \psi > dy [/tex]
,
[tex] | \psi_y > = | y> <y | \psi_y > [/tex]
,
[tex] P(x) dy = < \psi | x> < x | \psi > dx [/tex]
and our believe that
[tex] | \psi > = | x, y > < x,y | \psi > [/tex]
and
[tex] P(x,y) dx dy = < \psi | x,y> <x,y | \psi > dx dy [/tex]
, our quickiest approach will be:
[tex] < x, y | \psi > = < y | \psi_y > < x | \psi_x > [/tex]
and
[tex] < \psi | x,y > = \overline{< y | \psi_y > < x | \psi_x >} = [/tex]
[tex] < \psi_y | y > < \psi_x | x > [/tex]
.

This turned out it will satisfy both needs in representation of ket and probability.

Not only that, we also see
[tex] P(x,y) dx dy = P(x) p(y) dx dy [/tex]
.

This satisfies the general probability rule:
[tex] P ( x, y ) = P(x)P(y|x) [/tex]
; the issue here seems to be that
P(y) = P(y|x), which means x and y need to be independent to each other.

So, if we do find the eigenvalues of x and y independent to each other, we will try to expand our Hilbert space in the above way.

Our current ket space of | x,y,z,s > is of course an example.

4). If x and y are not independent, what will we get?

For example, I can easily produce a degenerate continuous spetrum by using function f(x) = |x|.
 
  • #124
E 3.1) B).

5). In 3), we actually have used the two property of "tensor product":

[tex] ( | x_1 > + | x_2 > ) \otimes | y > = | x_1 , y > + | x_2 , y > [/tex]
[tex] | x > \otimes ( | y_1 > + | y_2> ) = | x , y_1 > + | x , y_2 > [/tex]

The last property
[tex] \alpha | x, y > = | \alpha x> \otimes | y> = | x> \otimes | \alpha y > [/tex]
will be used in
[tex] < A > = < x,y | A | x,y > [/tex]
.
 
  • #125
Checked with a book, I found my answer to E3.1) B) will basically lead to "looking for the set of maximum commutable observables".

That is not what Eye asked any way. Eye's question already assumed there is a Hilbert space and we found multiple "generalized kets" for one eigenvalues of a continuous spectrum.

Any way, to look for that answer, I found I need to clarify there are only true discrete eigenkets in our assumption.

So I went sideway to find a proof why "separability" ensures "countable discrete eigenkets".

The proof is actually quite straightforward, after a few day's rumination:
1). A Hilbert space is a vector space with an inner product.
2). An inner product can define the "norm". The "norm" can define the distance between two vectors.
3). With the "distance", we can define the open sets and so the topology.
4). Definition of "separability" says if we have an open covering for the Hilbet space, then we have a "counatble" open covering as its subset.
5). For any two vectors belong to an orthonormal basis, their distance will be [tex] 2^{1/2} [/tex] .
6). If we define an open coverings as all open balls with a radius 0.25, we will have a countable subset that covers the Hilbert space.
7). Because any two orthonormal vectors has a distance greater than 0.25, so no two orthonormal vectors could be in one open ball; this implys the number of elements in the countable open coverings is greater than the number of all orthonormal vectors, so the basis is definitely countable.

Does this look good?
==================================================

Any way, if we assume the Hilbert space only has countable orthonormal basis, then we know there is no true eigenvector for a continuous spectrum because any continuous real interval has uncountable numbers in it.

Of course, I think there must be a proof that shows the space of square-integrable functions has a countable orthonormal basis and is separable. So, if the "space" has an isomorphich structure with the "functoion" space, it has only countable orthonormal basises and is separable.
 
  • #126
Response to posts #108, 109

The limiting procedure which you allude to in post #108 is not at all "well-defined". You have given no "structure" which tells how A changes with each incremental step in the supposed "limiting procedure".
____

P.S. I had to change this post and cut out all of the quotes because the LaTeX was doing really strange things!

P.P.S. I also wanted to respond to your posts #114-124, but LaTex is malfunctioning. I may not have another opportunity to post more until the beginning of next week.
 
Last edited:
  • #127
Response to posts #108, 109 (again!)

This is what I originally wanted to post.

Here is what you said in post #109 regarding post #108:
My previous post regarding the object [tex] \psi_q [/tex] seems to make sense, but I just got into troube when trying to verify that with the inner products or norms.

So, I guess it's not working any way.

You can just disregard that and just move on. ...
To what you have said here, I will only add that the limiting procedure which you allude to in post #108 is not at all "well-defined". For example, when you say:
Let's take
[tex] \triangle_n A = a_n | \psi_n > < \psi_n | [/tex]

[tex] \triangle_n a = a_n - a _{n-1} [/tex]
... you have given no "structure" which tells how A changes with each incremental step in the supposed "limiting procedure".
 
  • #128
Response to posts #114-124

Posts #114, 115

Your answers to E.3.3 and E.3.4 look fine.
______________

Post #116

In your answer to E.3.2, you began with:
[tex] \forall \psi \in H \ , [/tex] ...
I guess at that moment you forgot that H includes all of the vectors; i.e. even those with norm different from 1.

Later on you reach:
Now all I need to prove is
[tex] < \psi | ( \int_{S_{c(a)}} | a > < a | da ) | \psi > = [/tex]
[tex] \int_{S_{c(a)}} P(a) da [/tex]
The next step should have been to 'push' both the bra "<ψ|" and the ket "|ψ>" through and underneath the integral to get

S_c(A) <ψ|a><a|ψ> da .

For some reason, you were inclined to do this only with regard to the ket (with a slight 'abuse' of notation):
[tex] < \psi | ( \ \int_{S_{c(a)}} | a > < a | da \ ) | \psi > = [/tex]
[tex] < \psi | ( \ \int_{S_{c(a)}} | a > < a | \psi > da \ ) > = [/tex]
But you got around this by invoking an "inner product" (again, with a slight 'abuse' of notation), and then you convinced yourself (quite correctly) that <ψ|a> = <a|ψ>* ... which is what you needed to take the final step.

Let's put an end to this 'abuse' of Dirac notation. Here's what you wrote:
[tex] < (( \ \sum_{S_{P(a)}} P_{a_{n}} | \psi > \ ) + ( \ \int_{S_{c(a)}} | a \prime > < a \prime | \psi > da \prime \ ) )| ( \int_{S_{c(a)}} | a > < a | \psi > da ) > = [/tex]
What you had there was the ket

a Є S_p(A) Pa|ψ> + ∫S_c(A) |a><a|ψ> da ,

which you needed to 'turn around' into a bra. That's easy to do: kets go to bras, bras go to kets, numbers go to their complex conjugates, and operators go to their adjoints. In this case, we get

a Є S_p(A) <ψ|Pa + ∫S_c(A) <ψ|a><a| da .

... And that's all there is to it. Now, you just need "slam" this expression on it's right side with the expression

S_c(A) |a'><a'|ψ> da'

and you will get the desired result. This is how to 'use' Dirac notation without 'abuse'.

______________

Post #117

This answer for E.3.5 is fine ... except for the "dangling" da:
[tex] < \psi | a > = \overline{< a | \psi >} da [/tex]
But, later on, in post #119 you correct yourself.
______________

Posts #118, 119

In post #118 you say:
1) Let me start with
[tex] \sum_{S_{P(a)}} P_{a_{n}} + \int_{S_{c(a)}} | a > < a | da = I [/tex]
, so
[tex] \int_{S_{c(a)}} | a > < a | da = I - \sum_{S_{P(a)}} P_{a_{n}} [/tex]
.

2). Take derivative of a to it at the continuous part, then
[tex] \frac{d I}{ da } da = | a > < a | da [/tex]
or in other words,
[tex] \int_{S_{c(a)}} \frac{d I}{ da } da = \int_{S_{c(a)}} | a > < a | da [/tex]
The parameter "a" is the variable of integration. It is not "free" to take a derivative with respect to it.

What you do in post #119, along similar lines, however, is correct:
[tex] D ( a \prime ) = \int_{ - \infty}^{ a \prime } \overline{< a | \psi >} < a | \psi > da [/tex]
, then
[tex] P(a) = dD /da = \overline{< a | \psi >} < a | \psi > [/tex]
Indeed, you have a "free" parameter here.

Everything else looks fine (modulo a couple of minor typos).
______________

Posts #120, 121

These look fine.
______________

Post #122
E 3.1) a)

Even though my knowledge of "tensor product" of vector spaces might not be enough, I did give a thought on how to handle this exercise.

1). How do we know the spectrum of Q is degenerate?

Unless, we have another set of eigenbasis and self-adjoint operator, said E, and we found we have problems reconciling them with Q. For example,
[tex] < E | \psi > = \int < E | q > < q | \psi > dq [/tex]
does not hold steady or meet the expectation for a state [tex] | \psi > [/tex] .
I don't understand what you meant in the above.

Next:
2). We will then believe that we need to expand the Hilbert space by adding another one in. Why don't we choose "direct sum" instead of "tensor product"?

If we we choose "direct sum", the basis will be extended as
[tex] v_1, v_2, ... v_n, w_1, ... w_n [/tex]
; we are basically just adding more eigenvalues and eigenvectors in doing so.
... Right. So, that's not what we want.
We need something like
[tex] < x | \psi > = \int < xy | \psi > dy [/tex]
or
[tex] P (xy) dxdy = | < xy | \psi > |^2 dxdy [/tex]
.

So we need to define a "product" of vector spaces that fit our needs.
Yes, this is the idea.
______________

Post #123

Looks good (... I do see one small typo, though).

However, about:
, our quickiest approach will be:
[tex] < x, y | \psi > = < y | \psi_y > < x | \psi_x > [/tex]
This is not true in general; i.e. it is true only when the state is such that x and y are "independent".
___
Our current ket space of | x,y,z,s > is of course an example.
... where "s", I assume, refers to spin. This is precisely the example I had in mind (except that I split it up into two examples: |x,y,z> and |x,s>).
___
4). If x and y are not independent, what will we get?
We will get "correlations".

Next:
For example, I can easily produce a degenerate continuous spetrum by using function f(x) = |x|.
Yes, |Q| has a continuous, doubly degenerate spectrum.
______________

Post #124

E 3.1) B).

5). In 3), we actually have used the two property of "tensor product":

[tex] ( | x_1 > + | x_2 > ) \otimes | y > = | x_1 , y > + | x_2 , y > [/tex]
[tex] | x > \otimes ( | y_1 > + | y_2> ) = | x , y_1 > + | x , y_2 > [/tex]

The last property
[tex] \alpha | x, y > = | \alpha x> \otimes | y> = | x> \otimes | \alpha y > [/tex]
will be used in
[tex] < A > = < x,y | A | x,y > [/tex]
.
Yes ... except, the last relation should read

<A> = ∫∫ <x,y|A|x,y> dx dy .
____________________

P.S. I may not be able to post again for another 3 weeks.
 

Similar threads

Replies
19
Views
4K
Replies
5
Views
2K
Replies
13
Views
2K
Back
Top