- #36
bbbeard
- 192
- 4
IWantToLearn said:Hi,
I am wondering why the scalar product of two vectors the way it is, i mean why it wasn't ABtanθ instead of ABcosθ, or XAB, while X is any constant value, why mathematicians define it that way, the same story goes on for the vector product,
why we even had two types of products defined, while we don't have division defined,
and why it was "accidentally!" suitable as the best language to describe nature,
or i got the hole thing wrong and i have to think about it as it was defined as a new mathematics to best describe nature, in this case all of my questions will be answered as it had to be that way because we describe nature already and nature behave that way
my original thought that vector algebra is a topic of pure mathematics and it was developed not in mind describing nature, and after that physicists use it to describe nature
i need your opinion...
Well, here's my 2 cents. Maybe this is too technical, but the sooner you learn this, the better.
From the standpoint of physics, what makes something a vector is how it behaves under some set of coordinate transformations. In ordinary 3d space (as opposed to 4d Minkowski space, for example), the coordinate transformations we're usually interested in are rotations, which you are invited to think of in terms of rotation matrices. A vector is a "rank 1 tensor", i.e. a one-index object, say [itex]V_i[/itex] where i=1,2,3 represent x,y,z. A rotation matrix is a two-index object, say [itex]R_{ij}[/itex]. A rotation matrix acts on a vector, producing another vector
[itex]V'_i=\sum\limits_{j=1}^3 R_{ij}V_j[/itex]
where the sum is over j=1,2,3. We have a convention, called the Einstein summation convention, which says if we write a product like [itex]R_{ij}V_j[/itex] with a repeated index j, we are supposed to assume a summation over the repeated index, which allows us to write the expression above
[itex]V'_i=R_{ij}V_j[/itex]
(That Einstein guy saved the world a lot of ink.) Anyway, we say any three quantities that transform under a rotation [itex]R_{ij}[/itex] the way that the coordinates do is a "vector". Things like forces and momenta are vectors, but not every set of three quantities is a vector. For example, if we think about something like
X=(1/x,1/y,1/z),
this X is not a vector. That is, [itex]1/x_i[/itex] is not a vector -- it doesn't obey the vector transformation law [itex]V'_i=R_{ij}V_j[/itex].
We can also have two-index objects that obey a vector-like, or "tensor" transformation law. But we have to "rotate" both indices, that is, in order for [itex]T_{ij}[/itex] to be a tensor, it has to obey a law like
[itex]T'_{ij}=R_{ik}R_{jm}T_{km}[/itex].
And we can continue making higher-order tensors, but each index gets its own copy of the rotation matrix. An important rank three tensor is the fully-antisymmetric tensor, the so-called "epsilon" tensor (also called the Levi-Civita symbol), [itex]\epsilon_{ijk}[/itex]. You can read more at the link, but the basic idea is that [itex]\epsilon_{ijk}[/itex] switches signs when two indices are swapped, so that [itex]\epsilon_{123}=-\epsilon_{213}[/itex]. By convention [itex]\epsilon_{123}=+1[/itex]. Necessarily [itex]\epsilon_{ijk}=0[/itex] if any two of the indices are equal, e.g. [itex]\epsilon_{112}=0[/itex].
Now, here's the point: the outer product of two tensors is a tensor. So if we have two vectors [itex]V_i[/itex] and [itex]W_j[/itex], the object
[itex]T_{ij}=V_i W_j[/itex].
is a tensor. If we sum any two indices of a tensor, the result is a lower-rank tensor, e.g. if [itex]T_{ij}[/itex] is a rank two tensor, then [itex]T_{ii}[/itex] is a tensor of rank zero, also called a scalar (this is actually the trace of T if we think of T as a matrix). A scalar has no indices left over, so it doesn't change under rotations. In particular, if [itex]T_{ij}=V_i W_j[/itex] as above, then
[itex]T_{ii}=V_i W_i=V \bullet W[/itex]
is the inner product (also called the scalar product or dot product) of V and W. Now, the stuff about the magnitudes and the cosine of the angle is interesting, but the real utility of the inner product is that it takes two vectors and gives us a scalar.
Likewise, it turns out that we can take the epsilon tensor and two vectors V and W and make something that is a vector product:
[itex]U_{i}=\epsilon_{ijk}V_j W_k=V \times W[/itex].
Again, the stuff about the magnitudes and the sine of the angle is interesting, but the real interesting thing is that we have created a vector from two vectors. Actually, the cross-product is what we call a "pseudo-vector" because it changes sign under spatial inversion, but that's an aspect we don't have to explore here.
Now, hopefully, from this point you can see that there are lots of ways to make products from tensors. We could, say, make a rank-four tensor from a single vector using
[itex]T_{ijkm}=V_i V_j V_k V_m[/itex].
but there's no guarantee that this will be useful! There are rules for combining epsilons with different indices, and there's this thing called the Kronecker delta which bears a striking resemblance to the identity matrix, and it turns out that the del operator is also a rank one tensor, but that discussion is for another day...
BBB