Understanding the Dual Basis and Its Directions

In summary,The expert does not understand how to see the basis of the gradient as easily as she can see the basis of the original vectors.
  • #1
JTC
100
6
Please help.

I do understand the representation of a vector as: vi∂xi
I also understand the representation of a vector as: vidxi

So far, so good.

I do understand that when the basis transforms covariantly, the coordinates transform contravariantly, and v.v., etc.

Then, I study this thing called the gradient.
If I work out an indicial notation, I get this:
∇f=(∂f/∂xi)dxi

Now comes my trouble.

I can "envision" in my mind, that ∂xi are tangent to the coordinate curves and are, essentially, directions.

But I cannot see the "directions" of dxi
I cannot see them as "basis vectors" as easily as I see ∂xi

I do understand that dxi ∂xj = δij
And I understand how the direction of, say in 3D space, dx1 is perpedicular to the plane formed by ∂x2 and ∂x3.

But I cannot easily see the "directions" of the basis of the dual vectors as easily as I can see the basis of the original vectors (as tangent to the coordinate curves).

I cannot make the leap and replace dxi by e1, e2, e3, and easily as I can replace ∂xi with e1, e2, e3

I can begin with the definition of how I construct a dual basis... But I cannot easily make the leap to see the basis as this form dxi. I just don't see "directions" here.

Can someone provide some insight?

Also, given a metric, I can convert the basis of the gradient to a covariant form and the components to a contravariant. So why is the gradient called contravariant, when it can go either way with a metric?
 
Physics news on Phys.org
  • #2
In other words: whenever I see a representation of a gradient, I always see a basis that looks just like the basis of an original vector space (not the dual space)

And I KNOW the components transform contravariantly--but what are the directions of the bases?

I can only distinguish these animals algebraically (dx), and not geometrically (directions).

I am not sure I am asking this correctly.
 
Last edited:
  • #3
I think the simplest way to think about it is as follows. A derivative as a result of differentiation is a linear approximation in some direction. This means we have ##f(x_0+v) = f(x_0) + L_{x_0}(v) + r(v)## with the directional (in direction of ##v##) derivative ##L## and the error ##r##. So you can consider the vector ##L(v)## or the linear function ##L\, : \, v \mapsto L(v)\,.## It is the same coin with two different sides. An easy example would be ##(7+v)^2= 7^2 + (\left.\frac{d}{dx}\right|_{x=7}x^2)\cdot v + v^2 = 7^2 + 2 \cdot x|_{x=7} \cdot v + v^2##. If you ask a high school student about the differential of ##x^2## you'll get the answer ##2x##. But here the ##2## represents the slope of the tangent, the direction if you like, as well as the linear map ##2\, : \, x \mapsto 2\cdot x\,.##

Not sure whether this helps here, but I've tried to shed some lights on it with this insight article:
https://www.physicsforums.com/insights/the-pantheon-of-derivatives-i/

And here's "point of view of a derivative put to the extreme": The derivative
$$
D_{x_0}L_g(v)= \left.\frac{d}{d\,x}\right|_{x=x_0}\,L_g(x).v = J_{x_0}(L_g)(v)=J(L_g)(x_0;v)
$$
can be viewed as
  1. first derivative ##L'_g : x \longmapsto \alpha(x)##
  2. differential ##dL_g = \alpha_x \cdot d x##
  3. linear approximation of ##L_g## by ##L_g(x_0+\varepsilon)=L_g(x_0)+J_{x_0}(L_g)\cdot \varepsilon +O(\varepsilon^2) ##
  4. linear mapping (Jacobi matrix) ##J_{x}(L_g) : v \longmapsto \alpha_{x} \cdot v##
  5. vector (tangent) bundle ##(p,\alpha_{p}\;d x) \in (D\times \mathbb{R},\mathbb{R},\pi)##
  6. ##1-##form (Pfaffian form) ##\omega_{p} : v \longmapsto \langle \alpha_{p} , v \rangle ##
  7. cotangent bundle ##(p,\omega_p) \in (D,T^*D,\pi^*)##
  8. section of ##(D\times \mathbb{R},\mathbb{R},\pi)\, : \,\sigma \in \Gamma(D,TD)=\Gamma(D) : p \longmapsto \alpha_{p}##
  9. If ##f,g : D \mapsto \mathbb{R}## are smooth functions, then $$D_xL_y (f\cdot g) = \alpha_x (f\cdot g)' = \alpha_x (f'\cdot g + f \cdot g') = D_xL_y(f)\cdot g + f \cdot D_xL_y(g)$$ and ##D_xL_y## is a derivation on ##C^\infty(\mathbb{R})##.
  10. ##L_x^*(\alpha_y)=\alpha_{xy}## is the pullback section of ##\sigma: p \longmapsto \alpha_p## by ##L_x##.
And all of these, only to describe ##(x^2)'=2x\,.##
 
  • #4
I am afraid this has made things worse for me. My question is really simple.

I have a vector. I have a coordinate system. A vector is an operator on functions. I can see how the function changes in each coordinate direction and I can interpret this as a basis. I can see the basis.

I now have gradient. And can formulate it. But I cannot see the "directions" of the gradient. I cannot see the dxi as directions.
 
  • #5
JTC said:
I am afraid this has made things worse for me. My question is really simple.

I have a vector. I have a coordinate system. A vector is an operator on functions. I can see how the function changes in each coordinate direction and I can interpret this as a basis. I can see the basis.

I now have gradient. And can formulate it. But I cannot see the "directions" of the gradient. I cannot see the dxi as directions.
A matrix consists of a bunch of vectors. Can you "see" them, if you look at it? And if, what do you see? Vectors ##v^i## or linear forms ##x \mapsto ( v^i \mapsto \langle x,v^i \rangle)##, or do you take the entire thing as one transformation? Maybe it helps to consider ##dx^i## as the projection, that cuts out the ##i-## component of a vector. A vector space ##V## and its dual ##V^*## might be isomorphic, but this doesn't mean one "sees" it.
 
Last edited:
  • #6
JTC said:
I am afraid this has made things worse for me. My question is really simple.

I have a vector. I have a coordinate system. A vector is an operator on functions. I can see how the function changes in each coordinate direction and I can interpret this as a basis. I can see the basis.

I now have gradient. And can formulate it. But I cannot see the "directions" of the gradient. I cannot see the dxi as directions.
The dual vectors are not directions in the same sense as the tangent vectors, they belong to a completely different vector space, namely the vector space of linear maps from the tangent space to scalars.
 
  • #8
Thank you everyone. This has been a great help.

I am grateful for the attempts, but I think Orodruin saw what was troubling me.
 
  • #9
Say I give you a basis of vectors at a point. Forget about differential geometry for a moment, and let's just think about a vector space. So, you have an n-dimensional vector space, and I give you a set of n linearly-independent little arrows at the origin.

The little arrows don't have to be at right angles to each other; they only have to be linearly independent. So, say there is a point A in the vector space, and you want to give its location in terms of the little arrows I've given you. Well, there are two ways you can do this:

1. Interpret the little arrows as directions along which you make displacements, and take the appropriate linear combination of directed displacements to land at the point A.

This is probably the most straightforward thing you can do. But there is another option:

2. Interpret each subset of (n-1) arrows as specifying an oriented hyperplane. The meaning of the n-th coordinate is the amount by which you should displace the hyperplane of the remaining (n-1) arrows, parallel to itself. So, to reach the point A, we make parallel displacements of each of the n such hyperplanes, such that their new point of mutual intersection is the point A.

Now, if the basis I gave you was orthonormal to begin with, then the definitions 1 and 2 are actually the same (this should be easy to imagine in 3d space with three orthogonal unit vectors). But if the basis I gave you was *not* orthonormal, then definitions 1 and 2 describe different procedures (imagine 3d space with non-orthogonal vectors; you should be able to see that parallel displacements of, say, the XY plane do not correspond to displacements along the Z direction, since Z is not perpendicular to the XY plane).

Definition 2 is essentially what the dual basis means. There is the remaining question of the "weight" to give to these parallel displacements of hyperplanes (i.e., how much displacement corresponds to a given value of a coordinate), and this "weight" is determined by the defining formula,

$$\omega_i v^j = \delta_i^j$$

In any case, the basic idea is not hard to visualize and is actually helpful when thinking about differential geometry: Vectors measure displacements along a direction; dual vectors measure displacements from a hypersurface. This is why, e.g., vectors are appropriate for describing velocities, whereas dual vectors are appropriate for describing gradients.
 
  • Like
Likes fresh_42, JTC and Orodruin
  • #10
thank you, Ben
 

FAQ: Understanding the Dual Basis and Its Directions

1. How do I find the dual basis for a given vector space?

To find the dual basis for a vector space, you need to first determine the basis of the vector space. Then, you can use the Gram-Schmidt process to find the orthogonal basis. Finally, you can use the basis of the orthogonal basis to find the dual basis for the vector space.

2. What is the significance of the dual basis in linear algebra?

The dual basis is significant in linear algebra because it allows us to define a linear transformation from the dual space to the original vector space. This transformation helps us to understand and solve problems related to duality and linear transformations.

3. Can a vector space have more than one dual basis?

No, a vector space can have only one dual basis. This is because the dual basis is uniquely determined by the basis of the original vector space. Each vector in the dual basis corresponds to a specific linear functional on the original vector space.

4. How does the concept of dual basis relate to the concept of dual space?

The dual basis is a set of vectors that form a basis for the dual space. The dual space is the set of all linear functionals on a vector space. The dual basis and dual space are related in that the vectors in the dual basis correspond to the linear functionals in the dual space.

5. Can the dual basis be used to find the coordinates of a vector in a vector space?

Yes, the dual basis can be used to find the coordinates of a vector in a vector space. This is because the dual basis provides a way to represent a vector in terms of the linear functionals in the dual space. By using the coefficients from this representation, we can find the coordinates of the vector in the original vector space.

Similar threads

Replies
3
Views
481
Replies
18
Views
625
Replies
21
Views
17K
Replies
8
Views
2K
Replies
24
Views
2K
Replies
9
Views
3K
Back
Top