What is the role of the transpose matrix in tensor transformations?

In summary, the transformation of a tensor can be written in component form as F^{\mu\nu}=\Lambda^{\mu}_{\alpha}\Lambda^{\nu}_{\beta}F^{\beta\alpha}, or in matrix notation as F^{'}=LFL^{T}, where L is the Lorentz transformation matrix. The transpose matrix appears in the equation because of the definition of matrix multiplication, where the sum is always over an index that is a column index for the matrix on the left and a row index for the matrix on the right. This is why F must be the matrix on the left in the equation \Lambda^\mu{}_\alpha F^{\beta\alpha}. The understanding of this concept can
  • #1
peterjaybee
62
0
Hi,

In component form the transformation for the following tensor can be written as
[tex]F^{\mu\nu}=\Lambda^{\mu}_{\alpha}\Lambda^{\nu}_{\beta}F^{\beta\alpha}[/tex]

or in matrix notation, apparently as
[tex]F^{'}=LFL^{T}[/tex]
Here L is the Lorentz transformation matrix

Im happy with the component form, but I don't understand where the transpose matrix bit comes from in the matrix equation, and why it is on the RHS of the F tensor.
 
Mathematics news on Phys.org
  • #2
It follows immediately from the definition of the product of two matrices.

[tex](AB)_{ij}=A_{ik}B_{kj}[/tex]

(What does this definition say is on row i, column j of [tex]LFL^T[/tex]?)
 
  • #3
Im sorry, I still can't see it.
 
  • #4
No need to apologize. I know a lot of people are having difficulties with this. I'm genuinely interested in why that is, so when you do see it, I'd appreciate if you could tell me what it was that confused you.

If we write the component on row [itex]\mu[/itex], column [itex]\nu[/itex], of an arbitrary matrix X as [itex]X_{\mu\nu}[/itex], then

[tex](LFL^T)_{\mu\nu}=(LF)_{\mu\rho}(L^T)_{\rho\nu}=L_{\mu\sigma}F_{\sigma\rho}L_{\nu\rho}=L_{\mu\sigma}L_{\nu\rho}F_{\sigma\rho}[/tex]
 
  • #5
I strugle with this concept (and a lot of other index manipulations) because I find the index notation unfamiliar and a little alien. Because I don't understand it, my logic is flawed. For example when I initially saw
[tex]F^{\mu\nu}=\Lambda^{\mu}_{\alpha}\Lambda^{\nu}_{\beta}F^{\beta\alpha}[/tex],
I thought to get the transformed faraday components you just times two lorentz matricies together then right multiply by the untransformed faraday tensor. Even though I know this is wrong (having tried it) I do not understand why it is wrong. It is very difficult to describe.


Thanks to you, I now understand how to do the manipulation which is a relief :smile:. The manipulation itself makes sense, I just don't understand where my logic in the above fails if you see what I mean.

Ill try expressing it in another way if someone asked me...
"Can you get the transformed faraday components by just multiplying two lorentz matricies together then right multiply by the untransformed faraday tensor?"...
I would say no, but if they then asked me why not, I would be stuck.
 
  • #6
Maybe it will be help a little if we say it in coordinate free language. In spacetime, the distinction between a vector and a covector is only conceptually useful, but computationally, we can convert a vector into a covector or a covector into a vector using the metric tensor g(_,_). So if we have a vector v, then the covector corresponding to that is defined as g(v,_), i.e. the vector vμ corresponds to the covector vα through vα = gαμvμ

Similarly, if we have a contravariant tensor Fαβ, then we use this by contracting it with covectors v and w thus: Fαβvαwβ. But, as we have seen above, vα = gαμvμ and wβ = gβνvν. So Fαβvαwβ = Fαβgαμvμgβνvν = gβνgαμFαβvμvν.

So the contravariant tensor, which acts on pairs of covectors, can be made to act on their corresponding vectors by replacing it with the covariant tensor gβνgαμFαβ = Fμν.
 
  • #7
Just noticed the question was not about raising and lowering indices. Ignore my previous post.
 
  • #8
Look at the definition of matrix multiplication again, in #2. Note that the sum is always over an index that's a column index for the matrix on the left and a row index for the matrix on the right. Since [itex]\Lambda^\nu{}_{\beta}[/itex] is row [itex]\nu[/itex], column [itex]\beta[/itex] of a [itex]\Lambda[/itex], and [itex]F^{\beta\alpha}[/itex] is row [itex]\beta[/itex], column [itex]\alpha[/itex] of a [itex]F[/itex], the result

[tex]\Lambda^\nu{}_\beta F^{\beta\alpha}=(\Lambda F)^{\nu\alpha}[/tex]

follows immediately from the definition of matrix multiplication. But now look at

[tex]\Lambda^\mu{}_\alpha F^{\beta\alpha}[/tex]

Note that the sum is over the column index of F. If you have another look at the definition of matrix multiplication, you'll see that this means that if the above is a component of the product of two matrices, one of which is F, then F must be the matrix on the left. When you understand that, the rest should be easy.

Also note that you should LaTeX [itex]\Lambda^\mu{}_\nu[/itex] as \Lambda^\mu{}_\nu, so that the column index appears diagonally to the right below the row index. And check out the comment about the inverse here to see why the horizontal position of the indices matters.
 
  • #9
I finally get it! Its a miracle.

Thank you for your help.
 

FAQ: What is the role of the transpose matrix in tensor transformations?

What are tensor transformations?

Tensor transformations are mathematical operations that manipulate tensors, which are multidimensional arrays of numbers. These transformations can change the shape, orientation, and magnitude of tensors, allowing for more efficient calculations and analyses in fields such as physics, engineering, and computer science.

What is the difference between a tensor transformation and a tensor product?

A tensor transformation involves manipulating an existing tensor, while a tensor product involves combining two or more tensors to create a new tensor. Tensor transformations are often used to simplify or optimize tensor products, making them easier to perform and interpret.

How are tensor transformations used in machine learning?

In machine learning, tensor transformations are used to preprocess data and extract features, making it easier for algorithms to identify patterns and make predictions. They are also essential in operations such as gradient descent and backpropagation, which are used to optimize the parameters of deep learning models.

Can tensor transformations be applied to non-numerical data?

No, tensor transformations can only be applied to numerical data. This is because tensors are arrays of numbers, and any non-numerical data would need to be converted into numerical form before undergoing a tensor transformation. However, there are techniques such as one-hot encoding that can convert categorical data into numerical data for use in tensor transformations.

Are there any limitations or challenges associated with tensor transformations?

Yes, there are a few limitations and challenges when it comes to tensor transformations. One main challenge is the curse of dimensionality, where the number of dimensions in a tensor can become too large for efficient calculations. Another limitation is the need for specialized software and programming skills to perform tensor transformations, making it inaccessible for some researchers and practitioners.

Similar threads

Replies
4
Views
2K
Replies
3
Views
928
Replies
2
Views
571
Replies
30
Views
6K
Replies
3
Views
2K
Replies
4
Views
589
Replies
89
Views
13K
Back
Top