Valid Method for Proving Matrix Equation with Independent Variables

In summary, after a series of computations, the conversation discusses a matrix equation involving W1 and W2, which are correlated. The problem asks to show that the equation can be rewritten as W = RZ, where Z is a matrix with independent elements. The solution involves solving for Z = R^{-1}SY and showing that the elements of Z are independent.
  • #1
island-boy
99
0
after a series of computations, I was able to get the following matrix equation from the given of a problem:

[tex]\[\left( \begin{array} {ccc} W_1 \\ W_2 \end{array} \right)\] =
\[\left( \begin{array} {ccc} \frac{\sigma_{11}}{\sqrt{\sigma_{11}^2 + \sigma_{12}^2}} & \frac{\sigma_{12}}{\sqrt{\sigma_{11}^2 + \sigma_{12}^2}} \\ \frac{\sigma_{21}}{\sqrt{\sigma_{21}^2 + \sigma_{22}^2}} & \frac{\sigma_{22}}{\sqrt{\sigma_{21}^2 + \sigma_{22}^2}}\end{array} \right)\] \[\left( \begin{array} {ccc} Y_1 \\ Y_2 \end{array} \right)\] [/tex]

where Y1 and Y2 are independent processes.

the correlation of W1 and W2 was given as follows:

[tex]\rho = \frac{\sigma_{11}\sigma_{21} + \sigma_{12}\sigma_{22}}{\sqrt{\sigma_{11}^2 + \sigma_{12}^2}\sqrt{\sigma_{21}^2 + \sigma_{22}^2}}[/tex]

Now what the problem asks is that I be able to show the following matrix equation to be true:

[tex]\[\left( \begin{array} {ccc} W_1 \\ W_2 \end{array} \right)\] =
\[\left( \begin{array} {ccc} 1 & 0 \\ \rho & \sqrt{1 - \rho^2} \end{array} \right)\] \[\left( \begin{array} {ccc} Z_1 \\ Z_2 \end{array} \right)\] [/tex]

where Z1 and Z2 are independent
and rho is the correlation of W1 and W2.

My question is:
can I just let
[tex]\sigma_{11} = 1[/tex]
[tex]\sigma_{12} = 0[/tex]

and just let
Y1 = Z1
Y2 = Z2?

cause if I did so, then the matrix equation that I want to prove is satisfied.
that is, the following are now true:
[tex]\rho = \frac{\sigma_{21}}{\sqrt{\sigma_{21}^2 + \sigma_{22}^2}}[/tex]
[tex]\sqrt{1 - \rho^2} = \frac{\sigma_{22}}{\sqrt{\sigma_{21}^2 + \sigma_{22}^2}}[/tex]

Is this a valid way of proving?

or should I have to find a matrix linear transformation to transfrom the first equation that I got into the required equation? cause if that's the way that it shouldbe done, then I'm not sure how to proceed about it.

thanks for the help.
 
Last edited:
Physics news on Phys.org
  • #2
island-boy said:
after a series of computations, I was able to get the following matrix equation from the given of a problem:

[tex]\[\left( \begin{array} {ccc} W_1 \\ W_2 \end{array} \right)\] =
\[\left( \begin{array} {ccc} \frac{\sigma_{11}}{\sqrt{\sigma_{11}^2 + \sigma_{12}^2}} & \frac{\sigma_{12}}{\sqrt{\sigma_{11}^2 + \sigma_{12}^2}} \\ \frac{\sigma_{21}}{\sqrt{\sigma_{21}^2 + \sigma_{22}^2}} & \frac{\sigma_{22}}{\sqrt{\sigma_{21}^2 + \sigma_{22}^2}}\end{array} \right)\] \[\left( \begin{array} {ccc} Y_1 \\ Y_2 \end{array} \right)\] [/tex]
Let W denote the left side of this equation, let S denote the 2x2 matrix on the right side, and let Y denote what you think it denotes.
[tex]\[\left( \begin{array} {ccc} W_1 \\ W_2 \end{array} \right)\] =
\[\left( \begin{array} {ccc} 1 & 0 \\ \rho & \sqrt{1 - \rho^2} \end{array} \right)\] \[\left( \begin{array} {ccc} Z_1 \\ Z_2 \end{array} \right)\] [/tex]

where Z1 and Z2 are independent
and rho is the correlation of W1 and W2.
Let R denote the 2x2 matrix on the right side of this equation, and let Z denote what you think it does.

So you have W = SY and you want to show that if W = RZ, then the entries of Z are independent. You can't go setting [itex]\sigma _{ij}[/itex] and Zi to whatever you like. What kind of question would that be? Anyways, if you want to get W = RZ, then it's equivalent to get SY = RZ. In order to get this, what must Z be? Clearly, Z = R-1SY. So look at R-1SY, and show it's entires to be independent. Use the fact that the entries of Y are independent, and that the entries of W have correlation [itex]\rho[/itex].
 
  • #3
hi thanks for replying

and you want to show that if W = RZ, then the entries of Z are independent.

actually, what I need to show is that I can write W in the form RZ where Z are independent.
that is, defining W, S, Y, R, and Z below, and given W = SY where Y are independent, I must show that I can write W as W = RZ where Z are independent.

does what you have written still apply?
thanks again.
 
  • #4
Yes. ,
 
  • #5
hi AKG,

after thinking it throrughly, I think I understand what you meant.
I only need to solve for Z = R^{-1}SY
and then I show the elements of Z are independent of each other (possibly by showing that their covariance is 0)

Thanks for you help! cheers!
 

FAQ: Valid Method for Proving Matrix Equation with Independent Variables

What is a linear transformation?

A linear transformation is a mathematical function that maps one vector space to another in a way that preserves the structure of the original space. This means that the transformation must satisfy two properties: linearity and homogeneity.

How is a linear transformation represented?

A linear transformation can be represented using a matrix, which is a rectangular array of numbers. The size of the matrix depends on the dimension of the vector space being transformed. Each column of the matrix represents the transformation of a basis vector in the original space.

What is the difference between a linear transformation and a nonlinear transformation?

A linear transformation follows the rules of linearity and homogeneity, meaning that the transformation of a linear combination of vectors is equal to the same combination of the transformed vectors. A nonlinear transformation does not have this property and can result in distorted or curved transformations.

How can linear transformations be applied in real-world situations?

Linear transformations have many applications in fields such as physics, engineering, and computer graphics. They can be used to represent and manipulate geometric transformations, such as rotations, translations, and scaling. They can also be used in data analysis to transform and analyze data sets.

What is the inverse of a linear transformation?

The inverse of a linear transformation is another linear transformation that "undoes" the original transformation. It maps the transformed vectors back to their original positions in the original vector space. The inverse of a linear transformation can be found by calculating the inverse of the matrix representing the transformation.

Back
Top