The covariance of a sum of two random variables X and Y

In summary, the covariance of the sum of two random variables X and Y is expressed as Cov(X + Y) = Cov(X, X) + Cov(Y, Y) + 2Cov(X, Y). This formula illustrates that the covariance of the sum depends not only on the individual variances of X and Y but also on the covariance between them, indicating how they jointly vary.
  • #1
Ad VanderVen
169
13
TL;DR Summary
Suppose you have three random variables X, Y and K. Suppose X and Y are independent, but each correlated with K. Suppose Z = X+Y. Is it true that in probability theory the covariance of Z with K is equal to the sum of the covariance of X with K and the covariance of Y with K?
Suppose X and Y are random variables. Is it true that

Cov (Z,K) = Cov(X,K)+Cov(Y,K) where Z=X+Y?
 
Last edited:
Physics news on Phys.org
  • #2
The summary:

Suppose you have three random variables X, Y and K. Suppose X and Y are independent, but each correlated with K. Suppose Z = X+Y. Is it true that in probability theory the covariance of Z with K is equal to the sum of the covariance of X with K and the covariance of Y with K?

is incorect and should be:

Suppose you have three random variables X, Y and K. Suppose Z = X+Y. Is it true that in probability theory the covariance of Z with K is equal to the sum of the covariance of X with K and the covariance of Y with K?
 
  • #3
Covariance is linear in each variable. The random variables do not have to be independent. See this.
 
  • #5
I remember something along the lines that correlation( covariance?) was an inner -product in some space of Random Variables. I guess we have Cov<X,X>=Var(X)=norm(X)?
 
  • #6
How can X and Y be independent (cov=0) if they are both correlated with K?
 
  • Like
Likes Filip Larsen
  • #7
BWV said:
How can X and Y be independent (cov=0) if they are both correlated with K?
Suppose X and Y are any two independent variables and K = X+Y. Then cov(X,K) = cov(X, X+Y) = cov(X,X) + cov(X,Y) = cov(X,X) + 0 > 0.
 
  • Like
Likes Filip Larsen and BWV
  • #8
Let X and Y be independent uniformly distributed RVs and let K be some function of them.

Z=X+Y

But k could be anything, if it is X+Y it obviously perfectly correlated, but is could also be X-Y, or sin(x+y) etc, no?
 
  • #9
I believe the best we can use is bilinearity of coefficients, i.e.,
Cov(aX, bY)=abCov(X,Y).
But you're right, beyond that, I think there are no rules for f with f=f(X,Y).

Edit: I suspect the answer here may fall under propagation of errors/uncertainty







Wikipedia
Search

Propagation of uncertainty​


Article Talk

For the propagation of uncertainty through time, see Chaos theory § Sensitivity to initial conditions.
In statistics, propagation of uncertainty (or propagation of error) is the effect of variables' uncertainties (or errors, more specifically random errors) on the uncertainty of a function based on them. When the variables are the values of experimental measurements they have uncertainties due to measurement limitations (e.g., instrument precision) which propagate due to the combination of variables in the function.
The uncertainty u can be expressed in a number of ways. It may be defined by the absolute error Δx. Uncertainties can also be defined by the relative errorx)/x, which is usually written as a percentage. Most commonly, the uncertainty on a quantity is quantified in terms of the standard deviation, σ, which is the positive square root of the variance. The value of a quantity and its error are then expressed as an interval x ± u. However, the most general way of characterizing uncertainty is by specifying its probability distribution. If the probability distribution of the variable is known or can be assumed, in theory it is possible to get any of its statistics. In particular, it is possible to derive confidence limits to describe the region within which the true value of the variable may be found. For example, the 68% confidence limits for a one-dimensional variable belonging to a normal distribution are approximately ± one standard deviation σ from the central value x, which means that the region x ± σ will cover the true value in roughly 68% of cases.
If the uncertainties are correlated then covariance must be taken into account. Correlation can arise from two different sources. First, the measurement errors may be correlated. Second, when the underlying values are correlated across a population, the uncertainties in the group averages will be correlated.[1]
In a general context where a nonlinear function modifies the uncertain parameters (correlated or not), the standard tools to propagate uncertainty, and infer resulting quantity probability distribution/statistics, are sampling techniques from the Monte Carlo method family.[2] For very expansive data or complex functions, the calculation of the error propagation may be very expansive so that a surrogate model[3] or a parallel computing strategy[4][5][6] may be necessary.
In some particular cases, the uncertainty propagation calculation can be done through simplistic algebraic procedures. Some of these scenarios are described below.

Linear combinations​

Non-linear combinations​


Example formulae​

Example calculations​


See also​















References​



















Further reading​







External links​







Last edited 1 month ago by Hellacioussatyr
Wikipedia

 
Last edited:
  • #10
BWV said:
Let X and Y be independent uniformly distributed RVs and let K be some function of them.

Z=X+Y

But k could be anything,
Well, it can be a lot of things. If you start with any two uncorrelated variables, X and Y, many variables, Z, will have some correlation with both of them.
BWV said:
if it is X+Y it obviously perfectly correlated,
Not perfectly. X's correlation with X+Y is not perfect. The independent Y variable term prevents that.
BWV said:
but is could also be X-Y, or sin(x+y) etc, no?
Yes, there are a lot of examples where there is clearly a connection. There are also a lot of examples where two uncorrelated variables, X and Y might be correlated with a third variable, K, with no apparent reason.
 
  • #11
FactChecker said:
Well, it can be a lot of things. If you start with any two uncorrelated variables, X and Y, many variables, Z, will have some correlation with both of them.

Not perfectly. X's correlation with X+Y is not perfect. The independent Y variable term prevents that.

Yes, there are a lot of examples where there is clearly a connection. There are also a lot of examples where two uncorrelated variables, X and Y might be correlated with a third variable, K, with no apparent reason.
Was thinking of the correlation of z and k per the OP, obviously if Z=K the correlation is 1
 

FAQ: The covariance of a sum of two random variables X and Y

What is the covariance of a sum of two random variables X and Y?

The covariance of a sum of two random variables X and Y is given by Cov(X + Y) = Cov(X, X) + Cov(X, Y) + Cov(Y, X) + Cov(Y, Y). Since Cov(X, X) is the variance of X and Cov(Y, Y) is the variance of Y, this can be simplified to Var(X) + 2Cov(X, Y) + Var(Y).

How do you calculate the covariance between two random variables X and Y?

The covariance between two random variables X and Y is calculated using the formula Cov(X, Y) = E[(X - E[X])(Y - E[Y])], where E[X] and E[Y] are the expected values (means) of X and Y, respectively.

What does a positive or negative covariance indicate about the relationship between X and Y?

A positive covariance indicates that X and Y tend to increase or decrease together, meaning they have a positive linear relationship. A negative covariance indicates that as one variable increases, the other tends to decrease, indicating a negative linear relationship.

Is the covariance of X + Y always greater than the covariance of X and Y individually?

Not necessarily. The covariance of X + Y depends on the individual variances of X and Y as well as their covariance. If X and Y are independent, their covariance is zero, and the covariance of X + Y will simply be the sum of their variances. If they are not independent, the combined covariance can be greater or smaller depending on the sign and magnitude of Cov(X, Y).

How does the covariance of X and Y affect the variance of their sum?

The covariance of X and Y directly affects the variance of their sum. The variance of the sum of X and Y is given by Var(X + Y) = Var(X) + Var(Y) + 2Cov(X, Y). Therefore, if Cov(X, Y) is positive, it increases the variance of their sum, and if Cov(X, Y) is negative, it decreases the variance of their sum.

Similar threads

Replies
3
Views
2K
Replies
30
Views
3K
Replies
5
Views
2K
Replies
9
Views
1K
Replies
1
Views
878
Replies
5
Views
1K
Back
Top