Explaining Skewness: Why E[(x-u)^3] Determines Direction

  • Thread starter skwey
  • Start date
In summary, skewness is defined as E[((x-μ)/σ)^3] and is an odd moment about the mean. A non-zero skewness indicates that the distribution is not symmetric. When a distribution is skewed to the right, E[(x-μ)^3] is positive, while it is negative when skewed to the left. This is because values far from the mean have a greater impact on skewness and when the distribution is skewed to the right, there are more values on the right side of the mean. This can be compared to balancing weights on a scale. While there may be some variation in the nomenclature, the general idea is the same.
  • #1
skwey
17
0
Hey

In my textbook it says if E[(x-u)^3] is positive the skew is to the right, and negative to the left. u=expected value.

This is ok and I believe it, but I can't see why it is like this. Is there an easy explanation of why this has to hold?
 
Physics news on Phys.org
  • #2
The first thing to note is that skewness is defined as E[((x-μ)/σ)3], so your E[(x-μ)3] is just the skewness scaled by the cube of the standard deviation. If a distribution has positive (negative) skewness then E[(x-μ)3] is positive (negative) by definition.

The next thing to note is that skewness is an odd moment (e.g., E[(x-μ)1], E[(x-μ)3], E[(x-μ)5], ...) about the mean. All of these odd moments would be zero if the distribution was symmetric. So a non-zero value skewness necessarily means the distribution is not symmetric. (The converse is not true. A zero skewness does not necessarily mean the distribution is symmetric.)

So when talking about a distribution with a positive (negative) skewness we are talking about a distribution that is not symmetric. Think about what E[((x-μ)/σ)3] is measuring. Values far from the mean (|x-μ|/σ > 1) will weigh much more heavily than those near the mean (|x-μ|/σ < 1). It is those values that are far from the mean that determine the skewness. For the skewness to be negative, there must be more such values far from the mean for x<μ than for x>μ. The opposite is the case when the skewness is positive.
 
  • Like
Likes 1 person
  • #3
Thanks DH you obviously know what you are talking about. And sorry for not scaling it as you mentioned, I didn't do it because I only wanted to know why it was negative and why it was positive at times.

However I hope you can give me some more of your time, cause there is one thing that is still unclear:

Regarding the last paragraph of yours, I have had the same thought as you. But let's for argument sake look at the first moment E[(x-u)], with this moment you can use the same arguments as you use. If the distribution is skewed to the right x-u will tend to be bigger than |x-u| to the left. But the moment is still zero, because the p(x)-values to the left will tend to be bigger, so this will make the moment zero.

My question is: how do you disregard this when looking at the third moment, cause even though |x-u| will tend to be bigger at the skewed side. p(x) will tend to be bigger at the non-skewed side. So why does the effect of |x-u|^3 beat the p(x) "effect"? That is, if we have right skew, why does
(x-u)^3*p(x)dx tend to be more positive on the right, than (x-u)^3*p(x)dx is negative on the left side. Because p(x) will be bigger on the left, ex: chi-squared distributions.
 
Last edited:
  • #4
skwey said:
Thanks DH you obviously know what you are talking about. And sorry for not scaling it as you mentioned, I didn't do it because I only wanted to know why it was negative and why it was positive at times.

However I hope you can give me some more of your time, cause there is one thing that is still unclear:

Regarding the last paragraph of yours, I have had the same thought as you. But let's for argument sake look at the first moment E[(x-u)], with this moment you can use the same arguments as you use. If the distribution is skewed to the right x-u will tend to be bigger than |x-u| to the left. But the moment is still zero, because the p(x)-values to the left will tend to be bigger, so this will make the moment zero.

My question is: how do you disregard this when looking at the third moment, cause even though |x-u| will tend to be bigger at the skewed side. p(x) will tend to be bigger at the non-skewed side. So why does the effect of |x-u|^3 beat the p(x) "effect"? That is, if we have right skew, why does
(x-u)^3*p(x)dx tend to be more positive on the right, than (x-u)^3*p(x)dx is negative on the left side. Because p(x) will be bigger on the left, ex: chi-squared distributions.

Consider each term in the expectation: If there are more terms left of the mean than the right when considered with respect to the frequency of each value in the domain, it would make sense that you get a negative value.

If the distribution was skewed to the left (i.e. had a right tail), you would expect a negative result since the proportion of values of (x-mu)^3 would be negative and we would have a higher frequency of weight being negative.

Think of a scale where you have an object on the left scale, and weights on the right hand scale. You have to balance the weights with the object to get an exact reading: it works in the same sort of way for the skewness.
 
Last edited:
  • #5
chiro said:
If the distribution was skewed to the left (i.e. had a right tail), you would expect a negative result since the proportion of values of (x-mu)^3 would be negative and we would have a higher frequency of weight being negative..
That's a bit counter to the more or less standard mean of skewed to the left versus skewed to the right.

For example, this distribution with a long tail to the right is typically denoted as skewed right.
histogr6.gif


This nomenclature (skewed right means a longer tail to the right) is widespread but is not universal. Some would call the above distribution as being skewed to the left, evidenced by the mode and median being to the left of the mean.
 
  • #6
If my terminology was wrong I apologize to the OP (and think D H for the correction). In saying this the premise of my general message still stands.
 

FAQ: Explaining Skewness: Why E[(x-u)^3] Determines Direction

1. What is skewness?

Skewness is a measure of the asymmetry of a distribution, indicating how much the data is skewed to one side or the other. It is an important statistical concept in understanding the shape and characteristics of a dataset.

2. How is skewness calculated?

Skewness is typically calculated using a formula that measures the difference between the mean, median, and mode of a dataset. This value can be positive, negative, or zero, indicating a positive skew, negative skew, or no skew, respectively.

3. What does a positive skew indicate?

A positive skew indicates that the data is skewed to the right, with a longer tail on the right side of the distribution. This means that there are more values on the left side of the distribution and fewer values on the right side.

4. How does skewness affect data analysis?

Skewness can affect data analysis in several ways. For example, a highly skewed dataset may lead to biased results when using certain statistical methods, such as the mean. It can also impact the interpretation of data and make it difficult to compare different datasets.

5. Can skewness be corrected?

In some cases, skewness can be corrected by applying mathematical transformations to the data, such as logarithmic or square root transformations. However, it is important to consider the underlying reasons for the skewness and whether it is appropriate to correct it for the specific analysis being conducted.

Similar threads

Replies
8
Views
2K
Replies
1
Views
2K
Replies
8
Views
903
Replies
62
Views
3K
Replies
6
Views
922
Back
Top