Statistics: What is the efficient estimator of sigma^2?

  • Thread starter sanctifier
  • Start date
  • Tags
    Statistics
In summary, the efficient estimator of σ^2 from the data samples X_i is given as:\overline{X}_n = \frac{2v^2}{n} \lambda' (X|v) + v = \frac{2\sigma^4}{n}\lambda' (X|\sigma^2) + \sigma^2
  • #1
sanctifier
58
0

Homework Statement



[itex] X_{1},\; X_{2},\;\;...,\;\; X_{n} [/itex] are sampled from a normal random variable x of mean [itex] \mu =0 [/itex] and variance [itex] 4 \sigma ^2 [/itex]

Question: What is the efficient estimator of [itex] \sigma ^2 [/itex]

Homework Equations


Nothing Special.

The Attempt at a Solution


When [itex] \sigma [/itex] is given, the joint probability density function (p.d.f.) of [itex] X_{1},\; X_{2},\;\;...,\;\; X_{n} [/itex] is

[itex] f(X| \sigma ^2) = \frac{1}{( \sqrt[2]{2 \pi }2 \sigma )^n}exp\{- \frac{1}{2}\sum_{i=1}^{n} \frac{{X_i}^2}{4\sigma^2} \} [/itex]

Where [itex] X [/itex] denotes a vector of components [itex] X_{1},\; X_{2},\;\;...,\;\; X_{n} [/itex].

Let [itex] \lambda (X|\sigma^2) = -n*ln( \sqrt{2 \pi } 2\sigma)- \frac{1}{2}\sum_{i=1}^{n} \frac{{X_i}^2}{4\sigma^2} [/itex]

Namely, [itex] \lambda (X|v) = -n*ln( \sqrt{8 \pi v} )- \frac{1}{2}\sum_{i=1}^{n} \frac{{X_i}^2}{4v} [/itex] when [itex] v=\sigma^2 [/itex]

[itex] \lambda' (X|v) = \frac{d\lambda (X|v)}{dv} = - \frac{n}{2v} + \frac{n\overline{X} _n}{2v^2} [/itex]

Where [itex] \overline{X} _n = \sum_{i=1}^{n} X_i / n [/itex]

Hence, the efficient estimator of [itex] \sigma^2 [/itex] is

[itex] \overline{X} _n = \frac{2v^2}{n} \lambda' (X|v) + v = \frac{2\sigma^4}{n}\lambda' (X|\sigma^2) + \sigma^2 [/itex]

Is this correct?
 
Last edited:
Physics news on Phys.org
  • #2
Maybe it's my ignorance of the topic, but the question makes no sense to me. If you are given σ, why do you need an estimator for it?
 
  • #3
I think that he's asking for the best estimate of sigma^2 from the data samples X_i. I'm no mathematician so discount the following comments appropriately.

First, I am not following your approach, and your final result makes no sense to me--in the last line, in fact, you have equated a mean on the left to a variance on the right.

To point in a different direction, I know that the sample variance [tex]\hat{\sigma}^2=E[X_i^2]-E[\overline{X_i}^2][/tex] is related to the true population variance by [tex]E[\hat{\sigma}^2]=\frac{N-1}{N}\sigma^2.[/tex] Also I seem to recall that the sample variance is the maximum likelihood estimate of the population variance for the case of normally distributed random variables. Perhaps this will help (and I'll leave the demonstrations and derivations to you...).
 
  • #4
haruspex said:
Maybe it's my ignorance of the topic, but the question makes no sense to me. If you are given σ, why do you need an estimator for it?

You are not actually given σ; you are given that there is such a σ. This is a very standard type of question, and in the field the "..but σ is not known.." part of the statement is usually understood. However, calling the variance ##4 \sigma^2## instead of ##\sigma^2## is just silly, and is, I suspect, introduced in order to try to trick the student.
 
  • #5
haruspex said:
Maybe it's my ignorance of the topic, but the question makes no sense to me. If you are given σ, why do you need an estimator for it?

Thank you for reply.

Sorry for causing confusion, "σ is given but unknown" just means although σ remains in the formula, it acts like the unknown variable and needs to be estimated through the known values of [itex] X_{1},\; X_{2},\;\;...,\;\; X_{n} [/itex].

marcusl said:
I think that he's asking for the best estimate of sigma^2 from the data samples X_i. I'm no mathematician so discount the following comments appropriately.

First, I am not following your approach, and your final result makes no sense to me--in the last line, in fact, you have equated a mean on the left to a variance on the right.

To point in a different direction, I know that the sample variance [tex]\hat{\sigma}^2=E[X_i^2]-E[\overline{X_i}^2][/tex] is related to the true population variance by [tex]E[\hat{\sigma}^2]=\frac{N-1}{N}\sigma^2.[/tex] Also I seem to recall that the sample variance is the maximum likelihood estimate of the population variance for the case of normally distributed random variables. Perhaps this will help (and I'll leave the demonstrations and derivations to you...).

Thank you for mentioning the M.L.E.

Actually, an efficient estimator is based on the Cramér–Rao inequality. The equality in this inequality holds when a certain condition is met, i.e., [itex] T=u(\theta) \lambda_n ' (X|\theta)+v(\theta) [/itex], where [itex] X [/itex] denotes a vector of components [itex] X_{1},\; X_{2},\;\;...,\;\; X_{n} [/itex], actually it means the function [itex] \lambda (X|\theta) [/itex] is a multivariable function of [itex] X_{1},\; X_{2},\;\;...,\;\; X_{n} [/itex] whose parameter [itex] \theta [/itex] is unknown. [itex] u(\theta) [/itex] and [itex] v(\theta) [/itex] are functions not involving any part of [itex] X [/itex], and

[itex] \lambda_n(X|\theta) = lnf(X|\theta) [/itex]

[itex] \lambda_n ' (X|\theta) = \frac{d lnf(X|\theta)}{d\theta} [/itex]

Ray Vickson said:
You are not actually given σ; you are given that there is such a σ. This is a very standard type of question, and in the field the "..but σ is not known.." part of the statement is usually understood. However, calling the variance ##4 \sigma^2## instead of ##\sigma^2## is just silly, and is, I suspect, introduced in order to try to trick the student.

Ray knows the rules.

Can you tell me whether the answer is correct or not? Thank you in advance.
 
Last edited:
  • #6
sanctifier said:

Homework Statement



[itex] X_{1},\; X_{2},\;\;...,\;\; X_{n} [/itex] are sampled from a normal random variable x of mean [itex] \mu =0 [/itex] and variance [itex] 4 \sigma ^2 [/itex]

Question: What is the efficient estimator of [itex] \sigma ^2 [/itex]

Homework Equations


Nothing Special.


The Attempt at a Solution


When [itex] \sigma [/itex] is given, the joint probability density function (p.d.f.) of [itex] X_{1},\; X_{2},\;\;...,\;\; X_{n} [/itex] is

[itex] f(X| \sigma ^2) = \frac{1}{( \sqrt[2]{2 \pi }2 \sigma )^n}exp\{- \frac{1}{2}\sum_{i=1}^{n} \frac{{X_i}^2}{4\sigma^2} \} [/itex]

Where [itex] X [/itex] denotes a vector of components [itex] X_{1},\; X_{2},\;\;...,\;\; X_{n} [/itex].

Let [itex] \lambda (X|\sigma^2) = -n*ln( \sqrt{2 \pi } 2\sigma)- \frac{1}{2}\sum_{i=1}^{n} \frac{{X_i}^2}{4\sigma^2} [/itex]

Namely, [itex] \lambda (X|v) = -n*ln( \sqrt{8 \pi v} )- \frac{1}{2}\sum_{i=1}^{n} \frac{{X_i}^2}{4v} [/itex] when [itex] v=\sigma^2 [/itex]

[itex] \lambda' (X|v) = \frac{d\lambda (X|v)}{dv} = - \frac{n}{2v} + \frac{n\overline{X} _n}{2v^2} [/itex]

Where [itex] \overline{X} _n = \sum_{i=1}^{n} X_i / n [/itex]

Hence, the efficient estimator of [itex] \sigma^2 [/itex] is

[itex] \overline{X} _n = \frac{2v^2}{n} \lambda' (X|v) + v = \frac{2\sigma^4}{n}\lambda' (X|\sigma^2) + \sigma^2 [/itex]

Is this correct?

If you are trying to find the maximum of ##\lambda## then no, it is not correct. You need ##0 = \lambda' = (n/2)(-w + w^2 S)## where ##w = 1/v## and ##S = (1/n)\sum X_i^2##. Also, your final line makes no sense: you are giving a formula for an estimator of ##\sigma## that contains the value of ##\sigma## itself. The whole point of an estimator is that it contains only the observed quantities ##X_i## and ##n##, but not the true, underlying parameters of the distribution.
 

FAQ: Statistics: What is the efficient estimator of sigma^2?

What is the purpose of estimating sigma^2 in statistics?

The purpose of estimating sigma^2, also known as the variance, is to determine the spread or variability of a population or sample. It is a key measure in understanding the data and making statistical inferences.

What is the efficient estimator of sigma^2?

The efficient estimator of sigma^2 is the sample variance, denoted as s^2. This estimator is considered efficient because it has the smallest variance among all unbiased estimators of sigma^2.

How is the sample variance calculated?

The sample variance is calculated by taking the sum of squared differences between each data point and the sample mean, dividing it by the sample size minus one, and then taking the square root. The formula for the sample variance is: s^2 = Σ(x - x̄)^2 / (n - 1)

Can the sample variance be biased?

Yes, the sample variance can be biased if the sample data is not representative of the population or if there is a significant amount of outliers in the data. In these cases, a different estimator such as the adjusted sample variance may be used to reduce bias.

How is the sample variance related to the population variance?

The sample variance is an unbiased estimator of the population variance, meaning that on average, the sample variance will equal the population variance. As the sample size increases, the sample variance becomes a more accurate estimate of the population variance.

Similar threads

Replies
8
Views
1K
Replies
8
Views
2K
Replies
5
Views
2K
Replies
47
Views
3K
Replies
12
Views
2K
Back
Top