Linear Regression β: Estimating η with MLEs

In summary: I don't understand what it means by "let x be some additional given value of the explanatory variable"
  • #1
stukbv
118
0
β

Homework Statement



Data y1,y2...yn are modeled as observations of random variables Y1,..Yn given by

Yi = α + β(xi-xbar) + σεi

Where α , β and σ are unknown parameters x1,x2...xn are known constants and xbar is
(1/n)Ʃxi and εi's are independent random variables each with the gaussian distribution mean 0 and unknown variance 1.

Now let x be some additional given value of the explanatory variable. Construct from you estimates of the parameters a suitable estimate for η = α + β(x-xbar), the mean value of the response variable when the explanatory variable is x.

Homework Equations



From a previous question, for the initial problem where only xi's are the explanatory variable I calculate that;

∂l/∂α = Ʃyi - βƩ(xi-xbar) = nα where l is the log likelihood function.

And then obviously Ʃ(xi-xbar) = 0 , so the MLE(α) = Ʃyi /n in the original case.

and ∂l/∂β gives MLE(β) = Ʃyi(xi-xbar) / Ʃ(xi-xbar)2


The Attempt at a Solution



I assume that to find the MLE(η) I can just add the MLE(α) and MLE(β)(x-xbar)
where the MLE(β) and MLE(α) are the maximum likelihood estimates for ach parameter in the new regression, with the additional x. But this is the problem, how do I find the MLE's of this regression using the MLE's of the old regression?

What I have done is replaced xi with x in ∂l/∂α to get Ʃyi - βƩ(x-xbar) = nα which then gives Ʃyi - nβ(x-xbar) .= nα because the x's don't depend on i.
Which gives the new MLE(α) = Ʃyi/n - β(x-xbar)

Now for MLE(β) I did the same, replaced xi in the original MLE with x.
This gives MLE(β)(x-xbar) = Ʃyi/n - α(x-xbar).

I then replaced the α in the MLE(β) with the MLE(α) for this regression.
Solving that i get MLE(β)(x-xbar) = Ʃyi/n and MLE(α) = 0.

Giving the suitable estimate of η as Ʃyi/n which I think is wrong...
 
Physics news on Phys.org
  • #2
stukbv said:
β

Homework Statement



Data y1,y2...yn are modeled as observations of random variables Y1,..Yn given by

Yi = α + β(xi-xbar) + σεi

Where α , β and σ are unknown parameters x1,x2...xn are known constants and xbar is
(1/n)Ʃxi and εi's are independent random variables each with the gaussian distribution mean 0 and unknown variance 1.

Now let x be some additional given value of the explanatory variable. Construct from you estimates of the parameters a suitable estimate for η = α + β(x-xbar), the mean value of the response variable when the explanatory variable is x.

Homework Equations



From a previous question, for the initial problem where only xi's are the explanatory variable I calculate that;

∂l/∂α = Ʃyi - βƩ(xi-xbar) = nα where l is the log likelihood function.

And then obviously Ʃ(xi-xbar) = 0 , so the MLE(α) = Ʃyi /n in the original case.

and ∂l/∂β gives MLE(β) = Ʃyi(xi-xbar) / Ʃ(xi-xbar)2


The Attempt at a Solution



I assume that to find the MLE(η) I can just add the MLE(α) and MLE(β)(x-xbar)
where the MLE(β) and MLE(α) are the maximum likelihood estimates for ach parameter in the new regression, with the additional x. But this is the problem, how do I find the MLE's of this regression using the MLE's of the old regression?

What I have done is replaced xi with x in ∂l/∂α to get Ʃyi - βƩ(x-xbar) = nα which then gives Ʃyi - nβ(x-xbar) .= nα because the x's don't depend on i.
Which gives the new MLE(α) = Ʃyi/n - β(x-xbar)

Now for MLE(β) I did the same, replaced xi in the original MLE with x.
This gives MLE(β)(x-xbar) = Ʃyi/n - α(x-xbar).

I then replaced the α in the MLE(β) with the MLE(α) for this regression.
Solving that i get MLE(β)(x-xbar) = Ʃyi/n and MLE(α) = 0.

Giving the suitable estimate of η as Ʃyi/n which I think is wrong...

Every textbook on linear regression contains detailed formulas for doing all the calculations in your question. If you do not use a textbook, there are numerous treatments on-line that explain what needs to be done.

RGV
 
  • #3
Ok, but I can't find them because I don't know what I'm looking for.. I don't understand what it means by let x be some additional given value of the explanatory variable. I mean, does this mean we have n+1 explanatory variables x1,x2...xn,x
so that Y = α + β(x-xbar) + σε is the response variable to x and xbar now is
x1+x2+..+xn+x/n+1
OR
does it mean that we still have n explanatoty values but each has x added onto them, so that they are x1+x,x2+x...xn+x and xbar now is x1+..+xn/n + x and
Yi = α + β(xi+x-xbar) + σεi ?

I know this may be a stupid question but it's really confused me.
Thanks for your help
 
  • #4
stukbv said:
Ok, but I can't find them because I don't know what I'm looking for.. I don't understand what it means by let x be some additional given value of the explanatory variable. I mean, does this mean we have n+1 explanatory variables x1,x2...xn,x
so that Y = α + β(x-xbar) + σε is the response variable to x and xbar now is
x1+x2+..+xn+x/n+1
OR
does it mean that we still have n explanatoty values but each has x added onto them, so that they are x1+x,x2+x...xn+x and xbar now is x1+..+xn/n + x and
Yi = α + β(xi+x-xbar) + σεi ?

I know this may be a stupid question but it's really confused me.
Thanks for your help

If I understand your original question correctly, you have a single independent variable, x, and a single dependent variable, y, connected by an equation y = γ + βx. You have n observed values of x, call them x1, x2, ..., xn (not necessarily all different) and n observed values of y, call them y1, 2, ..., yn. Your model is that the observed values are yi = γ + βxi + σεi, where σ is a constant and the εi are iid N(0,1). You don't know γ, β or σ, but instead have to estimate them from the observed data, which you do via the least-squared method. These give estimates c of γ, b of β and s2 of σ^2. Of course, your c and b are just single observations of some random variables C and B. Normally, when we just have a single observation of a random variable, we cannot say very much at all about accuracy, etc., but in the present case the underlying random variables C and B are functions obtained by doing some calculations on n items of data, so we actually have (n-2) useful, leftover items that can be used to assess accuracy. Not surprisingly, the Student t-distribution with (n-2) degrees of freedom will be involved.

Now, why did you bother to fit an equation to data? Presumably, you would like to use that equation in the future to make estimates of y and so forth. So, you want to use the equation to make predictions. In other words, given a new value of x (maybe one you have not yet observed), what can you say about y? Well, if your underlying model really is correct, you have an unbiased estimate of the mean of y---namely, c + b*x. But, suppose you want to construct "error bars" around your estimate---that is, you want a "prediction interval". How could you do that? I think that is what your question is asking, and a Google search on 'linear regression + prediction", or something similar, will give you all you need. The formulas are a bit too lengthy to present here, and anyway, I don't personally believe in the value of just writing down formulas for somebody and saying "here they are". You need to read and absorb that material, and make it your own.

By the way, I chose to write c + b*x instead of a + b*(x - x_bar), but given a single data set you can pass easily between the two. The reason I prefer c + b*x is that the underlying form γ + β x does not depend on any particular data set, while the underlying form α + β (x - bar_x) is not really defined until we know bar_x. You see the difference?

RGV
 

FAQ: Linear Regression β: Estimating η with MLEs

What is the objective of Linear Regression?

The objective of Linear Regression is to find the best fitting line that represents the relationship between a dependent variable (η) and one or more independent variables (β). This line is used to make predictions about the dependent variable based on the values of the independent variables.

What is the role of MLEs in Linear Regression?

MLEs (Maximum Likelihood Estimates) are used in Linear Regression to estimate the parameters (β) of the regression line. The MLEs are calculated by maximizing the likelihood function, which measures the probability of the observed data given the estimated parameters. This allows us to find the most likely values for the parameters that best fit the data.

How do you interpret the β values in Linear Regression?

The β values (also known as regression coefficients) in Linear Regression represent the change in the dependent variable (η) for a one-unit change in the corresponding independent variable. For example, if the β value for a variable x is 0.5, it means that a one-unit increase in x is associated with a 0.5 unit increase in η.

What is the significance of the intercept term in Linear Regression?

The intercept term (β0) in Linear Regression represents the predicted value of the dependent variable (η) when all independent variables are equal to 0. This is important because it allows us to make predictions even when the independent variables do not have meaningful values (e.g. predicting the housing price when the number of bedrooms is 0).

How do you measure the goodness of fit in Linear Regression?

The most commonly used measure of goodness of fit in Linear Regression is the R-squared value. This value ranges from 0 to 1 and represents the percentage of variation in the dependent variable (η) that is explained by the independent variables (β). A higher R-squared value indicates a better fit of the regression line to the data.

Back
Top