Beta coefficient vs. Significance coefficient

In summary, the conversation discusses the use of significance values and beta values in multiple regression analysis for analyzing data from a study. The significance value is important in determining which variables should be included in the model, while the beta value helps characterize the linear model. The traditional use of the terms alpha and beta in statistics is also mentioned, with beta referring to the probability of Type II error and alpha referring to the probability of Type I error. The conversation also includes a story to explain the difference between Type I and Type II errors.
  • #1
Kontilera
179
24
Hello Physicsforums!
I've stumbeld on a problem in statistics (which is one of my weaknesses when it comes to math).

I´m currently trying to analyze data from a study and using multiple regression analysis.
From regression analysis with one independent variable I´m used to mainly focus on the significance values. When there are more independent variables however I understand that it is the beta and tolerance values that are of importance.

I am now wondering if anybody could help me understand why the significance value isn't relevant?

+ what determines wheater it is the standardized or not standardized coefficients that are relevant for my study?

Thanks in advance!
 
Physics news on Phys.org
  • #2
The statistical significance of a variable in a model addresses a different issue from it's standardized coefficient (aka beta). If a variable has a large beta, changing it tends to lead to a large change of the estimated variable, but it doesn't say how statistically strong that relationship is. A variable can have a large beta and a small statistical significance (or visa versa). When deciding if a variable should statistically be in the linear model, the statistical significance is important. Once the model is determined and the statistically best variables are in it, beta helps you characterize that linear model.
 
  • #3
Traditional use of the terms alpha and beta in statistics:
Beta is the probability of Type II error in any hypothesis test. It tells us if we are
incorrectly concluding no statistical significance.
Alpha is the probability of Type I error in any hypothesis test - wrongly claiming
statistical significance.

Tiger chow:
This dumb but essentially correct story helped a lot of my students to get Type I and Type II.

Have you ever seen faces in clouds? Are there really faces up there or is this a construct
of some hard-wiring of neurons in the brain?
Long time ago, two people are walking home. They see a bush way up ahead. Person A says:
It kinda looks like there is a tiger behind that bush. Person B says: I dunno, but that is
the shortest way home. Person A and B go their chosen ways. Most of the time B gets home,
but sometimes winds up as tiger chow. Long term people behaving like Person A contributed
more offspring to the gene pool. Person B is committing Type II errors long term. And hence not
around to have and raise kids.

Type II == tiger chow! Type I is more harmless. Why we often see faces when there are none there
most times. Or get that creepy uneasy feeling in some places.

Note: some texts and lots of software unfortunately use "beta" for population regression
coefficients.

Do not confuse the two concepts. I cannot tell if that is what is going on here.
 
  • #4
jim mcnamara said:
Traditional use of the terms alpha and beta in statistics:
Beta is the probability of Type II error in any hypothesis test. It tells us if we are
incorrectly concluding no statistical significance.
That is not the same beta that the OP is referring to. The OP beta refers "to how many standard deviations a dependent variable will change, per standard deviation increase in the predictor variable." (see https://en.wikipedia.org/wiki/Standardized_coefficient )
 
  • #5
@FactChecker - Wasn't clear to me. Thanks for the correction. And as you know there are several meanings for beta, so instead of the OP
being confused, I was.
 
  • Like
Likes FactChecker

FAQ: Beta coefficient vs. Significance coefficient

What is a beta coefficient?

A beta coefficient, also known as a regression coefficient, is a measure of the relationship between two variables in a linear regression model. It represents the change in the dependent variable for a one-unit change in the independent variable.

What is a significance coefficient?

A significance coefficient, also known as a p-value, is a measure of the probability that the relationship between two variables in a linear regression model is due to chance. It is used to determine if the relationship is statistically significant.

What is the difference between a beta coefficient and a significance coefficient?

The main difference between a beta coefficient and a significance coefficient is their purpose. Beta coefficients measure the strength and direction of the relationship between variables, while significance coefficients determine if this relationship is statistically significant or just due to chance.

Which is more important, a beta coefficient or a significance coefficient?

Both beta and significance coefficients are important in a linear regression model. Beta coefficients help to understand the relationship between variables, while significance coefficients indicate if this relationship is reliable. Therefore, both are necessary for interpreting the results of a regression analysis.

How do beta and significance coefficients affect each other?

Beta and significance coefficients are related but not dependent on each other. A beta coefficient can be significant or non-significant, and a significant p-value does not necessarily mean a strong relationship. However, a significant beta coefficient with a low p-value indicates a strong and reliable relationship between variables.

Back
Top