Basic statistics calibration question

In summary, the conversation discusses the use of standard deviation as a measure of accuracy for weight scales. The question posed in the interview was deemed incorrect as standard deviation is not necessarily the best indicator of accuracy. The conversation also touches on the importance of considering the average in determining accuracy and the use of a calibration formula for scales. The formula for standard deviation with a mean of 10 is used to calculate the accuracy of a set of data given in the conversation.
  • #1
dislect
166
0
Hi guys,

My friend was asked the following question in an interview:

There are 3 weight scales and a 10kg weight.
You weigh the 10kg on each scale 5 times (total of 15 measurements for the 3 scales), each measurement is a bit off the 10kg mark (lets say the first scale shows: 11kg, 9.4kg, 12.1kg, 10.6kg. 10.2kg).

You are suppose to choose the criteria of which scales is more accurate.
Why is it better to use standard deviation instead of the medium or average?
How do you create a calibration formula for the scale?


Thanks!
 
Mathematics news on Phys.org
  • #2
The question is wrong. It is not better to use the standard deviation. This was a trick question, I think.

I won't say more but think about why the standard deviation might be bad to use.
 
  • #3
I wish you would say more though :)
Standard deviation measures the variation from the average. If the standard deviation for one of the scales is smaller the of the others then the measurements are closer to the average... yet, the average could be something like 20kg which is very far away from the true 10kg. Is that why you ment its not a good indication for accuracy?
So is using the average of measurements for each scale and measuring how far the average is from the 10kg the bets way to measure the accuracy and choose the "best scale"?
How about using the standard deviation formula with the mean value as 10kg instead of the true average of that scale?

Plus, any clarifications about a calibration formula would be great.

thanks
 
  • #4
Yes, my point was that scale A could be less accurate even though it has a smaller standard deviation, if it has a crazy average. Without paying attention to the average, you couldn't calibrate the scale. For me it is more important.

I think by calibration formula they mean a list of steps to follow to calibrate one of the scales as much as possible. I don't you need many hints here, think of Hooke's law and how you would calibrate a scale at home.
 
  • #5
verty said:
Yes, my point was that scale A could be less accurate even though it has a smaller standard deviation, if it has a crazy average. Without paying attention to the average, you couldn't calibrate the scale. For me it is more important.

I think by calibration formula they mean a list of steps to follow to calibrate one of the scales as much as possible. I don't you need many hints here, think of Hooke's law and how you would calibrate a scale at home.

Consider the following set of data:

$$\begin{matrix}500 & 0 & 4922 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 4003 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 3 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \end{matrix}$$

What's its average? What's its standard deviation? :-p

This is something of a ridiculous example, but it shows that in the context of the problem, standard deviation IS the best indicator.

The problem says that each one is "a bit off" the actual value of 10, so the average will be about 10.

Thus, we use our formula for a sample standard deviation, replacing ##\bar{x}=10##, to obtain ##\displaystyle s=\sqrt{\frac{1}{n-1}\sum_{i=1}^{n}\left[(x_i-10)^2\right]}##. The scale with the smallest value of ##s## is the most accurate.
 
  • #6
Ok, great demonstration !
Follow up questions: Why 1/(n-1) and not 1/n?
How would you calibrate the following results: 11kg, 9.4kg, 12.1kg, 10.6kg. 10.2kg?
 
  • #7
With regards to n-1 its because n-1 provides an unbiased estimate of the population value.

In statistics we use random variables to estimate population parameters (like mu or sigma). An unbiased estimator means that E[theta_hat] = theta where theta is the true population value and theta_hat is an estimator that represents the estimator distribution of theta.
 
  • #8
dislect said:
Ok, great demonstration !
Follow up questions: Why 1/(n-1) and not 1/n?
How would you calibrate the following results: 11kg, 9.4kg, 12.1kg, 10.6kg. 10.2kg?
There are ##n-1## degrees of freedom.

For the data you gave, I would calculate ##s=\sqrt{\frac{(11-10)^2+(9.4-10)^2+(12.1-10)^2+(10.6-10)^2+(10.2-10)^2}{5-1}}=\sqrt{\frac{1+.36+4.41+.36+.04}{4}}=\frac{1}{2}\sqrt{6.17}\approx 1.241975 \text{ kg}##.

Note that the calculation assumes the "average" is 10, so it is not a standard form of standard deviation as much as a measure of the deviation from the "wanted" mean. The average of the data is not 10, but it should be fairly close.
 

Related to Basic statistics calibration question

1. What is meant by "basic statistics calibration"?

Basic statistics calibration refers to the process of adjusting or correcting data to improve its accuracy or reliability. This is typically done by comparing the data to a known standard or reference point and making necessary adjustments based on the differences.

2. Why is calibration important in statistical analysis?

Calibration is important because it helps to ensure that the data used in statistical analysis is accurate and reliable. Without proper calibration, there may be errors or biases in the data that can lead to incorrect conclusions and decisions.

3. How is calibration performed in statistics?

Calibration is typically performed by collecting data and comparing it to a known standard or reference point. Based on the differences between the data and the standard, adjustments can be made to improve the accuracy of the data. This process may involve using mathematical techniques and statistical methods.

4. What are some common types of calibration methods used in statistics?

Some common types of calibration methods include linear regression, least squares regression, and polynomial regression. These methods involve fitting a mathematical model to the data and using it to make adjustments or corrections.

5. Can calibration be applied to all types of data?

Yes, calibration can be applied to all types of data, as long as there is a known standard or reference point to compare it to. However, the specific calibration methods used may vary depending on the type of data and the purpose of the analysis.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
1K
  • MATLAB, Maple, Mathematica, LaTeX
Replies
6
Views
3K
  • Set Theory, Logic, Probability, Statistics
Replies
9
Views
1K
Replies
152
Views
5K
  • Introductory Physics Homework Help
Replies
9
Views
2K
Replies
1
Views
1K
  • Calculus and Beyond Homework Help
Replies
1
Views
2K
  • Programming and Computer Science
Replies
1
Views
2K
  • Electrical Engineering
Replies
2
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
6
Views
2K
Back
Top