Error propagation leads to different uncertainties, which do I choose?

In summary, I understand how to compute and propagate errors, but have trouble conceptualizing all things put together. I performed an experiment to determine a value for some quantity, which depends on two variables. The first variable has a constant fractional error, but the second variable's fractional error decreases as the magnetic flux density increases. I calculated the standard deviation, but found it to be too low. I narrowed it down to two questions: 1. How do I choose and uncertainty here? do I pick one from the higher end, like 5% or should I average the errors? that would make it about 4 %. It feels like I'm nit picking but I'm just thinking conceptually. 2. I calculated the standard deviation as well
  • #1
TheMercury79
24
5
Homework Statement
Propagation of uncertainties, standard deviation
Relevant Equations
$\frac{\deltaQ(x,y,z,...)}{Q(x,y,z,...)} = \sqrt{(\frac{\partial Q}{\partial x}\delta x)^2+(\frac{\partial Q}{\partial y}\delta y)^2+(\frac{\partial Q}{\partial z}\delta z)^2+...}$
I understand how to compute and propagate errors but have trouble with conceptualizing all things put together.

I have performed an experiment to determine a value for some quantity. This quantity depend on two variables. The first one depend in turn on some other quantities as well but I think I have managed to propagate errors correctly.

For the experiment I have made measurements for some ten or so points. The fractional error for the first variable is constant for all these points. The second variable is magnetic flux density B for which there is a constant error of say 5 mT at each point . As B increases, the fractional error in B then decreases and so the final outcome also has decreasing error as B gets higher. and gets lower and lower for each point. The quantity I want to measure then has errors that ranges from 5.5% to 3.5%. How do I choose an uncertainty from this? I'm thinking 5 % because the value that has this error is closest to the average value, but I don't know. The thing is 5 % is kind of a good outcome because then the true value is just within the error range. But 3.5 % is a little disappointing, so is it cheating if I choose among the highest error?

I narrowed it down to 2 questions:

1. How do I choose and uncertainty here? do I pick one from the higher end, like 5% or should I average the errors? that would make it about 4 %. It feels like I'm nit picking but I'm just thinking conceptually.

2. I calculated the standard deviation as well but that narrowed down the error to about 1 % but I feel that is too low. Do we use standard deviation only when we have values without uncertainties?

Just some pointers would be appreciated.
 
Physics news on Phys.org
  • #2
I think your question would be easier to answer if you told us exactly what you have done rather than speak in abstract terms. Or at least be a bit more specific.
 
  • Like
Likes BvU
  • #3
Hi,

Please use double # or $ as delimiters for ##\LaTeX##...
And preview before posting ...

Did you mean
$$\bigl (\Delta Q(x,y,z,...)\bigr )^2 = \left (\frac {\partial Q}{\partial x} \Delta x\right )^2+
\left (\frac {\partial Q}{\partial y} \Delta y\right )^2 +\left (\frac {\partial Q}{\partial z} \Delta z\right )^2+ ...\ ?$$

From your rather vague story I gather you have varied ##B## and measured something. What ?
There is some dependence (linear ?) so you must have made a plot. Can you post it ?

TheMercury79 said:
The quantity I want to measure
Did you mean 'determine' as opposed to measure ?

Ther is too much to guess and assume so I refrain from answering until your description is clearer.

##\ ##
 
  • Like
Likes Orodruin
  • #4
TheMercury79 said:
Homework Statement:: Propagation of uncertainties, standard deviation
Relevant Equations:: $$\frac{\delta Q(x,y,z,...)}{Q(x,y,z,...)} = \sqrt{(\frac{\partial Q}{\partial x}\delta x)^2+(\frac{\partial Q}{\partial y}\delta y)^2+(\frac{\partial Q}{\partial z}\delta z)^2+...}$$

1. How do I choose and uncertainty here? do I pick one from the higher end, like 5% or should I average the errors? that would make it about 4 %. It feels like I'm nit picking but I'm just thinking conceptually.

2. I calculated the standard deviation as well but that narrowed down the error to about 1 % but I feel that is too low. Do we use standard deviation only when we have values without uncertainties?
You need to distinguish two things:
  • the uncertainty in each individual value calculated
  • the overall uncertainty when they are combined.
You could ignore the first and just compute the second from the standard deviation. But the uncertainty is not actually the standard deviation. That would tend to increase with more datapoints, whereas it should reduce. Instead, use the "standard error of the mean": https://en.wikipedia.org/wiki/Standard_error.

Clearly it would be better to use both sources of information, but I am unaware of any standard procedure for doing so. @Orodruin may well know one.
 
  • #5
BvU said:
Hi,

Please use double # or $ as delimiters for ##\LaTeX##...
And preview before posting ...

Did you mean
$$\bigl (\Delta Q(x,y,z,...)\bigr )^2 = \left (\frac {\partial Q}{\partial x} \Delta x\right )^2+
\left (\frac {\partial Q}{\partial y} \Delta y\right )^2 +\left (\frac {\partial Q}{\partial z} \Delta z\right )^2+ ...\ ?$$

From your rather vague story I gather you have varied ##B## and measured something. What ?
There is some dependence (linear ?) so you must have made a plot. Can you post it ?Did you mean 'determine' as opposed to measure ?

Ther is too much to guess and assume so I refrain from answering until your description is clearer.

##\ ##
What is up with your answer?. It is obvious I just forgot an extra '$' so it is not necessary to ask about my equation. Yes I said 'measure', you're right I should have said 'determine', but you still know what I meant. It depends on two quantitites, their fractional errors add in quadrature during progagation, what is unclear about this? Should I average the final errors or choose the highest one? (oh, I probably should have said 'greatest one' here, right?)
 
  • #6
TheMercury79 said:
it is not necessary to ask about my equation
But it is: the equation is not correct. You can check that by looking at the dimensions.

I did not mean to be rude. Do you ?

##\ ##
 
  • Like
Likes Orodruin
  • #7
haruspex said:
Clearly it would be better to use both sources of information, but I am unaware of any standard procedure for doing so.

Here's a start. You have values ##(y_i)## with a priori variances ##\sigma_i##.
Taking each to be normally distributed about some unknown ideal value ##\hat y##, the marginal probability of the values is ##p((y_i))=(2\pi)^{-n/2}\frac 1{\Pi\sigma_i}e^{-\Sigma{\frac{(y_i-\hat y)^2}{(2\sigma_i)^2}}}##.
To maximise that we would want ##\hat y## to be ##y'=S\Sigma\frac{y_i}{\sigma_i^2}## where ##1/S=\Sigma\frac 1{\sigma_i^2}##. So that is our best estimate for ##\hat y##.
To find the mean error in that, we need to estimate ##E[(y'-\hat y)^2]##.
I got as far as the expression ##S^2\frac{(2\pi)^{-n/2}}{\Pi\sigma_i}\int..\int(\Sigma\frac{y_i-\hat y}{\sigma_i^2})^2e^{-\Sigma{\frac{(y_i-\hat y)^2}{(2\sigma_i)^2}}}d(y_i)##.
I see hints that this should simplify greatly, but it could take me a long time to do it without error - and that's if I've not already made some.
 
  • #8
TheMercury79 said:
Should I average the final errors or choose the highest one? (oh, I probably should have said 'greatest one' here, right?)
You should use the formula as written. Why do otherwise?
 
  • #9
hutchphd said:
You should use the formula as written. Why do otherwise?
See post #4.
 
  • #10
haruspex said:
Here's a start. You have values ##(y_i)## with a priori variances ##\sigma_i##.
Taking each to be normally distributed about some unknown ideal value ##\hat y##, the marginal probability of the values is ##p((y_i))=(2\pi)^{-n/2}\frac 1{\Pi\sigma_i}e^{-\Sigma{\frac{(y_i-\hat y)^2}{(2\sigma_i)^2}}}##.
To maximise that we would want ##\hat y## to be ##y'=S\Sigma\frac{y_i}{\sigma_i^2}## where ##1/S=\Sigma\frac 1{\sigma_i^2}##. So that is our best estimate for ##\hat y##.
To find the mean error in that, we need to estimate ##E[(y'-\hat y)^2]##.
I got as far as the expression ##S^2\frac{(2\pi)^{-n/2}}{\Pi\sigma_i}\int..\int(\Sigma\frac{y_i-\hat y}{\sigma_i^2})^2e^{-\Sigma{\frac{(y_i-\hat y)^2}{(2\sigma_i)^2}}}d(y_i)##.
I see hints that this should simplify greatly, but it could take me a long time to do it without error - and that's if I've not already made some.
Generally this works for independent measurements. However, it is not clear to me from the OP’s description that there are no correlations between measurements.
 
  • #11
haruspex said:
See post #4.
I am not seeing the complication here. One assumes the errors to independent (absent information to the contrary) and the result for the two methods, if correctly applied, should be the same. If they are not, then the errors are likely correlated. In this case either a root correlation or bias needs to be identified. Absent that one should choose the larger (more conservative) error estimate.
The Central Limit Theorem is your friend here.
 
  • #12
hutchphd said:
the result for the two methods, if correctly applied, should be the same
No, because of the heteroscedasticity.
Indeed, without taking the a priori distributions into account one would use the wrong value for the best estimate.
 
  • #13
Orodruin said:
Generally this works for independent measurements. However, it is not clear to me from the OP’s description that there are no correlations between measurements.
I see no hint that they are not independent, but I am rather unclear as to how the a priori errors vary. I just assume @TheMercury79 can work out the likely variance of each datapoint.

@TheMercury79, you have not clarified how you used the standard deviation of the measurements to get an estimate of the error in your answer. Did you use the standard error of the mean formula? If that produces a much smaller error estimate than you expected then I can think of three explanations:
  1. the a priori errors you allowed were excessive
  2. the a priori errors are largely systematic
  3. the very low s.d. observed was a fluke
I think we can rule out the last. The first gives you a big problem since it does not encompass the true value. So I would look for systematic errors.
 
  • #14
haruspex said:
No, because of the heteroscedasticity.
Indeed, without taking the a priori distributions into account one would use the wrong value for the best estimate.
I have never seen the term before (good thing my clients never knew that). But isn't that what the partial derivatives in the expansion are there to address? (If the recovery curve Q is correct)
I do not pretend to understand all the nuances, but it got me throughmany FDA submissions.
 
  • #15
hutchphd said:
isn't that what the partial derivatives in the expansion are there to address?
I should have started by asking you which formula you meant in post #8.

The formula quoted in post #1 won't do it alone because it only estimates the error for a single datapoint. Taking the largest of the these would be crazy since taking more readings would tend to increase that instead of reducing it. Taking the average is hardly better.

Bullet 2 in post #1 implies @TheMercury79 tried using the formula for standard deviation (or maybe for standard error of the mean) instead. This ignores the estimates of the individual datapoint errors, so cannot account for heteroscedasticity.
 
  • Like
Likes hutchphd
  • #16
haruspex said:
You have values ##(y_i)## with a priori variances ##\sigma_i##.
Taking each to be normally distributed about some unknown ideal value ##\hat y##, the marginal probability of the values is ##p((y_i))=(2\pi)^{-n/2}\frac 1{\Pi\sigma_i}e^{-\Sigma{\frac{(y_i-\hat y)^2}{(2\sigma_i)^2}}}##.
To maximise that we would want ##\hat y## to be ##y'=S\Sigma\frac{y_i}{\sigma_i^2}## where ##1/S=\Sigma\frac 1{\sigma_i^2}##. So that is our best estimate for ##\hat y##.
To find the mean error in that, we need to estimate ##E[(y'-\hat y)^2]##.
I got as far as the expression ##S^2\frac{(2\pi)^{-n/2}}{\Pi\sigma_i}\int..\int(\Sigma\frac{y_i-\hat y}{\sigma_i^2})^2e^{-\Sigma{\frac{(y_i-\hat y)^2}{(2\sigma_i)^2}}}d(y_i)##.
I see hints that this should simplify greatly,
It simplified alright: ##E[(y'-\hat y)^2]=S=\frac 1{\Sigma\frac 1{\sigma_i^2}}##
Unfortunately, that's not what I set out to achieve. The scatter of the actual ##y_i## should play a part.
 
Last edited:
  • Like
Likes hutchphd

FAQ: Error propagation leads to different uncertainties, which do I choose?

What is error propagation and how does it lead to different uncertainties?

Error propagation is the process of determining the uncertainties in a final measurement or result based on the uncertainties in the individual measurements or inputs. This is done using mathematical formulas that take into account the magnitude and direction of the uncertainties. As each measurement or input has its own uncertainty, the final result will also have its own uncertainty, which may be different from the individual uncertainties.

Why do different uncertainties arise from error propagation?

Different uncertainties arise from error propagation because each measurement or input has its own unique uncertainty that contributes to the overall uncertainty of the final result. Additionally, the type of mathematical operation used to combine the measurements or inputs can also affect the final uncertainty. For example, adding or subtracting values with uncertainties will result in a larger uncertainty compared to multiplying or dividing values with uncertainties.

How do I choose which uncertainty to use in my final result?

The uncertainty to use in your final result should be chosen based on the type of measurement or input that has the largest contribution to the overall uncertainty. This can be determined by calculating the percent uncertainty for each measurement or input and selecting the one with the highest value. Additionally, it is important to consider the type of mathematical operation used and how it affects the final uncertainty.

Can I use the average uncertainty for all measurements in my final result?

No, using the average uncertainty for all measurements may not accurately represent the uncertainty in the final result. This is because the average uncertainty does not take into account the magnitude and direction of each individual uncertainty, which can greatly affect the overall uncertainty. It is important to use the appropriate uncertainty for each measurement or input in order to accurately determine the final uncertainty.

How can I minimize the uncertainties in my final result?

To minimize the uncertainties in your final result, it is important to minimize the uncertainties in each individual measurement or input. This can be done by using precise and accurate instruments, taking multiple measurements, and reducing sources of error. Additionally, choosing the appropriate mathematical operation and considering the direction and magnitude of uncertainties can also help minimize the overall uncertainty in the final result.

Similar threads

Back
Top