Download Limitations in Analytical Accuracy, Part II: Theories to Describe the

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Degrees of freedom (statistics) wikipedia , lookup

History of statistics wikipedia , lookup

Psychometrics wikipedia , lookup

Misuse of statistics wikipedia , lookup

Student's t-test wikipedia , lookup

Transcript
20 Spectroscopy 22(2)
w w w. s p e c t r o s c o p y o n l i n e . c o m
February 2007
Chemometrics In Spectroscopy
Limitations in Analytical
Accuracy, Part II: Theories to
Describe the Limits in Analytical
Accuracy and Comparing Test
Results for Analytical Uncertainty
This column continues from the previous discussion on the limits of analytical accuracy
and uncertainty. It introduces several tools used for evaluating and comparing analytical
results.
Jerome Workman, Jr. and Howard Mark
Limits in Analytical Accuracy
Y
ou might recall from our previous column (1) how
Horwitz throws down the gauntlet to analytical scientists, stating that a general equation can be formulated for the representation of analytical precision based
upon analyte concentration (2). He states this as follows:
CV(%) 2(10.5logC)
where C is the mass fraction as concentration expressed
in powers of 10 (for example, 0.1% analyte is equal to C 103).
A paper published by Hall and Selinger (3) points out an
empirical formula relating the concentration (c) to the coefficient of variation (CV), also known as the precision ().
They derive the origin of the “trumpet curve” using a binomial distribution explanation. Their final derived relationship becomes
CV c 0.15
50
They further simplify the Horwitz trumpet relationship in
two forms as follows:
CV(%) 0.006c0.5
and
0.006c0.5
They then derive their own binomial model relationships
using Horwitz’s data with variable apparent sample size.
CV(%) 0.02c0.15
and
0.02c0.85
22 Spectroscopy 22(2)
Both sets of relationships depict relative error as inversely proportional to
analyte concentration.
In yet a more detailed incursion
into this subject, Rocke and Lorenzato
(4) describe two disparate conditions
in analytical error: concentrations
near zero and macro level concentrations, say greater than 0.5% for argument’s sake. They propose that analytical error comprises two types,
additive and multiplicative. So their
derived model for this condition is
x e
where x is the measured concentration; is the true analyte concentration; and is a normally distributed
analytical error with mean 0 and standard deviation . It should be noted
that represents the multiplicative or
proportional error with concentration
and represents the additive error
demonstrated at small concentrations.
Using this approach, the critical
level at which the CV is a specific
value can be found by solving for x
using the following relationship:
(CVx)2 (
x)2 ()2
where x is the measured analyte concentration as the practical quantitation level (PQL used by the U.S. Environmental Protection Agency [EPA]).
This relationship is simplified to
x
w w w. s p e c t r o s c o p y o n l i n e . c o m
February 2007
1
2
(CV 2 ση )
where CV is the critical level at which
the coefficient of variation is a preselected value to be achieved using a specific analytical method, and is the
standard deviation of the multiplicative
or measurement error of the method.
For example, if the desired CV is 0.3
and is 0.1, then the PQL or x is computed as 3.54. This is the lowest analyte
concentration that can be determined
given the parameters used.
The authors describe the earlier
model as a linear exponential calibration curve as
y e
where y is the observed measurement
data. This model approximates a consistent or constant standard deviation
model at low concentrations and approximates a constant CV model for
high concentrations. In this model, the
multiplicative error varies as e
.
certainty range of 5.014 to 5.714 with
an uncertainty interval of 0.7. Therefore, if we have a relatively unbiased
analytical method, there is a 95%
probability that our true analyte value
lies between these upper and lower
concentration limits.
Detection Limit for
Concentrations Near Zero
Comparison Test for a Single Set
of Measurements Versus a True
Analytical Result
Finally, detection limit (D) is estimated using
3σ
D r
where is the standard deviation of
the measurement error measured at
low (near zero) concentration, and r is
the number of replicate measurements
made.
Uncertainty in an Analytical
Measurement
By making replicate analytical measurements, one can estimate the certainty of
the analyte concentration using a computation of the confidence limits. As an
example, given five replicate measurement results as: 5.30%, 5.44%, 5.78%,
5.00%, and 5.30%, the precision (or
standard deviation) is computed using
the following equation:
r
s
(x − x )
(r − 1)
i 1
2
i
where s represents the precision, means summation of all the (xi –x)2
values, xi is an individual replicate analytical result, –x is the mean of the
replicate results, and r is the total
number of replicates included in the
group (this is often represented as n).
For the previous set of replicates, s 0.282. The degrees of freedom are indicated by r 1 4. If we want to
calculate the 95% confidence level, we
note that the t-value is 2.776. So the
uncertainty (U) of our measurement
result is calculated as
U x t •
s
r
The example case results in an un-
Let’s start this discussion by assuming
we have a known analytical value by
artificially creating a standard sample
using impossibly precise weighing and
mixing methods so that the true analytical value is 5.2% analyte. We make
one measurement and obtain a value
of 5.7%. Then we refer to errors using
statistical terms as follows:
Measured value: 5.7%
“True” value: 5.2%
Absolute error: Measured value True value 0.5%
Relative percent error: 0.5/5.2 100
9.6%
Then we recalibrate our instrumentation and obtain the results: 5.10, 5.20,
5.30, 5.10, and 5.00. Thus, our mean
value (x–) is 5.14.
Our precision as the standard deviation (s) of these five replicate measurements is calculated as 0.114 with n 1
4 degrees of freedom. The t-value
from the t table, 0.95, degrees of
freedom as 4, is 2.776.
To determine if a specific test result
is significantly different from the true
or mean value, we use the test statistic
(Te):
Te x −
• n
s
For this example, Te 1.177. We note
that there is no significant difference
in the measured value versus the expected or true value if Te t-value.
And there is a significant difference
between the set of measured values
and the true value if Te t-value. We
must then conclude here that there is
no difference between the measured
set of values and the true value, as
1.177 2.776.
24 Spectroscopy 22(2)
Comparison Test for Two Sets of
Measurements
If we take two sets of five measurements using two calibrated instruments and the mean results are x–1 5.14 and x–2 5.16, we would like to
know if the two sets of results are statistically identical. So we calculate the
standard deviation for both sets and
find s1 0.114 and s2 0.193. The
pooled standard deviation s1,2 0.079. The degrees of freedom in this
case is n1 1 equals 5 1 4. The tvalue at 0.95, d.f. 4, is 2.776.
To determine if one set of measurements is significantly different from
the other set of measurements, we use
the test statistic (Te):
Te1, 2 w w w. s p e c t r o s c o p y o n l i n e . c o m
February 2007
x 1 x 2
1
s•
n1 n 2
For this example, Te1,2 0.398. So if
there is no significant difference in the
sets of measured values, we would expect Te t-value, because 0.398 2.776. And if there is a significant difference between the sets of measured
values, we expect Te t-value. We
must conclude here that there is no
difference between the sets of measured values.
Calculating the Number of
Measurements Required to
Establish a Mean Value (or
Analytical Result) with a
Prescribed Uncertainty (Accuracy)
If error is random and follows probabilistic (normally distributed) variance
phenomena, we must be able to make
additional measurements to reduce
the measurement noise or variability.
This is certainly true in the real world
to some extent. Most of us with some
basic statistical training will recall the
concept of calculating the number of
measurements required to establish a
mean value (or analytical result) with
a prescribed accuracy. For this calculation, one would designate the allowable error (e), and a probability (or
risk) that a measured value (m) would
be different by an amount (d).
We begin this estimate by computing the standard deviation of measurements. This is determined by first calculating the mean, then taking the
difference of each control result from
the mean, squaring that difference, dividing by n 1, then taking the
square root. All these operations are
included in the equation:
n
s
(x x )
(n 1)
i 1
2
i
where s represents the standard deviation; means summation of all the (xi
–x)2 values; xi is an individual control
result; –x is the mean of the control results; and n is the total number of control results included in the group.
If we were to follow a cookbook approach for computing the various parameters, we would proceed as follows:
● Compute an estimate of (s) for the
method (see previous);
● Choose the allowable margin of
error (d);
● Choose the probability level as
alpha (), as the risk that our measurement value (m) will be off by
more than d;
● Determine the appropriate t-value
for t1 /2 for n 1 degrees of freedom.
● Finally the formula for n (the number of discrete measurements
required) for a given uncertainty is
as follows:
⎛ t 2 •s 2 ⎞
n ⎜ 2 ⎟ 1
⎝ d ⎠
Problem Example: We want to learn
the average value for the quantity of
toluene in a test sample for a set of hydrocarbon mixtures.
s 1, 0.95, d 0.1. For this
problem, t1 /2 1.96 (from t table),
and thus n is computed as follows:
⎛ 1.96 2 •12 ⎞
n ⎜⎜
2
⎟⎟ 1 385
⎝ 0 .1
⎠
So if we take 385 measurements, we
conclude with a 95% confidence that
the true analyte value (mean value)
will be between the average of the 385
results and 0.1, or –x 0.1.
The Q-Test for Outliers (5–7)
We make five replicate measurements
using an analytical method to calculate basic statistics regarding the
method. Then we want to determine if
a seemingly aberrant single result is
indeed a statistical outlier. The five
replicate measurements are: 5.30%,
5.44%, 5.78%, 5.00%, and 5.30%. The
result we are concerned with is 6.0%.
Is this result an outlier? To find out, we
first calculate the absolute values of
the individual deviations, as in Table I.
Thus the minimum deviation
(DMin) is 0.22; the maximum deviation 1.00; and the deviation range (R)
is 1.00 0.22 0.78. We then calculate the Q-test value as Qn using:
Qn D Min
R
This results in the Qn of 0.22/0.78 0.28 for n 5.
Using the Q-value table (90% confidence level as Table II), we note that if
Qn Q-value, then the measurement
is not an outlier. Conversely, if Qn Q-value, then the measurement is an
outlier.
So because 0.28 0.642, this test
value is not considered an outlier.
Summation of Variance from
Several Data Sets
We sum the variance from several separate sets of data by computing the
variance of each set of measurements;
this is determined by first calculating
the mean for each set, then taking the
difference of each result from the
mean, squaring that difference, and dividing by r 1 where r is the number
of replicates in each individual data
Table I: Absolute values of individual
deviations
Compute Deviation
5.30–6.00
5.44–6.00
5.78–6.00
5.00–6.00
5.30–6.00
Absolute Deviation
0.70
0.56
0.22
1.00
0.70
26 Spectroscopy 22(2)
w w w. s p e c t r o s c o p y o n l i n e . c o m
February 2007
Table II: Q-value table (at different confidence levels)
n:
3
Q(90%): 0.941
Q(95%): 0.970
Q(99%): 0.994
4
5
6
7
8
9
10
0.765
0.829
0.926
0.642
0.710
0.821
0.560
0.625
0.740
0.507
0.568
0.680
0.468
0.526
0.634
0.437
0.493
0.598
0.412
0.466
0.568
set. All these operations are included
in the following equation:
r
s
2
(x x )
(r 1)
i 1
2
i
where s2 represents the variance for
each set; means summation of all
the (xi –x)2 values; xi is an individual
result; –x is the mean of each set of results; and r is the total number of results included in each set.
The pooled variance (s2p) is given as
s 2 s 22 ... s 2k
s 2p 1
k
where s2k represents the variance for
each data set, and k is the total number of data sets included in the pooled
group.
The pooled standard deviation p is
given as:
σ p s 2p
References
(1) J. Workman and H. Mark, Spectroscopy 21(9), 18–24 (2006).
(2) W. Horwitz, Anal. Chem. 54(1),
67A–76A (1982).
(3) P. Hall and B. Selinger, Anal. Chem.
61, 1465–1466 (1989).
(4) D. Rocke and S. Lorenzato, Technometrics 37(2), 176–184 (1995).
(5) J.C. Miller and J.N. Miller, Statistics for
Analytical Chemistry (second edition)
(Ellis Horwood, Upper Saddle River,
New Jersey, 1992), pp. 63–64.
(6) W.J. Dixon and F.J. Massey, Jr., Introduction to Statistical Analysis (fourth
edition), W.J. Dixon, Ed. (McGraw-Hill,
New York, 1983), pp. 377, 548.
(7) D.B. Rohrabacher, Anal. Chem. 63,
139 (1991).
Circle 44
Erratum
An astute reader, Charles R. Lytle,
PhD, pointed out a correction and
a clarification in our previous column (1). The Correction: Equation
1 is missing the denominator (r).
The Clarification: For Equations 1
and 2, r represents the number of
replicates (subsamples) for n as the
sample number; subscripts i and j
are indices for subsamples. The narrative is referring to multiple analyses of the same sample, better referred to as subsamples of the same
analytical sample.
(1) Spectroscopy 21(9), 18–24
(2006).
Jerome Workman,
Jr. serves on the Editorial
Advisory Board of
Spectroscopy and is director of research and technology for the Molecular
Spectroscopy &
Microanalysis division of Thermo Fisher
Scientific Inc. He can be reached by e-mail
at: [email protected].
Howard Mark
serves on the Editorial
Advisory Board of
Spectroscopy and runs a
consulting service, Mark
Electronics (Suffern, NY).
He can be reached via email at:
[email protected].