〈1210〉 Statistical Tools for Procedure Validation
〈1210〉 Statistical Tools for Procedure Validation
Printed by: Dang Van Vu Official Date: Official as of 01-May-2018 @2023 USPC
Do Not Distribute DOI Ref: saf9m DOI: https://doi.org/10.31003/USPNF_M8646_07_01
1
1. INTRODUCTION
This chapter describes utilization of statistical approaches in procedure validation as described in Validation of Compendial
Procedures á1225ñ. For the purposes of this chapter, “procedure validation” refers to the analytical procedure qualification stage
al
of the method life cycle, following design and development and prior to testing.
Chapter á1225ñ explains that capabilities of an analytical procedure must be validated based on the intended use of the
analytical procedure. Chapter á1225ñ also describes common types of uses and suggests procedure categories (I, II, III, or IV)
based on the collection of performance parameters appropriate for these uses. Performance parameters that may need to be
established during validation include accuracy, precision, specificity, detection limit [limit of detection, (LOD)], quantitation
ci
limit, linearity, and range. In some situations (e.g., biological assay), relative accuracy takes the place of accuracy. This chapter
focuses on how to establish analytical performance characteristics of accuracy, precision, and LOD. For quantitative analytical
procedures, accuracy can only be assessed if a true or accepted reference value is available. In some cases, it will be necessary
to assess relative accuracy. In many analytical procedures, precision can be assessed even if accuracy cannot be assessed. The
section addressing LOD can be applied to limit tests in Category II.
The other analytical performance characteristics noted in á1225ñ, which include specificity, robustness, and linearity, are out
ffi
of scope for this chapter.
Because validation must provide evidence of a procedure’s fitness for use, the statistical hypothesis testing paradigm is
commonly used to conduct validation consistent with á1225ñ. Although some statistical interval examples are provided in 3.
Accuracy and Precision, these methods are not intended to represent the only approach for data analysis, nor to imply that
alternative methods are inadequate.
O
Table 1 provides terminology used to describe an analytical procedure in this chapter. The definitions for individual
determination and reportable value are in alignment with General Notices, 7.10 Interpretation of Requirements.
Material created by any physical manipulation of the laboratory sample, such as crushing or
Analytical sample grinding
Test portion The quantity (aliquot) of material taken from the analytical sample for testing
The solution resulting from chemical manipulation of the test portion such as chemical deriva-
Test solution tization of the analyte in the test portion or dissolution of the test portion
Individual determination (ID) The measured numerical value from a single unit of test solution
Reportable value Average value of readings from one or more units of a test solution
Not all analytical procedures have all stages shown in Table 1. For example, liquid laboratory samples that require no further
manipulations immediately progress to the test solution stage. Demonstration that a reportable value is fit for a particular use
is the focus of analytical validation.
Table 2 provides an example of the Table 1 terminology for a solid oral dosage form.
Analytical sample 20 tablets are removed from the laboratory sample and are crushed in a mortar and pestle
Replicate 1: 1 g of crushed powder aliquot from the Replicate 2: 1 g of crushed powder aliquot from the
Test portion analytical sample analytical sample
https://online.uspnf.com/uspnf/document/1_GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US 1/12
www.webofpharma.com
Printed on: Wed Feb 08 2023, 11:00:23 PM(EST) Status: Currently Official on 09-Feb-2023 DocId: GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US
Printed by: Dang Van Vu Official Date: Official as of 01-May-2018 @2023 USPC
Do Not Distribute DOI Ref: saf9m DOI: https://doi.org/10.31003/USPNF_M8646_07_01
2
Procedure validation is a cornerstone in the process of establishing an analytical procedure. The aim of procedure validation
is to demonstrate that the procedure, when run under standard conditions, will satisfy the requirement of being fit for use. To
maximize the likelihood of a successful validation, it is imperative that all aspects of the procedure be well understood prior to
the validation. Surprising discoveries (whether "good" or "bad") during validation should be carefully evaluated to determine
whether the procedure was adequately developed. Moreover, pre-validation work can reveal suitable approaches to reduce the
total size of the validation experiment without increasing the risk of drawing the wrong conclusion. General principles and plans
for sample preparation, general principles, experimental design, data collection, statistical evaluation, and choice of acceptance
criteria should be documented in a validation experimental protocol signed before initiation of the formal validation.
Questions considered prior to validation may include the following:
• What are the allowable ranges for operational parameters, such as temperature and time, that impact the performance of
the analytical procedure?
al
○ Robustness of these ranges can be determined using a statistical design of experiments (DOE).
• What are the ruggedness factors that impact precision?
○ Factors such as analyst, day, reagent lot, reagent supplier, and instrument that impact the precision of a test procedure
are called ruggedness factors. When ruggedness factors impact precision, reportable values within the same
ruggedness grouping (e.g., analyst) are correlated. Depending on the strength of the correlation, a statistical analysis
• Do accepted reference values or results from an established procedure exist for validation of accuracy?
○ If not, as stated in International Council for Harmonisation (ICH) Q2, accuracy may be inferred once precision,
linearity, and specificity have been established.
• How many individual determinations will compose the reportable value, and how will they be aggregated?
○ To answer this question, it is necessary to understand the contributors to the procedure variance and the ultimate
purpose of the procedure. Estimation of variance components during pre-validation provides useful information for
making this decision.
• What are appropriate validation acceptance criteria?
○ The validation succeeds when there is statistical evidence that the assay is no worse than certain pre-specified levels
for each relevant validation parameter.
○ What defines the assay as fit for use, and how does this relate to acceptance criteria?
• How large a validation experiment is necessary?
○ Validation experiments should be properly powered to ensure that there are sufficient data to conclude that the
accuracy and precision can meet pre-specified acceptance criteria. Computer simulation is a useful tool for performing
power calculations.
○ Efficiencies (both cost and statistical) can be gained if assessment of linearity, accuracy, and precision can be
combined.
On the basis of the answers to these and similar questions, one can design a suitable validation experimental protocol.
� = τ + β + � (1)
Y = a reportable value
τ = true or accepted reference value
β = systematic bias of the procedure
https://online.uspnf.com/uspnf/document/1_GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US 2/12
www.webofpharma.com
Printed on: Wed Feb 08 2023, 11:00:23 PM(EST) Status: Currently Official on 09-Feb-2023 DocId: GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US
Printed by: Dang Van Vu Official Date: Official as of 01-May-2018 @2023 USPC
Do Not Distribute DOI Ref: saf9m DOI: https://doi.org/10.31003/USPNF_M8646_07_01
3
Both τ (tau) and β (beta) are fixed statistical parameters, and E is a normal random variable with a mean of zero and standard
deviation σ (sigma). The magnitude of σ depends on the number of individual readings averaged to obtain the reportable value.
Accuracy of an analytical procedure expresses the closeness of agreement between τ and Y. Closeness is expressed as the
long-run average of (Y − τ). This long-run average is called the systematic bias and is represented with β. To estimate β, it is
necessary to know the true value, τ. Chapter á1225ñ notes that a reference standard or a well-characterized orthogonal procedure
can be used to assign the value of τ. Accuracy should be established across the required range of the procedure.
Precision of an analytical procedure is the degree of agreement among reportable values when the procedure is applied
repeatedly (possibly under different conditions) to multiple test portions of a given analytical sample. The most common
precision metric is the standard deviation σ. This is denoted in Equation 4 with the variable S. The term σ2 is called the variance.
Precision improves as σ decreases. Many commonly used statistical procedures rely on the assumption of the normal distribution,
for which σ is a natural descriptor of variability.
Change to read:
3.1 Methods for Estimating Accuracy and Precision
An example is provided to demonstrate the test procedure for lot release using statistical analysis. This example uses
high-performance liquid chromatography (HPLC). The measured drug substance (DS) is a USP compendial substance, so
information concerning τ is available (1). Three different quantities of reference standard were weighted to correspond to three
different percentages of the test concentrations: 50%, 100%, and 150%. The unit of measurement on each reportable value
is the mass fraction of DS expressed in units of mg/g and does not change as the level of concentration varies. The value of τ
al
is 1000 mg/g for all three concentrations. The computed statistics from the validation data set include the sample mean (Y),
the sample standard deviation (S), and the number of reportable values (n). Table 3 presents the n = 9 reportable values and
the computed statistics.
50 2 988.43
ffi
50 3 995.90
100 4 987.22
100 5 990.53
100 6 999.39
O
150 7 996.33
150 8 993.67
150 9 987.76
Several assumptions are made for purposes of this example, which allows analysis of the combined data set in Table 3:
1. All n = 9 reportable values are independent.
2. The standard deviation of the reportable value is constant across all three concentration levels. If this condition is not
met, data transformations may still allow combination of all the data in Table 3 (pooling). If transformations are not
successful, each concentration level must be validated for precision separately.
3. The average reportable value is equal across concentration levels. If this condition does not hold, it is necessary to employ
an analysis of variance model and validate accuracy for each concentration level separately.
The point estimator for unknown bias β is:
β = � − τ (2)
β = systematic bias
Y = sample mean
τ = true or accepted reference value
where
�
Σ �
�=1 �
�= �
(3)
https://online.uspnf.com/uspnf/document/1_GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US 3/12
www.webofpharma.com
Printed on: Wed Feb 08 2023, 11:00:23 PM(EST) Status: Currently Official on 09-Feb-2023 DocId: GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US
Printed by: Dang Van Vu Official Date: Official as of 01-May-2018 @2023 USPC
Do Not Distribute DOI Ref: saf9m DOI: https://doi.org/10.31003/USPNF_M8646_07_01
4
Y = sample mean
Yi = individual values
n = number of reportable values
� 2
Σ �� − �
�=1
�= �−1
(4)
Because point estimators have uncertainty associated with them, best practice requires calculation of a statistical confidence
interval to quantify the uncertainty. Statistical confidence intervals provide a range of plausible values for β and σ for a given
level of confidence. A 100(1 − 2α)% two-sided confidence interval for the bias β is
�
� − � ± �1 − �: � − 1 × (5)
�
al
β = 100(1 − 2α)% two-sided confidence interval of bias
t1−α:n−1 = percentile of central t-distribution with area 1 − α to the left and (n − 1) degrees of freedom
S = result found from Equation 4
n = number of reportable values
ci
For example, with ɑ = 0.05 and n = 9, t0.95:8 = 1.860 provides a 100(1 − 2 × ▲0.05▲ (ERR 1-Mar-2018))% = 90% two-sided confidence
interval for β. Using the example data in Table 3 with τ = 1000 mg/g, the 90% confidence interval on β is
�
� − � ± �1 − �: � − 1 ×
�
4.44
ffi
992.81 − 1000 ± 1.86 × (6)
9
− 9.94 to − 4.44 mg/g
β = 100(1 − 2α)% two-sided confidence interval of bias
t1 − α:n − 1 = percentile of central t-distribution with area 1 − α to the left and (n − 1) degrees of freedom
S = result found from Equation 4
O
For the standard deviation, one is concerned with only the 100(1 − α)% upper confidence bound since typically, it needs to
be shown that the standard deviation is not too large. An upper 100(1 − α)% confidence bound for σ is
�−1
�=� 2
(7)
��: � − 1
For example, if α = 0.05 and n = 9, then χ20.05:8 = 2.73. Using the data in Table 3,
�−1
�=�
�2
�: � − 1
9−1
� = 4.44 2.73 = 7.60 mg/g (8)
The confidence intervals in Equations 5 and 7 can be used to perform statistical tests against criteria included in the validation
protocol. Use of point estimates only does not provide the required scientific rigor. In particular, the two-sided confidence
interval in Equation 5 can be used to perform a two one-sided test (TOST) of statistical equivalence (2). Assume in the present
example that the accuracy requirement is validated if evidence demonstrates that the absolute value of β is NMT 15 mg/g.
Since the computed confidence interval from −9.94 to −4.44 mg/g falls entirely within the range from −15 to +15 mg/g, the
bias criterion is satisfied. Most typically, the TOST employs a type I error rate of α = 0.05. This error rate represents the maximum
risk of declaring that the acceptance criterion is satisfied, when in truth it is not satisfied. Thus, with α = 0.05, the two-sided
confidence interval in Equation 5 is 100(1 − 2α)% = 90%.
https://online.uspnf.com/uspnf/document/1_GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US 4/12
www.webofpharma.com
Printed on: Wed Feb 08 2023, 11:00:23 PM(EST) Status: Currently Official on 09-Feb-2023 DocId: GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US
Printed by: Dang Van Vu Official Date: Official as of 01-May-2018 @2023 USPC
Do Not Distribute DOI Ref: saf9m DOI: https://doi.org/10.31003/USPNF_M8646_07_01
5
The upper bound in Equation 7 is used to validate precision. Suppose the pre-defined acceptance criterion for precision
requires σ to be <20 mg/g. The computed upper bound of 7.60 mg/g in Equation 8 represents the largest value we expect for
σ with 95% confidence. Since 7.60 mg/g is <20 mg/g, precision has been successfully validated with a confidence of 95%.
Change to read:
3.2 Combined Validation of Accuracy and Precision
When assessing whether an analytical procedure is fit for its intended purpose, it is often useful to consider the combined
impact of bias and precision. The degree to which β impacts the usefulness of an analytical procedure depends in part on σ.
That is, a procedure with a relatively small value of σ can accommodate a relatively greater value of β than a procedure with a
greater value of σ. For this reason, it is useful to establish a single criterion that can be used to simultaneously validate both
accuracy and precision. One such criterion is proposed in a series of articles by Hubert et al. (3–5) and seeks to ensure that
Pr − � < � − � < � ≥ �, or
al
Equation 9 has a dual interpretation. It can be interpreted as either (i) the probability that the next reportable value falls in
the range from (−λ + τ) to (λ + τ) is ≥P, or (ii) the proportion of all future reportable values falling between (−λ + τ) and (λ + τ)
is ≥P. Accordingly, two statistical intervals have been proposed to demonstrate that Equation 9 is true:
1. A prediction interval (also referred to as an expectation tolerance interval) is used to demonstrate (i).
ci
2. A tolerance interval (also referred to as a content tolerance interval) is used to demonstrate (ii).
Hahn and Meeker (6) note that the prediction interval is also referred to as an expectation tolerance interval and that the
tolerance interval is also referred to as a content tolerance interval. Because the inference associated with the tolerance interval
concerns a larger set of values, the tolerance interval is always wider than the prediction interval. Selection of an interval will
depend on the desire to validate ▲▲ (ERR 1-May-2018) either (i) or (ii) and a company’s risk profile.
ffi
Either interval can be used in the following manner to evaluate accuracy and precision simultaneously through Equation 9.
1. Compute the appropriate statistical interval using Equation 10 for the prediction interval and Equation 11 for the tolerance
interval.
2. If the computed interval falls completely in the range from (−λ + τ) to (λ + τ), the criterion in Equation 9 is satisfied, and
the procedure is validated for both accuracy and precision.
The prediction interval used to validate Equation 9 is
O
1
� ± � 1 + � /2: � − 1 × � 1 + � (10)
Y = sample mean
t (1 + P)/2:n − 1 = percentile of a central t-distribution with area (1 + P)/2 to the left and (n − 1) degrees of freedom
S = result found from Equation 4
�± �× �
�21 + � /2 × � − 1 1
�= 2
× 1 + � (11)
��: � − 1
Y = sample mean
K = result found from Equation 11
S = result found from Equation 4
Z2(1 + P)/2 = ▲the square of the standard normal▲ (ERR 1-May-2018) percentile with area (1 + P)/2 to the left
n = number of reportable values
χ2α:n − 1 = a chi-squared percentile with area α to the left and (n − 1) degrees of freedom
The formula for K is based on an approximation by Howe (7), although exact tabled values can be found in several sources.
The approximation works well in practical situations if exact values are not available.
For the data in Table 3 with P = 0.90, the interval for Equation 10 is computed as
1
� ± � 1 + � /2, � − 1 × � 1 + �
1
992.81 ± 1.86 × 4.44 1 + 9 (12)
https://online.uspnf.com/uspnf/document/1_GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US 5/12
www.webofpharma.com
Printed on: Wed Feb 08 2023, 11:00:23 PM(EST) Status: Currently Official on 09-Feb-2023 DocId: GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US
Printed by: Dang Van Vu Official Date: Official as of 01-May-2018 @2023 USPC
Do Not Distribute DOI Ref: saf9m DOI: https://doi.org/10.31003/USPNF_M8646_07_01
6
�21 + � /2 × � − 1 1
�= × 1+ �
�2
�: � − 1
2
1.64 × 9 − 1 1
�= 3.49
× 1 + 9 = 2.63
�± �× �
992.81 ± 2.63 × 4.44
981.2 to 1004.5 mg/g (13)
The exact value for K is 2.637, and the approximation is seen to work quite well. As predicted earlier, the interval for Equation
13 is wider than the interval for Equation 12.
Suppose the criterion for Equation 9 is designed to ensure that the difference between Y and τ is <2% of τ with a probability
NLT P = 0.90. Thus,
al
Since both Equations 12 and 13 fall in the range from 980 to 1020 mg/g, the procedure is validated using either interval.
It is also possible to estimate Pr(−λ < Y − τ < λ) in Equation 9 directly using either the confidence interval described by Mee
(8) or a Bayesian approach. The validation criterion is thus satisfied if this estimated probability exceeds P. A Bayesian tolerance
interval is provided in Wolfinger (9) and can be computed using the statistical software package WinBUGS (10,11). Bayesian
analyses can be challenging, and the aid of an experienced statistician is recommended.
ci
4. LIMITS OF DETECTION AND QUANTITATION
The LOD and limit of quantitation (LOQ) are two related quantities determined in the validation of Category II procedures
of á1225ñ. These are procedures for the determination of impurities or degradation products in DS and finished pharmaceutical
ffi
products. Only one is needed for each use: LOQ for quantitative tests and LOD for qualitative limit tests. These limits are also
known under other names, including detection limit (DL) for LOD and lower limit of quantitation (LLOQ) for LOQ.
The following definitions are consistent with á1225ñ and ICH Q2:
• The LOD is the lowest amount of analyte in a sample that can be detected, but not necessarily quantitated, under the
stated experimental conditions.
• The LOQ is the lowest amount of analyte in a sample that can be determined with acceptable precision and accuracy
O
�� = � + �1 − ��� (15)
https://online.uspnf.com/uspnf/document/1_GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US 6/12
www.webofpharma.com
Printed on: Wed Feb 08 2023, 11:00:23 PM(EST) Status: Currently Official on 09-Feb-2023 DocId: GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US
Printed by: Dang Van Vu Official Date: Official as of 01-May-2018 @2023 USPC
Do Not Distribute DOI Ref: saf9m DOI: https://doi.org/10.31003/USPNF_M8646_07_01
7
al
ci
Figure 1. Determination of RC and RD.
For example, if α = 0.05, 1 − α = 0.95 and Z0.95 = 1.645. This determination depends on the distribution of values obtained
when analyzing blanks. The LOD in the signal space (RD) is defined as that value, which if true, is such that RC is exceeded with
ffi
probability 1 − β. That is,
�� = �� + �1 − ��� (16)
�� = � + �1 − � + �1 − � �� (17)
Note that this definition allows for two values to be selected by the laboratory: α and β (which need not be equal). The α
represents the type I or false-positive error rate, and β represents the type II or false-negative error rate. In Figure 1, RC and RD
are illustrated with α = β = 0.05 for normally distributed data so that Z1 − α = Z1 − β =1.645. Although the values of α and β need
not be equal, this choice leads to a common rule for RD, namely B + 3.3σE (3.3 ≅ 2 × 1.645).
The LOD on the concentration scale is then found by converting the value in the signal scale, RD, to one in the concentration
scale, LOD, as shown in Figure 2.
https://online.uspnf.com/uspnf/document/1_GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US 7/12
www.webofpharma.com
Printed on: Wed Feb 08 2023, 11:00:23 PM(EST) Status: Currently Official on 09-Feb-2023 DocId: GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US
Printed by: Dang Van Vu Official Date: Official as of 01-May-2018 @2023 USPC
Do Not Distribute DOI Ref: saf9m DOI: https://doi.org/10.31003/USPNF_M8646_07_01
8
al
ci
ffi
Figure 2. Determination of LOD from RD.
O
This step requires that the signal (R) versus concentration (X) line, R = B + mX, as well as σE, be known exactly. The formulation
provided in this section assumes the regression measurements are independent.
The LOD on the concentration scale is then calculated as
�� − � � + � δ
1−α 1−β �
LOD = �
= �
(18)
As a statistical procedure, the LOD definition in Equation 18 is unsatisfactory for two reasons. First, since σE is generally
unknown, it must be determined how to best estimate this parameter. This is complicated because σE is typically concentration
dependent. Two common estimates are (i) the standard deviation of the blank responses and (ii) the standard deviation obtained
from deviations about the regression line of signal on concentration. The choice needs to be the value that best represents σE
in the neighborhood of the LOD. Laboratories will often pick a worst-case value for σE. If the LOD is still suitable for its intended
use, the laboratories are protected against understating the LOD. Understatement of the LOD results in an inflated type II error
rate (β) and a deflated type I error rate (α).
The second statistical concern with Equation 18 is how to incorporate uncertainty due to the fact that the exact slope of the
regression line of signal on concentration is unknown. Because the regression line is estimated, the definition of RD in Equation
17 is itself an estimate. This is corrected by using a statistical prediction interval that takes into account the uncertainty in the
estimated line as well as the variability associated with a future observation. The expanded formula for the critical value, RC,
originally defined in Equation 15 that accounts for this uncertainty is
https://online.uspnf.com/uspnf/document/1_GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US 8/12
www.webofpharma.com
Printed on: Wed Feb 08 2023, 11:00:23 PM(EST) Status: Currently Official on 09-Feb-2023 DocId: GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US
Printed by: Dang Van Vu Official Date: Official as of 01-May-2018 @2023 USPC
Do Not Distribute DOI Ref: saf9m DOI: https://doi.org/10.31003/USPNF_M8646_07_01
9
1 �2
�� = � + �1 − �: � − 2 × � 1 + � + � 2
Σ � − �
�=1 �
� 2
Σ � − � − ���
�=1 �
�= � − 2
(19)
Equation 19 differs from Equation 15 because the t-distribution is used instead of the normal distribution for the multiplier, and
two additional terms appear in the square root to capture the uncertainty of the regression line.
A second equation for RC answers the question, “Above which concentration can we be confident that we will obtain signals
that are distinguishable from background?” This question is answered by using the lower 100(1 − β)% prediction bound of the
al
calibration curve as shown in Figure 3.
ci
ffi
O
Figure 3 is similar to Figure 2, but uses two dashed curves instead of the solid calibration line. Here
https://online.uspnf.com/uspnf/document/1_GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US 9/12
www.webofpharma.com
Printed on: Wed Feb 08 2023, 11:00:23 PM(EST) Status: Currently Official on 09-Feb-2023 DocId: GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US
Printed by: Dang Van Vu Official Date: Official as of 01-May-2018 @2023 USPC
Do Not Distribute DOI Ref: saf9m DOI: https://doi.org/10.31003/USPNF_M8646_07_01
10
1 ��� − � 2
�� = � + LOD × � − �1 − β: � − 2 × � 1 + � + �
(20)
2
Σ � − �
�=1 �
After equating Equation 19 and Equation 20, and cancelling the B terms,
1 �2 1 ��� − � 2
�1 − α: � − 2 × � 1 + � + �
= ��� × � − �1 − β: � − 2 × � 1 + � + � (21)
2 2
Σ �� − � Σ � − �
�=1 �=1 �
al
t1 − α:n − 2 = percentile of a central t-distribution with area 1 − α to the left and (n − 2) degrees of freedom
S = standard error of regression line
X2 = average concentration value squared
LOD = limit of detection
m = slope
t1 − β:n − 2 = percentile of a central t-distribution with area 1 − β to the left and (n − 2) degrees of freedom
n
Xi
X
ci
= number of observations used in the regression analysis
= concentration value used in determining the line
= average concentration value
Equation 21 is a quadratic equation for LOD that can be solved exactly or by using iterative search tools available in
ffi
spreadsheets. A slightly conservative (overly large) approximation for LOD that does not require a quadratic solution is obtained
by assuming that LOD is negligible compared to X [i.e., (LOD − X)2 is replaced with X2]. The resulting equation under this
simplification is
�2
O
� 1
��� = �1 − α: � − 2 + �1 − β: � − 2 × � 1 + � + �
(22)
2
Σ �� − �
�=1
which is similar in form to Equation 18. Equations 18 and 22 both allow the two error probabilities, α and β, to differ. Often they
are both taken as equal to 0.05.
The data in Table 4 are used to demonstrate calculation of the LOD.
0.02 0.00602
0.05 0.01547
0.10 0.03078
0.15 0.04576
0.25 0.07592
https://online.uspnf.com/uspnf/document/1_GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US 10/12
www.webofpharma.com
Printed on: Wed Feb 08 2023, 11:00:23 PM(EST) Status: Currently Official on 09-Feb-2023 DocId: GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US
Printed by: Dang Van Vu Official Date: Official as of 01-May-2018 @2023 USPC
Do Not Distribute DOI Ref: saf9m DOI: https://doi.org/10.31003/USPNF_M8646_07_01
11
Fitting the linear regression to these data yields the regression line:
so that m = 0.3032 and B = 0.000235. Values needed to compute LOD shown in Equation 22 with α = β = 0.05 are provided in
Table 5.
m (slope) 0.3032
S 0.00019
� 2
Σ � − �
�=1 �
0.0419
al
The value of LOD computed from Equation 22 is
� 1 �2
��� = �1 − α: � − 2 + �1 − β: � − 2 × � 1 + � + 2
Σ�
ci 0.00019
� − �
�=1 �
m = slope
n = number of observations used in the regression analysis
Xi = concentration value used in determining the line
https://online.uspnf.com/uspnf/document/1_GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US 11/12
www.webofpharma.com
Printed on: Wed Feb 08 2023, 11:00:23 PM(EST) Status: Currently Official on 09-Feb-2023 DocId: GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US
Printed by: Dang Van Vu Official Date: Official as of 01-May-2018 @2023 USPC
Do Not Distribute DOI Ref: saf9m DOI: https://doi.org/10.31003/USPNF_M8646_07_01
12
5. CONCLUDING REMARKS
This chapter presented some simple statistical methods that can be used in procedure validation as described in á1225ñ.
These methods may not be applied in all situations, and other statistical approaches, both more and less sophisticated, may be
appropriate for any particular situation.
Re-evaluation of a procedure should be considered whenever use of the procedure changes. For example, if a new product
strength is introduced, the procedure is transferred to a new lab, samples are to be tested following a new type of stress test,
or specifications change, a re-validation is most likely appropriate. In some situations, a re-assessment of existing data to revised
acceptance limits is sufficient.
Finally, although not part of procedure validation, it is recommended that some type of statistical process control be used
to monitor the performance of the procedure. Such a process provides early warning of “drift” in the analytical performance
characteristics of accuracy and precision. Such changes in performance are not uncommon, and often occur as a result of worn
equipment, change of routines, or aging reagents.
REFERENCES
1. Weitzel MLJ. The estimation and use of measurement uncertainty for a drug substance test procedure validated according
to USP á1225ñ. Accred Qual Assur. 2012;17:139–146.
2. Schuirmann DJ. A comparison of the two one-sided tests procedure and the power approach for assessing the equivalence
of average bioavailability. J Biopharmacokinet Biopharmaceut. 1987;15:657–680.
3. Hubert P, Nguyen-Huu JJ, Boulanger B, Chapuzet E, Chiap P, Cohen N, et al. Harmonization of strategies for the validation
al
of quantitative analytical procedures. A SFSTP proposal—part I. J Pharm Biomed Anal. 2004;36(3):579–586.
4. Hubert P, Nguyen-Huu JJ, Boulanger B, Chapuzet E, Chiap P, Cohen N, et al. Harmonization of strategies for the validation
of quantitative analytical procedures. A SFSTP proposal—part II. J Pharm Biomed Anal. 2007;45(1):70–81.
5. Hubert P, Nguyen-Huu JJ, Boulanger B, Chapuzet E, Chiap P, Cohen N, et al. Harmonization of strategies for the validation
of quantitative analytical procedures. A SFSTP proposal—part III. J Pharm Biomed Anal. 2007;45(1):82–96.
ci
6. Hahn GJ, Meeker WQ. Statistical Intervals: A Guide for Practitioners. New York: John Wiley & Sons; 1991.
7. Howe WG. Two-sided tolerance limits for normal populations—Some improvements. J Am Stat Assoc. 1969;64(326):610–
620.
8. Mee RW. Estimation of the percentage of a normal distribution lying outside a specified interval. Commun Stat Theory
ffi
Methods. 1988;17(5):1465–1479.
9. Wolfinger RD. Tolerance intervals for variance component models using Bayesian simulation. J Qual Technol. 1998;30(1):
18–32.
10. Ntzoufras I. Bayesian Modeling Using WinBUGS. New York: John Wiley & Sons; 2009.
11. Spiegelhalter D, Thomas A, Best N, Gilks W. BUGS 0.5 Examples: Volume 1 (version i). Cambridge, UK: MRC Biostatistics
Unit; 1996. http://users.aims.ac.za/~mackay/BUGS/Manual05/Examples1/bugs.html. Accessed 12 Jul 2016.
O
▲ (USP41)
https://online.uspnf.com/uspnf/document/1_GUID-13ED4BEB-4086-43B5-A7D7-994A02AF25C8_7_en-US 12/12
www.webofpharma.com