Abstract

Objectives

We examined the false acceptance rate (FAR) and false rejection rate (FRR) of varying precision verification experimental designs.

Methods

Analysis of variance was applied to derive the subcomponents of imprecision (ie, repeatability, between-run, between-day imprecision) for complex matrix experimental designs (day × run × replicate; day × run). For simple nonmatrix designs (1 day × multiple replicates or multiday × 1 replicate), ordinary standard deviations were calculated. The FAR and FRR in these different scenarios were estimated.

Results

The FRR increased as more samples were included in the precision experiment. The application of an upper verification limit, which seeks to cap FRR at 5% for multiple experiments, significantly increased the FAR. The FRR decreases as the observed imprecision increases relative to the claimed imprecision and when a greater number of days, runs, or replicates are included in the verification design. Increasing the number or days, runs, or replicates also reduces the FAR for between-day imprecision and repeatability.

Conclusions

Design of verification experiments should incorporate the local availability of resources and analytical expertise. The largest imprecision component should be targeted with a greater number of measurements. Consideration of both FAR and FRR should be given when committing a platform into service.

Key Points

• Design of precision verification experiments should incorporate the local availability of resources and analytical expertise.

• The largest imprecision component should be targeted with a greater number of measurements.

• Consideration of both false acceptance rate and false rejection rate should be given when committing a platform into service.

Precision describes the reproducibility characteristic of an analytical system. Simplistically, it describes how tightly quantitative measurements cluster together when testing the same material repeatedly over time. Precision can be divided into several components, including repeatability (ie, within-run), between-run, between-day, and within-laboratory (ie, total) imprecision.1 It is often expressed as a standard deviation (σ) in the measurand’s units or coefficient of variation (CV) generally as a percentage. The precision profile of an analytical system is the analytical σ or CV at different measurand concentrations and in most cases provided by the manufacturer. Laboratories are required to verify the precision claimed by the manufacturer before committing the system to clinical service.2,3 It is critically important to ensure that the precision of an analytical system conforms to the manufacturer’s claim, as increased imprecision can lead to an increase risk of erroneous laboratory results, subsequently affcting clinical interpretation and patient management.4-6

While guidelines exist for precision verification, such as those from the Clinical and Laboratory Standards Institute (CLSI; EP15-A3),1 these are variably followed.7 Variations to these guidelines include the number of days, number of runs per day, and number of replicates per run. Factors influencing the choice of experimental design include financial, operational, and analytical expertise. In general, the greater the number of tests performed, the more resource intensive and complex the exercise becomes. As complexity in the experimental design increases, the greater the requirement for advanced statistical tools for data analysis becomes. For example, an experimental design that considers days, runs per day, and replicates per run requires nested analysis of variance (ANOVA) to derive the corresponding variability of each component.1

The CLSI guideline EP15 for end-user laboratories recommends a 5 × 5 experimental design, which entails measuring the same material five times a day over 5 days.1 In addition, EP15 introduces an upper verification limit (UVL) factor that seeks to keep false rejection rates within 5%. A false rejection is the scenario of failing the precision verification excessive by chance alone, when the system is in fact performing within the manufacturer’s precision claim. The main rationale for introducing the UVL is to avoid the operational costs associated with unnecessarily repeating the verification exercise.

However, the CLSI EP15 does not provide information on the related false acceptance, which is the case of having a system passing the precision verification exercise by chance, when in actual fact the precision performance is worse than the stated manufacturer’s claim. From a patient harm-minimization perspective, it is more important to ensure that the false acceptance rate is minimized. A laboratory should give balanced consideration to the rate of both false rejections and false acceptances, to ensure delivery of safe clinical care within available resource constraints. To do so, the laboratory needs to have information on both aspects of false acceptance and rejection rates to make an informed decision on the experimental design and assist in interpretation of the results generated.

The objective of this report is to evaluate the effect of precision verification experimental design on the false acceptance and false rejection rates for instrument/assay verification prior to implementation of an analytical system into clinical practice.

Materials and Methods

Types of Experimental Designs

The precision verification study can be broadly divided into matrix designs, in which several measurements are made each day, in comparison to simplistic designs, in which only one measurement is made daily for several days (ie, days × 1 replicate) or several measurements are made for only 1 day (replicates × 1 day). For the complex matrix designs, the daily measurements can be performed across several runs and several days (ie, days × runs × replicates) or in a single run over several days (ie, days × replicates). Formal definitions of terminology related to precision verification can be found in the Supplemental Material (all supplemental materials can be found at American Journal of Clinical Pathology online).

Day × Replicate Matrix Design

Under this experimental design, a fixed number of replicates are measured consecutively in a single run each day and repeated over multiple days. For day × replicate designs, it is necessary to employ one-way ANOVA analysis to derive the components of imprecision.

For clinical laboratory assay verification, experiments are conducted to determine if the observed standard deviations for within-laboratory imprecision and repeatability imprecision, denoted as sday and srep respectively, are within the claimed specifications.

When the experimental design is balanced (ie, no missing data), the mathematical model for measurements obtained from the precision verification for a day × replicate study is as follows:

where Yij = observed measurement on day i and replicate j; μ = mean, estimated as the average of all readings; Di = error due to day-to-day variation on day i; and  εj(i) = within-run error for replicate j on day i.

The results obtained from the verification experiments can be summarized in a one-way ANOVA table:

Source of VariationSum of SquaresDegrees of FreedomMean SquareExpected Mean Square
Between-run (day-to-day)SS1dfday= nday1MS1nrepσday2+σrep2
Within-run (repeatability)SS2dfrep= N ndayMS2σrep2
TotalSStotalN1Total
Source of VariationSum of SquaresDegrees of FreedomMean SquareExpected Mean Square
Between-run (day-to-day)SS1dfday= nday1MS1nrepσday2+σrep2
Within-run (repeatability)SS2dfrep= N ndayMS2σrep2
TotalSStotalN1Total
Source of VariationSum of SquaresDegrees of FreedomMean SquareExpected Mean Square
Between-run (day-to-day)SS1dfday= nday1MS1nrepσday2+σrep2
Within-run (repeatability)SS2dfrep= N ndayMS2σrep2
TotalSStotalN1Total
Source of VariationSum of SquaresDegrees of FreedomMean SquareExpected Mean Square
Between-run (day-to-day)SS1dfday= nday1MS1nrepσday2+σrep2
Within-run (repeatability)SS2dfrep= N ndayMS2σrep2
TotalSStotalN1Total

where nday = number of days (assuming one run per day), nrep = number of repetitions per day, N=nrepnday, σday2 = day-to-day variance, and σrep2 = repeatability variance.

Here, σday2 refers to the “pure” between-day variance, that is, between-day variance that has been corrected for the contribution of within-day variance. If the matrix study is balanced, the degrees of freedom for repeatability, dfrep=N nday, can be obtained directly from the ANOVA table.

To determine a variance that involves a linear combination of two or more independent mean squares, the complex estimate of variance will be adopted.8 Instead of using the degrees of freedom in the ANOVA table, the effective degrees of freedom for within-laboratory imprecision are derived based on the Satterthwaite formulation8,9:

Similarly, the effective degrees of freedom for a between-day variance can be derived as

Alternatively, depending on the manufacturer’s claim, dfWL can also be written with σWLσrep instead of σdayσrep

where σWL=σday2+σrep2

The derivations of these equations are detailed in the Supplementary Materials.

To derive the false acceptance rate when the imprecision is actually greater than the manufacturer’s claim, the ratio σdayσrep is adopted to allow independent examination of the increase in σday and σrep while σWL is dependent on σrep. Hence, dfWL,eff will not be used in this study, but the formula is listed here for completeness. The statistical derivation of false acceptance rate is described in the section below. The false rejection rate can be derived as described previously.1

It should be noted that the effective degrees of freedom, dfeff, are derived based on the ANOVA table, which is established from observed experimental data. However, as described in CLSI EP15, dfeff are estimated based on the variances claimed by the instrument manufacturers rather than the observed variances. This approximation will allow users to evaluate the precision verification experimental design and estimate the corresponding false acceptance rate before collecting the data from the full experiments. In practice, the value of dfeff calculated from observed and claimed variance typically does not differ significantly.

Day × Run × Replicate Matrix Design

Under this experimental design, a fixed number of replicates are measured consecutively within a fixed number of times (runs) per day and repeated across multiple days. This experimental design is generally more resource intensive and less commonly employed for imprecision verification. For day × run × replicate designs, it is necessary to employ two-way nested ANOVA analysis to derive the components of imprecision.

The mathematical model for measurements obtained from performance evaluation for a day × run × replicate design is as follows:

where Yij = observed measurement on day i and replicate j; μ = mean, estimated as the average of all readings; Di = error due to day-to-day variation on day i; Rj(i) = error due to run-to-run variation for run j on day i; and εk(ij) = within-run error for replicate k for run j on day i.

The results obtained from the verification experiments can be summarized in the ANOVA table for a two-way nested design.

Source of VariationSum of SquaresDegrees of FreedomMean SquareExpected Mean Square
Between-daySS1dfday= nday1MS1nrunnrepσday2+nrepσrun2+σrep2
Between-run (nested in day)SS2dfrun=(nrun1)ndayMS2nrepσrun2+σrep2
Within-run (error)SS3dfrep=NndaynrunMS3σrep2
TotalSStotalN1Total
Source of VariationSum of SquaresDegrees of FreedomMean SquareExpected Mean Square
Between-daySS1dfday= nday1MS1nrunnrepσday2+nrepσrun2+σrep2
Between-run (nested in day)SS2dfrun=(nrun1)ndayMS2nrepσrun2+σrep2
Within-run (error)SS3dfrep=NndaynrunMS3σrep2
TotalSStotalN1Total
Source of VariationSum of SquaresDegrees of FreedomMean SquareExpected Mean Square
Between-daySS1dfday= nday1MS1nrunnrepσday2+nrepσrun2+σrep2
Between-run (nested in day)SS2dfrun=(nrun1)ndayMS2nrepσrun2+σrep2
Within-run (error)SS3dfrep=NndaynrunMS3σrep2
TotalSStotalN1Total
Source of VariationSum of SquaresDegrees of FreedomMean SquareExpected Mean Square
Between-daySS1dfday= nday1MS1nrunnrepσday2+nrepσrun2+σrep2
Between-run (nested in day)SS2dfrun=(nrun1)ndayMS2nrepσrun2+σrep2
Within-run (error)SS3dfrep=NndaynrunMS3σrep2
TotalSStotalN1Total

where

nday = number of days, nrun = number of runs per day, nrep = number of repetitions per run, and N=ndaynrunnrep.

The degrees of freedom for repeatability, dfrep=Nndaynrun, can be obtained directly from the ANOVA table. The degrees of freedom for between-day, between-run, and within-laboratory imprecision are derived based on the Satterthwaite formulation10:

Take note that dfday,eff is dependent on two ratios, σdayσrep and σrunσrep.

Although dfWL,eff will not be calculated in the Results section, it is listed here for completeness.

where

The day × run × rep design can be employed to verify the claims on σday, σrun, and σrep based on the observed standard deviation sday, srun, and srep, respectively. The false acceptance rate and false rejection rate can be derived as described in the section below.

Days × 1 Replicate and Replicates × 1 Day (Nonmatrix) Design

This is the most simplistic experimental design in which a single measurement is performed over multiple days or multiple replicate measurements are performed in a single day. The mean, σ, and CV can be simply calculated across the days (σday) or the replicate measurements (σrep). The within-laboratory imprecision (σWL) is calculated using all the data (ie, all the replicates and days). This study design does not require ANOVA analysis and complex formulation for determination of the degrees of freedom for each variance component.

Hypothesis Test

The imprecision of the assay can be assessed by comparing the observed standard deviation, s, with the manufacturer’s claimed σ using a hypothesis testing. The null hypothesis H0 and the alternative hypothesis H1 are as follows:

H0 suggests that the observed verification experimental variance s falls within the manufacturer’s claimed σ while H1 states that the observed s is larger than the manufacturer’s σ. The χ2 test of a single variance can be applied to inspect the validity of H0 by determining the probability of observing standard deviations of varying magnitude relative to σ. The value of χ2 is given by

where dfeff = effective degree of freedom.

False Rejection Rate

In its simplest form, a precision verification can reject an assay if any observed s is larger than manufacturer’s σ by direct comparison. However, even when the actual precision of the assay is equal to or better than the supplier’s claim, there is a possibility that the assay may be rejected through random sampling of the actual imprecision, leading to a false rejection. The false rejection rate can be obtained from the critical value from the χ2 distribution. When this simple direct comparison of standard deviations is used, the critical value or threshold value for χ2 is just the degrees of freedom as s is equal to σ (ie, H0 is true).

The area under the curve of the χ2 distribution to the right of this critical χ2 indicates the probability for obtaining s values that are higher than σ values, which is the false rejection rate α, given as (Supplemental Figure 1)

When multiple samples (ie, concentrations) are examined, it is assumed that the assay will be rejected if any one of the samples demonstrates a standard deviation, which is greater than the claimed value. Therefore, the total false rejection rate is given by

Impact of Adopting the UVL on the False Acceptance Rate

CLSI EP15 introduced the UVL to reduce the false rejection rate in assay precision verification experiments.2 The UVL factor, F, is a multiplier to the manufacturer’s claimed values introduced to limit the false rejection rate for σday and σrep to 5% each under the day × replicate experimental design.

where χα,df2 is the critical χ2 value with α of 5% for a single sample experiment. The dfeff of the χ2 distribution is determined by the calculation described in the previous sections. The value of α is dependent on the number of samples at different concentration levels, nsamples, employed in the verification experiments:

If the observed precision values are smaller than the claimed values multiplied by F (ie, s<Fσ⁠), assay verification can be accepted. However, the introduction of UVL increases the risk of false acceptance of assays that exceed the manufacturer’s claimed specifications (Supplemental Figure 1). To date, the CLSI EP15 document does not specify this false acceptance rate for a precision experiment with varying degrees of freedom.

False Acceptance Rate

If an assay imprecision is ψ times larger than the claimed value, the corresponding false acceptance rate β can be derived for the acceptance criteria based on simple comparison with CLSI EP15 guidelines.

For direct comparisons, the β value for each sample is given by

If UVL is applied, the β value for each sample is instead given by

When multiple samples are used, it is assumed that the assay will be accepted only when all the samples exhibit standard deviations that are smaller than the claimed value (Supplemental Figure 2). The total false acceptance rate is thus given by

The determination of UVL factor for a day × run × replicate matrix design follows the same procedures as the run × rep design.

Results

Day × Replicate (Balanced ANOVA) Design

Table 1 examines the effect of the number of samples included in the 5-day × 5-replicate design. Under this experimental design, the probability of having an observed imprecision, s, that is larger than the claimed σ, when the actual imprecision is equal to σ, is summarized in the false rejection rate column. When more samples are included in the precision experiment, they increase the false rejection rate as a consequence of the increased number of hypothesis tests being performed. In addition, in Table 1 the UVL keeps the false rejection rate at 5% for experiments involving a different number of samples but significantly increases the false acceptance rate. Also noteworthy is that the false acceptance rate reduces when the observed imprecision increases relative to the manufacturer’s claimed imprecision.

Table 1

Effect of the Number of Samples on False Acceptance and Rejection Rate With and Without the UVL for a 5-Day × 5-Replicate Designa

Between-Day ImprecisionRepeatabilityBetween-Day ImprecisionRepeatability
No. of SamplesFalse Acceptance Rate Without Upper Verification Limit, %False Rejection Rate Without Upper Verification Limit, %
Actual Factor of Increase From Claimed ValueObserved Factor of Increase From Claimed Value
1.11.21.51.11.21.5≥1.1≥1.2≥1.5≥1.1≥1.2≥1.5
152.144.527.931.716.41.630.422.98.023.49.20.1
227.119.87.810.02.70.051.640.515.441.317.50.2
314.18.82.23.20.40.066.354.222.255.025.10.3
47.43.90.61.00.10.076.664.628.565.532.00.4
53.81.70.20.30.00.083.772.734.273.638.20.6
62.00.80.00.10.00.088.779.039.579.843.90.7
False Acceptance Rate With Upper Verification Limit, %False Rejection Rate With Upper Verification Limit, %
188.685.767.683.364.916.75.05.05.05.05.00.1
287.082.756.880.155.45.45.05.05.05.05.00.2
385.980.649.478.149.12.05.05.05.05.05.00.3
485.179.043.876.544.40.85.05.05.05.05.00.4
584.577.739.475.240.70.45.05.05.05.05.00.6
683.976.535.774.137.60.25.05.05.05.05.00.7
Between-Day ImprecisionRepeatabilityBetween-Day ImprecisionRepeatability
No. of SamplesFalse Acceptance Rate Without Upper Verification Limit, %False Rejection Rate Without Upper Verification Limit, %
Actual Factor of Increase From Claimed ValueObserved Factor of Increase From Claimed Value
1.11.21.51.11.21.5≥1.1≥1.2≥1.5≥1.1≥1.2≥1.5
152.144.527.931.716.41.630.422.98.023.49.20.1
227.119.87.810.02.70.051.640.515.441.317.50.2
314.18.82.23.20.40.066.354.222.255.025.10.3
47.43.90.61.00.10.076.664.628.565.532.00.4
53.81.70.20.30.00.083.772.734.273.638.20.6
62.00.80.00.10.00.088.779.039.579.843.90.7
False Acceptance Rate With Upper Verification Limit, %False Rejection Rate With Upper Verification Limit, %
188.685.767.683.364.916.75.05.05.05.05.00.1
287.082.756.880.155.45.45.05.05.05.05.00.2
385.980.649.478.149.12.05.05.05.05.05.00.3
485.179.043.876.544.40.85.05.05.05.05.00.4
584.577.739.475.240.70.45.05.05.05.05.00.6
683.976.535.774.137.60.25.05.05.05.05.00.7

aThe claimed ratio of σ day/σ rep is set to a constant of 1.5 for all samples in all analyses. Factor = factorial difference in actual (for false acceptance) or observed (for false rejection) imprecision relative to the claimed imprecision. With upper verification limit (UVL), the false rejection rate is capped at 5%, but this only applies if the rejection event occurs. For example, with UVL, the rejection event will not happen at all if the observed standard deviation is 1.1× or 1.2× (ie, less than the UVL) of the claimed value.

Table 1

Effect of the Number of Samples on False Acceptance and Rejection Rate With and Without the UVL for a 5-Day × 5-Replicate Designa

Between-Day ImprecisionRepeatabilityBetween-Day ImprecisionRepeatability
No. of SamplesFalse Acceptance Rate Without Upper Verification Limit, %False Rejection Rate Without Upper Verification Limit, %
Actual Factor of Increase From Claimed ValueObserved Factor of Increase From Claimed Value
1.11.21.51.11.21.5≥1.1≥1.2≥1.5≥1.1≥1.2≥1.5
152.144.527.931.716.41.630.422.98.023.49.20.1
227.119.87.810.02.70.051.640.515.441.317.50.2
314.18.82.23.20.40.066.354.222.255.025.10.3
47.43.90.61.00.10.076.664.628.565.532.00.4
53.81.70.20.30.00.083.772.734.273.638.20.6
62.00.80.00.10.00.088.779.039.579.843.90.7
False Acceptance Rate With Upper Verification Limit, %False Rejection Rate With Upper Verification Limit, %
188.685.767.683.364.916.75.05.05.05.05.00.1
287.082.756.880.155.45.45.05.05.05.05.00.2
385.980.649.478.149.12.05.05.05.05.05.00.3
485.179.043.876.544.40.85.05.05.05.05.00.4
584.577.739.475.240.70.45.05.05.05.05.00.6
683.976.535.774.137.60.25.05.05.05.05.00.7
Between-Day ImprecisionRepeatabilityBetween-Day ImprecisionRepeatability
No. of SamplesFalse Acceptance Rate Without Upper Verification Limit, %False Rejection Rate Without Upper Verification Limit, %
Actual Factor of Increase From Claimed ValueObserved Factor of Increase From Claimed Value
1.11.21.51.11.21.5≥1.1≥1.2≥1.5≥1.1≥1.2≥1.5
152.144.527.931.716.41.630.422.98.023.49.20.1
227.119.87.810.02.70.051.640.515.441.317.50.2
314.18.82.23.20.40.066.354.222.255.025.10.3
47.43.90.61.00.10.076.664.628.565.532.00.4
53.81.70.20.30.00.083.772.734.273.638.20.6
62.00.80.00.10.00.088.779.039.579.843.90.7
False Acceptance Rate With Upper Verification Limit, %False Rejection Rate With Upper Verification Limit, %
188.685.767.683.364.916.75.05.05.05.05.00.1
287.082.756.880.155.45.45.05.05.05.05.00.2
385.980.649.478.149.12.05.05.05.05.05.00.3
485.179.043.876.544.40.85.05.05.05.05.00.4
584.577.739.475.240.70.45.05.05.05.05.00.6
683.976.535.774.137.60.25.05.05.05.05.00.7

aThe claimed ratio of σ day/σ rep is set to a constant of 1.5 for all samples in all analyses. Factor = factorial difference in actual (for false acceptance) or observed (for false rejection) imprecision relative to the claimed imprecision. With upper verification limit (UVL), the false rejection rate is capped at 5%, but this only applies if the rejection event occurs. For example, with UVL, the rejection event will not happen at all if the observed standard deviation is 1.1× or 1.2× (ie, less than the UVL) of the claimed value.

Table 2 displays the effect of the number of days and repetitions on the false acceptance and false rejection rate for experimental designs with three samples. Not surprisingly, the false rejection rate decreases as the observed imprecision increases relative to the claimed imprecision and when a greater number of days or replicates are included in the verification exercise. Increasing the number of days or replicates also reduces the false acceptance for between-day imprecision and repeatability. Nevertheless, there is a diminishing return on false acceptance and rejection rates with each incremental increase in the number of days or replicates.

Table 2

Effect of the Number of Days and Repetitions on False Acceptance and False Rejection Rate for Experimental Designs With Three Samples

Between-Day ImprecisionRepeatabilityBetween-Day ImprecisionRepeatability
False Acceptance Rate, %False Rejection Rate, %
No. of Replicates5 Days × nrep
Actual Factor of Increase From Claimed ValueObserved Factor of Increase From Claimed Value
1.11.21.51.11.21.5≥1.1≥1.2≥1.5≥1.1≥1.2≥1.5
214.18.82.298.495.269.666.354.222.265.950.013.4
314.18.82.276.555.59.666.354.222.262.439.83.8
414.18.82.227.710.10.266.354.222.258.631.61.1
514.18.82.23.20.40.066.354.222.255.025.10.3
614.18.82.20.10.00.066.354.222.251.620.00.1
74.32.20.30.00.00.066.352.117.248.515.90.0
204.32.20.30.00.00.066.352.117.222.21.00.0
No. of Daysnday × 5 Replicates
269.261.642.589.776.727.261.354.435.063.943.66.2
335.927.111.557.633.22.465.455.628.460.936.32.3
414.18.82.219.86.10.166.354.222.257.930.10.9
514.18.82.23.20.40.066.354.222.255.025.10.3
64.32.20.30.20.00.066.352.117.252.320.90.1
71.10.40.00.00.00.065.950.013.449.717.40.0
200.00.00.00.00.00.057.930.10.926.41.80.0
Between-Day ImprecisionRepeatabilityBetween-Day ImprecisionRepeatability
False Acceptance Rate, %False Rejection Rate, %
No. of Replicates5 Days × nrep
Actual Factor of Increase From Claimed ValueObserved Factor of Increase From Claimed Value
1.11.21.51.11.21.5≥1.1≥1.2≥1.5≥1.1≥1.2≥1.5
214.18.82.298.495.269.666.354.222.265.950.013.4
314.18.82.276.555.59.666.354.222.262.439.83.8
414.18.82.227.710.10.266.354.222.258.631.61.1
514.18.82.23.20.40.066.354.222.255.025.10.3
614.18.82.20.10.00.066.354.222.251.620.00.1
74.32.20.30.00.00.066.352.117.248.515.90.0
204.32.20.30.00.00.066.352.117.222.21.00.0
No. of Daysnday × 5 Replicates
269.261.642.589.776.727.261.354.435.063.943.66.2
335.927.111.557.633.22.465.455.628.460.936.32.3
414.18.82.219.86.10.166.354.222.257.930.10.9
514.18.82.23.20.40.066.354.222.255.025.10.3
64.32.20.30.20.00.066.352.117.252.320.90.1
71.10.40.00.00.00.065.950.013.449.717.40.0
200.00.00.00.00.00.057.930.10.926.41.80.0
Table 2

Effect of the Number of Days and Repetitions on False Acceptance and False Rejection Rate for Experimental Designs With Three Samples

Between-Day ImprecisionRepeatabilityBetween-Day ImprecisionRepeatability
False Acceptance Rate, %False Rejection Rate, %
No. of Replicates5 Days × nrep
Actual Factor of Increase From Claimed ValueObserved Factor of Increase From Claimed Value
1.11.21.51.11.21.5≥1.1≥1.2≥1.5≥1.1≥1.2≥1.5
214.18.82.298.495.269.666.354.222.265.950.013.4
314.18.82.276.555.59.666.354.222.262.439.83.8
414.18.82.227.710.10.266.354.222.258.631.61.1
514.18.82.23.20.40.066.354.222.255.025.10.3
614.18.82.20.10.00.066.354.222.251.620.00.1
74.32.20.30.00.00.066.352.117.248.515.90.0
204.32.20.30.00.00.066.352.117.222.21.00.0
No. of Daysnday × 5 Replicates
269.261.642.589.776.727.261.354.435.063.943.66.2
335.927.111.557.633.22.465.455.628.460.936.32.3
414.18.82.219.86.10.166.354.222.257.930.10.9
514.18.82.23.20.40.066.354.222.255.025.10.3
64.32.20.30.20.00.066.352.117.252.320.90.1
71.10.40.00.00.00.065.950.013.449.717.40.0
200.00.00.00.00.00.057.930.10.926.41.80.0
Between-Day ImprecisionRepeatabilityBetween-Day ImprecisionRepeatability
False Acceptance Rate, %False Rejection Rate, %
No. of Replicates5 Days × nrep
Actual Factor of Increase From Claimed ValueObserved Factor of Increase From Claimed Value
1.11.21.51.11.21.5≥1.1≥1.2≥1.5≥1.1≥1.2≥1.5
214.18.82.298.495.269.666.354.222.265.950.013.4
314.18.82.276.555.59.666.354.222.262.439.83.8
414.18.82.227.710.10.266.354.222.258.631.61.1
514.18.82.23.20.40.066.354.222.255.025.10.3
614.18.82.20.10.00.066.354.222.251.620.00.1
74.32.20.30.00.00.066.352.117.248.515.90.0
204.32.20.30.00.00.066.352.117.222.21.00.0
No. of Daysnday × 5 Replicates
269.261.642.589.776.727.261.354.435.063.943.66.2
335.927.111.557.633.22.465.455.628.460.936.32.3
414.18.82.219.86.10.166.354.222.257.930.10.9
514.18.82.23.20.40.066.354.222.255.025.10.3
64.32.20.30.20.00.066.352.117.252.320.90.1
71.10.40.00.00.00.065.950.013.449.717.40.0
200.00.00.00.00.00.057.930.10.926.41.80.0

Table 3 shows a comparison of different experimental designs of day × replicate for three samples with a similar total number of measurements performed. A higher number of day/replicates included in the experimental design have the effect of improving the confidence in the estimation of the imprecision, thereby leading to reduced false acceptance and false rejection rates.

Table 3

Comparison of Different Experiment Design for Three Samples With Similar Total Number of Measurements Performed

Experiment DesignFalse Acceptance Rate for Between-Day Imprecision, %False Acceptance Rate for Repeatability, %
Actual Factor of Increase From Claimed Value
1.11.21.51.11.21.5
5 days × 5 replicates14.18.82.23.20.40.0
6 days × 4 replicates4.32.20.38.81.80.0
8 days × 3 replicates1.10.40.019.86.10.0
12 days × 2 replicates0.00.00.057.633.22.4
15 days + 10 replicates4.61.00.06.82.30.0
10 days + 15 replicates6.82.30.04.61.00.0
5 days + 20 replicates11.96.61.13.40.50.0
20 days + 5 replicates3.40.50.011.96.61.1
5 days × 4 replicates14.18.82.227.710.10.2
10 days × 2 replicates0.20.00.076.555.59.6
10 days + 10 replicates6.82.30.16.82.30.1
Experiment DesignFalse Acceptance Rate for Between-Day Imprecision, %False Acceptance Rate for Repeatability, %
Actual Factor of Increase From Claimed Value
1.11.21.51.11.21.5
5 days × 5 replicates14.18.82.23.20.40.0
6 days × 4 replicates4.32.20.38.81.80.0
8 days × 3 replicates1.10.40.019.86.10.0
12 days × 2 replicates0.00.00.057.633.22.4
15 days + 10 replicates4.61.00.06.82.30.0
10 days + 15 replicates6.82.30.04.61.00.0
5 days + 20 replicates11.96.61.13.40.50.0
20 days + 5 replicates3.40.50.011.96.61.1
5 days × 4 replicates14.18.82.227.710.10.2
10 days × 2 replicates0.20.00.076.555.59.6
10 days + 10 replicates6.82.30.16.82.30.1
Table 3

Comparison of Different Experiment Design for Three Samples With Similar Total Number of Measurements Performed

Experiment DesignFalse Acceptance Rate for Between-Day Imprecision, %False Acceptance Rate for Repeatability, %
Actual Factor of Increase From Claimed Value
1.11.21.51.11.21.5
5 days × 5 replicates14.18.82.23.20.40.0
6 days × 4 replicates4.32.20.38.81.80.0
8 days × 3 replicates1.10.40.019.86.10.0
12 days × 2 replicates0.00.00.057.633.22.4
15 days + 10 replicates4.61.00.06.82.30.0
10 days + 15 replicates6.82.30.04.61.00.0
5 days + 20 replicates11.96.61.13.40.50.0
20 days + 5 replicates3.40.50.011.96.61.1
5 days × 4 replicates14.18.82.227.710.10.2
10 days × 2 replicates0.20.00.076.555.59.6
10 days + 10 replicates6.82.30.16.82.30.1
Experiment DesignFalse Acceptance Rate for Between-Day Imprecision, %False Acceptance Rate for Repeatability, %
Actual Factor of Increase From Claimed Value
1.11.21.51.11.21.5
5 days × 5 replicates14.18.82.23.20.40.0
6 days × 4 replicates4.32.20.38.81.80.0
8 days × 3 replicates1.10.40.019.86.10.0
12 days × 2 replicates0.00.00.057.633.22.4
15 days + 10 replicates4.61.00.06.82.30.0
10 days + 15 replicates6.82.30.04.61.00.0
5 days + 20 replicates11.96.61.13.40.50.0
20 days + 5 replicates3.40.50.011.96.61.1
5 days × 4 replicates14.18.82.227.710.10.2
10 days × 2 replicates0.20.00.076.555.59.6
10 days + 10 replicates6.82.30.16.82.30.1

Day × Run × Replicate (Balanced ANOVA) Design

Under this experimental design, the false acceptance rate is similarly reduced by increasing the number of samples, number of days, number of runs, and number of replicates performed, as well as the degree the actual imprecision exceeds the claimed imprecision Table 4 and Table 5. As seen in the day × replicate experimental design above, there is a clear diminishing return on increasing the number of days, runs, or replicates on the false acceptance rate, shown in Tables 4 and 5.

Table 4

Effect of the Number of Samples, Days, Runs, and Replicates on False Acceptance Rate for Experimental Designsa

20-Day × 2-Run × 2-Replicate Design
No. of SamplesFalse Acceptance Rate for Between-Day Imprecision, %False Acceptance Rate for Between-Run Imprecision, %False Acceptance Rate for Repeatability, %
Actual Factor of Increase From Claimed Value
1.11.21.51.11.21.51.11.21.5
140.828.58.938.625.56.422.67.20.1
216.78.10.814.96.50.45.10.50.0
36.82.30.15.81.70.01.20.00.0
42.80.70.02.20.40.00.30.00.0
51.10.20.00.90.10.00.10.00.0
60.50.10.00.30.00.00.00.00.0
No. of Daysn Days × 2 Runs × 2 Replicates With 3 Samples
517.812.54.614.18.82.26.32.00.0
1011.96.61.110.35.20.63.20.40.0
159.24.10.37.52.80.11.90.10.0
206.82.30.15.81.70.01.20.00.0
255.81.70.05.01.20.00.80.00.0
No. of Runs20 Days × n Runs × 2 Replicates With 3 Samples
26.82.30.15.81.70.01.20.00.0
35.81.70.03.00.40.00.50.00.0
45.01.20.01.80.10.00.20.00.0
No. of Replicates20 Days × 2 Runs × nrep With 3 Samples
26.82.30.15.81.70.01.20.00.0
36.82.30.15.01.20.00.20.00.0
46.82.30.14.61.00.00.10.00.0
20-Day × 2-Run × 2-Replicate Design
No. of SamplesFalse Acceptance Rate for Between-Day Imprecision, %False Acceptance Rate for Between-Run Imprecision, %False Acceptance Rate for Repeatability, %
Actual Factor of Increase From Claimed Value
1.11.21.51.11.21.51.11.21.5
140.828.58.938.625.56.422.67.20.1
216.78.10.814.96.50.45.10.50.0
36.82.30.15.81.70.01.20.00.0
42.80.70.02.20.40.00.30.00.0
51.10.20.00.90.10.00.10.00.0
60.50.10.00.30.00.00.00.00.0
No. of Daysn Days × 2 Runs × 2 Replicates With 3 Samples
517.812.54.614.18.82.26.32.00.0
1011.96.61.110.35.20.63.20.40.0
159.24.10.37.52.80.11.90.10.0
206.82.30.15.81.70.01.20.00.0
255.81.70.05.01.20.00.80.00.0
No. of Runs20 Days × n Runs × 2 Replicates With 3 Samples
26.82.30.15.81.70.01.20.00.0
35.81.70.03.00.40.00.50.00.0
45.01.20.01.80.10.00.20.00.0
No. of Replicates20 Days × 2 Runs × nrep With 3 Samples
26.82.30.15.81.70.01.20.00.0
36.82.30.15.01.20.00.20.00.0
46.82.30.14.61.00.00.10.00.0

aThe claimed ratios of σ day/σ rep and σ run/σ rep are set to 1.5 and 1.2, respectively.

Table 4

Effect of the Number of Samples, Days, Runs, and Replicates on False Acceptance Rate for Experimental Designsa

20-Day × 2-Run × 2-Replicate Design
No. of SamplesFalse Acceptance Rate for Between-Day Imprecision, %False Acceptance Rate for Between-Run Imprecision, %False Acceptance Rate for Repeatability, %
Actual Factor of Increase From Claimed Value
1.11.21.51.11.21.51.11.21.5
140.828.58.938.625.56.422.67.20.1
216.78.10.814.96.50.45.10.50.0
36.82.30.15.81.70.01.20.00.0
42.80.70.02.20.40.00.30.00.0
51.10.20.00.90.10.00.10.00.0
60.50.10.00.30.00.00.00.00.0
No. of Daysn Days × 2 Runs × 2 Replicates With 3 Samples
517.812.54.614.18.82.26.32.00.0
1011.96.61.110.35.20.63.20.40.0
159.24.10.37.52.80.11.90.10.0
206.82.30.15.81.70.01.20.00.0
255.81.70.05.01.20.00.80.00.0
No. of Runs20 Days × n Runs × 2 Replicates With 3 Samples
26.82.30.15.81.70.01.20.00.0
35.81.70.03.00.40.00.50.00.0
45.01.20.01.80.10.00.20.00.0
No. of Replicates20 Days × 2 Runs × nrep With 3 Samples
26.82.30.15.81.70.01.20.00.0
36.82.30.15.01.20.00.20.00.0
46.82.30.14.61.00.00.10.00.0
20-Day × 2-Run × 2-Replicate Design
No. of SamplesFalse Acceptance Rate for Between-Day Imprecision, %False Acceptance Rate for Between-Run Imprecision, %False Acceptance Rate for Repeatability, %
Actual Factor of Increase From Claimed Value
1.11.21.51.11.21.51.11.21.5
140.828.58.938.625.56.422.67.20.1
216.78.10.814.96.50.45.10.50.0
36.82.30.15.81.70.01.20.00.0
42.80.70.02.20.40.00.30.00.0
51.10.20.00.90.10.00.10.00.0
60.50.10.00.30.00.00.00.00.0
No. of Daysn Days × 2 Runs × 2 Replicates With 3 Samples
517.812.54.614.18.82.26.32.00.0
1011.96.61.110.35.20.63.20.40.0
159.24.10.37.52.80.11.90.10.0
206.82.30.15.81.70.01.20.00.0
255.81.70.05.01.20.00.80.00.0
No. of Runs20 Days × n Runs × 2 Replicates With 3 Samples
26.82.30.15.81.70.01.20.00.0
35.81.70.03.00.40.00.50.00.0
45.01.20.01.80.10.00.20.00.0
No. of Replicates20 Days × 2 Runs × nrep With 3 Samples
26.82.30.15.81.70.01.20.00.0
36.82.30.15.01.20.00.20.00.0
46.82.30.14.61.00.00.10.00.0

aThe claimed ratios of σ day/σ rep and σ run/σ rep are set to 1.5 and 1.2, respectively.

Table 5

Comparison of Different Experimental Designs With an Equal Number of Total Measurements With Three Samplesa

Experimental Design: Day × Run × ReplicateFalse Acceptance Rate for Between-Day Imprecision, %False Acceptance Rate for Between-Run Imprecision, %False Acceptance Rate for Repeatability, %
Actual Factor of Increase From Claimed Value
1.11.21.51.11.21.51.11.21.5
20 × 2 × 26.82.30.15.81.70.01.20.00.0
20 × 1 × 44.30.90.00.50.00.0
10 × 4 × 29.24.10.34.10.80.01.20.00.0
10 × 2 × 411.96.61.18.23.40.20.50.00.0
10 × 1 × 87.52.80.10.30.00.0
5 × 4 × 414.18.82.25.81.70.00.50.00.0
5 × 2 × 817.812.54.611.96.61.10.30.00.0
5 × 1 × 1611.96.61.10.30.00.0
Experimental Design: Day × Run × ReplicateFalse Acceptance Rate for Between-Day Imprecision, %False Acceptance Rate for Between-Run Imprecision, %False Acceptance Rate for Repeatability, %
Actual Factor of Increase From Claimed Value
1.11.21.51.11.21.51.11.21.5
20 × 2 × 26.82.30.15.81.70.01.20.00.0
20 × 1 × 44.30.90.00.50.00.0
10 × 4 × 29.24.10.34.10.80.01.20.00.0
10 × 2 × 411.96.61.18.23.40.20.50.00.0
10 × 1 × 87.52.80.10.30.00.0
5 × 4 × 414.18.82.25.81.70.00.50.00.0
5 × 2 × 817.812.54.611.96.61.10.30.00.0
5 × 1 × 1611.96.61.10.30.00.0

aThe claimed ratios of σ day/σ rep and σ run/σ rep are set to 1.5 and 1.2, respectively.

Table 5

Comparison of Different Experimental Designs With an Equal Number of Total Measurements With Three Samplesa

Experimental Design: Day × Run × ReplicateFalse Acceptance Rate for Between-Day Imprecision, %False Acceptance Rate for Between-Run Imprecision, %False Acceptance Rate for Repeatability, %
Actual Factor of Increase From Claimed Value
1.11.21.51.11.21.51.11.21.5
20 × 2 × 26.82.30.15.81.70.01.20.00.0
20 × 1 × 44.30.90.00.50.00.0
10 × 4 × 29.24.10.34.10.80.01.20.00.0
10 × 2 × 411.96.61.18.23.40.20.50.00.0
10 × 1 × 87.52.80.10.30.00.0
5 × 4 × 414.18.82.25.81.70.00.50.00.0
5 × 2 × 817.812.54.611.96.61.10.30.00.0
5 × 1 × 1611.96.61.10.30.00.0
Experimental Design: Day × Run × ReplicateFalse Acceptance Rate for Between-Day Imprecision, %False Acceptance Rate for Between-Run Imprecision, %False Acceptance Rate for Repeatability, %
Actual Factor of Increase From Claimed Value
1.11.21.51.11.21.51.11.21.5
20 × 2 × 26.82.30.15.81.70.01.20.00.0
20 × 1 × 44.30.90.00.50.00.0
10 × 4 × 29.24.10.34.10.80.01.20.00.0
10 × 2 × 411.96.61.18.23.40.20.50.00.0
10 × 1 × 87.52.80.10.30.00.0
5 × 4 × 414.18.82.25.81.70.00.50.00.0
5 × 2 × 817.812.54.611.96.61.10.30.00.0
5 × 1 × 1611.96.61.10.30.00.0

aThe claimed ratios of σ day/σ rep and σ run/σ rep are set to 1.5 and 1.2, respectively.

Days × 1 Replicate or Replicates × 1 Day Nonmatrix Design

Finally, some laboratories may not adopt the matrix experimental design (balanced ANOVA). Instead, the between-day imprecision is estimated by measuring the samples once per day across multiple days. The repeatability is estimated by running the samples multiple times in a single day. Under this experimental design, the claimed ratio of σ day/σ rep is not relevant here, as the degrees of freedom for each variance component are independent of the claimed ratio. The degrees of freedom for each variance component are simply the number of measurements used in the calculation minus 1. In addition, the role of days and replicates is interchangeable in such designs.

As with the matrix experimental designs, increasing the number of samples, days, and replicates, as well as the magnitude of increase in imprecision relative to the manufacturer’s claim, reduces the false acceptance rate with diminishing returns, as shown in Table 6. When nday = 1, between-day imprecision cannot be defined; similarly, when nrep = 1, repeatability within the same day cannot be determined. Therefore, the cases for nday or nrep with the value of 1 are not shown in Table 6. In general, for a given total number of measurements, the nonmatrix experimental design performs well compared to the day × replicate (matrix) experimental design when an equal number of measurements are allocated for day and replicate, respectively, as shown in Table 3.

Table 6

Effect of Number of Samples, Number of Days, and Number of Replicates on the False Acceptance and False Rejection Rate for Between-Day Imprecision and Repeatability for the 10-Day + Replicates and 10-Replicate + Days Experiment Designa

False Acceptance Rate, %False Rejection Rate, %
Between-Day ImprecisionRepeatabilityBetween-Day ImprecisionRepeatability
No. of Samples10 Days + 10 Replicates
Actual Factor of Increase From Claimed ValueObserved Factor of Increase From Claimed Value
1.11.21.51.11.21.5≥1.1≥1.2≥1.5≥1.1≥1.2≥1.5
140.828.58.940.828.58.928.316.41.628.316.41.6
216.78.10.816.78.10.848.630.23.348.630.23.3
36.82.30.16.82.30.163.241.74.863.241.74.8
42.80.70.02.80.70.073.651.36.473.651.36.4
51.10.20.01.10.20.081.159.37.981.159.37.9
60.50.10.00.50.10.086.566.09.586.566.09.5
No. of Replicates in 1 Day10 Days + nrep With 3 Samples
26.82.30.125.821.112.163.241.74.861.354.435.0
36.82.30.117.812.54.663.241.74.865.455.628.4
46.82.30.114.18.82.263.241.74.866.354.222.2
56.82.30.111.96.61.163.241.74.866.352.117.2
66.82.30.110.35.20.663.241.74.865.950.013.4
76.82.30.19.24.10.363.241.74.865.347.810.3
86.82.30.18.23.40.263.241.74.864.745.78.0
96.82.30.17.52.80.163.241.74.863.943.66.2
106.82.30.16.82.30.163.241.74.863.241.74.8
156.82.30.14.61.00.063.241.74.859.433.01.4
206.82.30.13.40.50.063.241.74.855.726.30.4
No. of Daysnday + 10 Replicates With 3 Samples
225.821.112.16.82.30.161.354.435.063.241.74.8
317.812.54.66.82.30.165.455.628.463.241.74.8
414.18.82.26.82.30.166.354.222.263.241.74.8
511.96.61.16.82.30.166.352.117.263.241.74.8
610.35.20.66.82.30.165.950.013.463.241.74.8
79.24.10.36.82.30.165.347.810.363.241.74.8
88.23.40.26.82.30.164.745.78.063.241.74.8
97.52.80.16.82.30.163.943.66.263.241.74.8
106.82.30.16.82.30.163.241.74.863.241.74.8
154.61.00.06.82.30.159.433.01.463.241.74.8
203.40.50.06.82.30.155.726.30.463.241.74.8
False Acceptance Rate, %False Rejection Rate, %
Between-Day ImprecisionRepeatabilityBetween-Day ImprecisionRepeatability
No. of Samples10 Days + 10 Replicates
Actual Factor of Increase From Claimed ValueObserved Factor of Increase From Claimed Value
1.11.21.51.11.21.5≥1.1≥1.2≥1.5≥1.1≥1.2≥1.5
140.828.58.940.828.58.928.316.41.628.316.41.6
216.78.10.816.78.10.848.630.23.348.630.23.3
36.82.30.16.82.30.163.241.74.863.241.74.8
42.80.70.02.80.70.073.651.36.473.651.36.4
51.10.20.01.10.20.081.159.37.981.159.37.9
60.50.10.00.50.10.086.566.09.586.566.09.5
No. of Replicates in 1 Day10 Days + nrep With 3 Samples
26.82.30.125.821.112.163.241.74.861.354.435.0
36.82.30.117.812.54.663.241.74.865.455.628.4
46.82.30.114.18.82.263.241.74.866.354.222.2
56.82.30.111.96.61.163.241.74.866.352.117.2
66.82.30.110.35.20.663.241.74.865.950.013.4
76.82.30.19.24.10.363.241.74.865.347.810.3
86.82.30.18.23.40.263.241.74.864.745.78.0
96.82.30.17.52.80.163.241.74.863.943.66.2
106.82.30.16.82.30.163.241.74.863.241.74.8
156.82.30.14.61.00.063.241.74.859.433.01.4
206.82.30.13.40.50.063.241.74.855.726.30.4
No. of Daysnday + 10 Replicates With 3 Samples
225.821.112.16.82.30.161.354.435.063.241.74.8
317.812.54.66.82.30.165.455.628.463.241.74.8
414.18.82.26.82.30.166.354.222.263.241.74.8
511.96.61.16.82.30.166.352.117.263.241.74.8
610.35.20.66.82.30.165.950.013.463.241.74.8
79.24.10.36.82.30.165.347.810.363.241.74.8
88.23.40.26.82.30.164.745.78.063.241.74.8
97.52.80.16.82.30.163.943.66.263.241.74.8
106.82.30.16.82.30.163.241.74.863.241.74.8
154.61.00.06.82.30.159.433.01.463.241.74.8
203.40.50.06.82.30.155.726.30.463.241.74.8

aFor the experiment that varies number of samples, it is assumed that the measurements were performed within the 10-day + 10-replicate design.

Table 6

Effect of Number of Samples, Number of Days, and Number of Replicates on the False Acceptance and False Rejection Rate for Between-Day Imprecision and Repeatability for the 10-Day + Replicates and 10-Replicate + Days Experiment Designa

False Acceptance Rate, %False Rejection Rate, %
Between-Day ImprecisionRepeatabilityBetween-Day ImprecisionRepeatability
No. of Samples10 Days + 10 Replicates
Actual Factor of Increase From Claimed ValueObserved Factor of Increase From Claimed Value
1.11.21.51.11.21.5≥1.1≥1.2≥1.5≥1.1≥1.2≥1.5
140.828.58.940.828.58.928.316.41.628.316.41.6
216.78.10.816.78.10.848.630.23.348.630.23.3
36.82.30.16.82.30.163.241.74.863.241.74.8
42.80.70.02.80.70.073.651.36.473.651.36.4
51.10.20.01.10.20.081.159.37.981.159.37.9
60.50.10.00.50.10.086.566.09.586.566.09.5
No. of Replicates in 1 Day10 Days + nrep With 3 Samples
26.82.30.125.821.112.163.241.74.861.354.435.0
36.82.30.117.812.54.663.241.74.865.455.628.4
46.82.30.114.18.82.263.241.74.866.354.222.2
56.82.30.111.96.61.163.241.74.866.352.117.2
66.82.30.110.35.20.663.241.74.865.950.013.4
76.82.30.19.24.10.363.241.74.865.347.810.3
86.82.30.18.23.40.263.241.74.864.745.78.0
96.82.30.17.52.80.163.241.74.863.943.66.2
106.82.30.16.82.30.163.241.74.863.241.74.8
156.82.30.14.61.00.063.241.74.859.433.01.4
206.82.30.13.40.50.063.241.74.855.726.30.4
No. of Daysnday + 10 Replicates With 3 Samples
225.821.112.16.82.30.161.354.435.063.241.74.8
317.812.54.66.82.30.165.455.628.463.241.74.8
414.18.82.26.82.30.166.354.222.263.241.74.8
511.96.61.16.82.30.166.352.117.263.241.74.8
610.35.20.66.82.30.165.950.013.463.241.74.8
79.24.10.36.82.30.165.347.810.363.241.74.8
88.23.40.26.82.30.164.745.78.063.241.74.8
97.52.80.16.82.30.163.943.66.263.241.74.8
106.82.30.16.82.30.163.241.74.863.241.74.8
154.61.00.06.82.30.159.433.01.463.241.74.8
203.40.50.06.82.30.155.726.30.463.241.74.8
False Acceptance Rate, %False Rejection Rate, %
Between-Day ImprecisionRepeatabilityBetween-Day ImprecisionRepeatability
No. of Samples10 Days + 10 Replicates
Actual Factor of Increase From Claimed ValueObserved Factor of Increase From Claimed Value
1.11.21.51.11.21.5≥1.1≥1.2≥1.5≥1.1≥1.2≥1.5
140.828.58.940.828.58.928.316.41.628.316.41.6
216.78.10.816.78.10.848.630.23.348.630.23.3
36.82.30.16.82.30.163.241.74.863.241.74.8
42.80.70.02.80.70.073.651.36.473.651.36.4
51.10.20.01.10.20.081.159.37.981.159.37.9
60.50.10.00.50.10.086.566.09.586.566.09.5
No. of Replicates in 1 Day10 Days + nrep With 3 Samples
26.82.30.125.821.112.163.241.74.861.354.435.0
36.82.30.117.812.54.663.241.74.865.455.628.4
46.82.30.114.18.82.263.241.74.866.354.222.2
56.82.30.111.96.61.163.241.74.866.352.117.2
66.82.30.110.35.20.663.241.74.865.950.013.4
76.82.30.19.24.10.363.241.74.865.347.810.3
86.82.30.18.23.40.263.241.74.864.745.78.0
96.82.30.17.52.80.163.241.74.863.943.66.2
106.82.30.16.82.30.163.241.74.863.241.74.8
156.82.30.14.61.00.063.241.74.859.433.01.4
206.82.30.13.40.50.063.241.74.855.726.30.4
No. of Daysnday + 10 Replicates With 3 Samples
225.821.112.16.82.30.161.354.435.063.241.74.8
317.812.54.66.82.30.165.455.628.463.241.74.8
414.18.82.26.82.30.166.354.222.263.241.74.8
511.96.61.16.82.30.166.352.117.263.241.74.8
610.35.20.66.82.30.165.950.013.463.241.74.8
79.24.10.36.82.30.165.347.810.363.241.74.8
88.23.40.26.82.30.164.745.78.063.241.74.8
97.52.80.16.82.30.163.943.66.263.241.74.8
106.82.30.16.82.30.163.241.74.863.241.74.8
154.61.00.06.82.30.159.433.01.463.241.74.8
203.40.50.06.82.30.155.726.30.463.241.74.8

aFor the experiment that varies number of samples, it is assumed that the measurements were performed within the 10-day + 10-replicate design.

Discussion

Precision verification is a critical preimplementation method evaluation study. However, it can be resource intensive as it involves repeated testing that consumes reagents, quality control materials, pooled patient samples, and other consumables. In addition, it can be operationally demanding as it requires dedicated suitably trained personnel to perform the measurements and collect and analyze the data, resulting in a time-consuming process. While false rejection of the imprecision verification data may result in a new round of verification exercises that is similarly resource and time-consuming, a false acceptance may potentially lead to adverse clinical outcomes. Hence, laboratories should carefully consider the available resources and design the precision verification study to address these analytical and clinical requirements in an informed and balanced manner. Often, the laboratory is constrained by the total number of measurements it can perform for the entire precision verification study.

From the results of this study, several broad recommendations can be made.

Experimental Design

Experimental design can be adjusted based on which of the imprecision components the laboratory wishes to emphasize. In general, more importance should be given to verifying the imprecision component with the largest variation as reported by the manufacturer, which is most often the between-day imprecision. Accordingly, the experimental design should perform the verification over more days, even at the expense of less replicates/runs per day. For example, if the laboratory has determined that it can only perform a total of 20 measurements for a given concentration for a precision verification study, an experimental design with 10 days × 2 replicates will better verify the between-day imprecision compared with a 5-day × 4-replicate experimental design—but at the significant expense of the repeatability component (Table 3).

Importantly, laboratory practitioners should bear in mind the diminishing returns of increasing the number of measurements for a particular experiment component (day, run, replicate, sample). While a matrix experimental design may allow the laboratory to complete the verification exercise more expediently compared with a nonmatrix design, it is generally associated with higher false acceptance rates. The inclusion of more measurements in the verification exercise reduces false acceptance and false rejection rates by reducing the uncertainty and subsequent confidence limits of the estimates.

Statistical Approach

To optimize the data collected, the laboratory should apply the correct statistical approaches. A matrix experimental design is optimally evaluated using ANOVA techniques. Software for analysis of simple matrix designs can be performed using Excel (Microsoft). However, more complex designs may require use of dedicated software packages such as the VCA package11 in the R statistical environment.12 Nonmatrix experimental designs can be simplistically examined using simple standard deviations and coefficients of variation.

Consider Both False Acceptance and False Rejection Rates

While there is an instinctive preference to reduce false rejection rates and avoid unnecessarily repeating experiments, it is also important to consider the probability of falsely accepting an assay that has increased imprecision. An increased imprecision has significant clinical impact, producing results with greater variability that can lead to an erroneous diagnosis or obscure the trends of a measurement when serially monitored.4-6 Moreover, it can also increase the frequency of quality control and proficiency testing failures that require significant efforts to investigate, leading to long-term operational inefficiencies.

Replace Missing Values

During an imprecision verification, a measurement value(s) may be missing due to outlier exclusion or missed measurement(s). However, it is important to ensure that the missing values of a verification exercise are suitably replaced. This is particularly important for the matrix experimental designs since most statistical software available to the laboratories assumes a balanced design (ie, no missing values). To replace the missing value(s), the experiment for the entire day should be repeated to ensure the variability of the assay is again adequately captured.

Considerations for Repeating a Failed Imprecision Verification Exercise

When a verification exercise fails to meet the manufacturer’s claim, the laboratory may elect to repeat the entire process. In such instances, the laboratory should select an experimental design that best assesses the particular failed component. For example, if the repeatability failed to verify against the manufacturer’s claim using a matrix experimental design with 10 days × 2 replicates, a repeatability verification exercise by measuring 10 replicates in a single day should be considered.

What to Do When the Precision Verification Fails More Than Once

Assuming that the precision study is conducted properly, and the laboratory is satisfied that there are no pre- or postanalytical issues at fault (eg, instability of materials, erroneous material handling, data entry errors), a repeated failure of the precision verification is likely to indicate that the assay is not performing to the manufacturer’s claims. This is especially true if the verification failed despite application of the UVL, since this limit is meant to cap the false rejection rate at 5%. A repeat failure means that there is only 0.25% (5% × 5%) probability that the assay is performing within the manufacturer’s claim.

References

1.

Clinical and Laboratory Standards Institute.
User Verification of Precision and Estimation of Bias; Approved Guideline.
3rd ed.
Wayne, PA
:
CLSI
;
2014
.
CLSI Document EP15-A3.

2.

International Organization for Standardization.
Medical Laboratories—Requirements for Quality and Competence.
3rd ed.
Geneva, Switzerland: International Organization of Standardization
;
2012
.

3.

College of American Pathologists.
All Common Checklist
. COM.40300 Verification of Test Performance Specifications: FDA-Cleared/Approved Tests. Northfield, IL: College of American Pathology; 2020.

4.

Petersen
PH
,
Jørgensen
LG
,
Brandslund
I
, et al.
Consequences of bias and imprecision in measurements of glucose and HbA1c for the diagnosis and prognosis of diabetes mellitus
.
Scand J Clin Lab Invest Suppl.
2005
;
240
:
51
-
60
.

5.

Hyltoft Petersen
P
,
Klee
GG
.
Influence of analytical bias and imprecision on the number of false positive results using Guideline-Driven Medical Decision Limits
.
Clin Chim Acta.
2014
;
430
:
1
-
8
.

6.

Chai
JH
,
Ma
S
,
Heng
D
, et al.
Impact of analytical and biological variations on classification of diabetes using fasting plasma glucose, oral glucose tolerance test and HbA1c
.
Sci Rep.
2017
;
7
:
13721
.

7.

Hand
M
,
Crampton
A
,
Thomas
A
, et al.
A survey of clinical laboratory instrument verification in the UK and New Zealand
.
Ann Clin Biochem.
2019
;
56
:
275
-
282
.

8.

Satterthwaite
FE
.
An approximate distribution of estimates of variance components
.
Biometrics.
1946
;
2
:
110
-
114
.

9.

Graybill
FA.
An Introduction to Linear Statistical Models.
Vol. 1.
New York, NY
:
McGraw-Hill
;
1961
.

10.

Satterthwaite
FE
.
Synthesis of variance
.
Psychometrika.
1941
;
6
:
309
-
316
.

11.

Schuetzenmeister
A
,
Dufey
F
.
VCA: variance component analysis. R package version 1.4.3.
https://CRAN.R-project.org/package=VCA. Accessed
January 31,
2021
.

12.

R Core Team.
R: A Language and Environment for Statistical Computing.
Vienna, Austria
:
R Foundation for Statistical Computing; 2020
.

This article is published and distributed under the terms of the Oxford University Press, Standard Journals Publication Model (https://dbpia.nl.go.kr/pages/standard-publication-reuse-rights)