# 15.4: Evaluating Quality Assurance Data

- Page ID
- 5344

In the previous section we described several internal methods of quality assessment that provide quantitative estimates of the systematic errors and the random errors in an analytical method. Now we turn our attention to how we incorporate this quality assessment data into a complete quality assurance program. There are two general approaches to developing a quality assurance program: a prescriptive approach, in which we prescribe an exact method of quality assessment, and a performance-based approach in which we can use any form of quality assessment, provided that we can demonstrate an acceptable level of statistical control.^{3}

### 15.4.1 Prescriptive Approach

With a prescriptive approach to quality assessment, duplicate samples, blanks, standards, and spike recoveries are measured following a specific protocol. We compare the result for each analysis to a single predetermined limit, taking an appropriate corrective action if the limit is exceeded. Prescriptive approaches to quality assurance are common for programs and laboratories subject to federal regulation. For example, the Food and Drug Administration (FDA) specifies quality assurance practices that must be followed by laboratories analyzing products regulated by the FDA.

Figure 15.2 provides a typical example of a prescriptive approach to quality assessment. Two samples, A and B, are collected at the sample site. Sample A is split into two equal-volume samples, A_{1} and A_{2}. Sample B is also split into two equal-volume samples, one of which, B_{SF}, is spiked in the field with a known amount of analyte. A field blank, D_{F}, also is spiked with the same amount of analyte. All five samples (A_{1}, A_{2}, B, B_{SF}, and D_{F}) are preserved if necessary and transported to the laboratory for analysis.

After returning to the lab, the first sample that is analyzed is the field blank. If its spike recovery is unacceptable—an indication of a systematic error in the field or in the lab—then a laboratory method blank, D_{L}, is prepared and analyzed. If the spike recovery for the method blank is unsatisfactory, then the systematic error originated in the laboratory; this is something we can find and correct before proceeding with the analysis. An acceptable spike recovery for the method blank, however, indicates that the systematic error occurred in the field or during transport to the laboratory, casting uncertainty on the quality of the samples. The only recourse is to discard the samples and return to the field to collect new samples.

If the field blank is satisfactory, then sample B is analyzed. If the result for B is above the method’s detection limit, or if it is within the range of 0.1 to 10 times the amount of analyte spiked into B_{SF}, then a spike recovery for B_{SF} is determined. An unacceptable spike recovery for B_{SF} indicates the presence of a systematic error involving the sample. To determine the source of the systematic error, a laboratory spike, B_{SL}, is prepared using sample B, and analyzed. If the spike recovery for B_{SL} is acceptable, then the systematic error requires a long time to have a noticeable effect on the spike recovery. One possible explanation is that the analyte has not been properly preserved or it has been held beyond the acceptable holding time. An unacceptable spike recovery for B_{SL} suggests an immediate systematic error, such as that due to the influence of the sample’s matrix. In either case the systematic errors are fatal and must be corrected before the sample is reanalyzed.

This is one reason that environmental testing is so expensive.

If the spike recovery for B_{SF} is acceptable, or if the result for sample B is below the method’s detection limit, or outside the range of 0.1 to 10 times the amount of analyte spiked in B_{SF}, then the duplicate samples A_{1} and A_{2} are analyzed. The results for A_{1} and A_{2} are discarded if the difference between their values is excessive. If the difference between the results for A_{1} and A_{2} is within the accepted limits, then the results for samples A_{1} and B are compared. Because samples collected from the same sampling site at the same time should be identical in composition, the results are discarded if the difference between their values is unsatisfactory, and accepted if the difference is satisfactory.

**Figure 15.2:** Example of a prescriptive approach to quality assurance for laboratories monitoring waters and wastewaters. Adapted from Environmental Monitoring and Support Laboratory, U. S. Environmental Protection Agency, “Handbook for Analytical Quality Control in Water and Wastewater Laboratories,” March 1979.

The protocol in Figure 15.2 requires four to five evaluations of quality assessment data before the result for a single sample is accepted, a process that we must repeat for each analyte and for each sample. Other prescriptive protocols are equally demanding. For example, Figure 3.7 in Chapter 3 shows a portion of a quality assurance protocol for the graphite furnace atomic absorption analysis of trace metals in aqueous solutions. This protocol involves the analysis of an initial calibration verification standard and an initial calibration blank, followed by the analysis of samples in groups of ten. Each group of samples is preceded and followed by continuing calibration verification (CCV) and continuing calibration blank (CCB) quality assessment samples. Results for each group of ten samples are accepted only if both sets of CCV and CCB quality assessment samples are acceptable.

The advantage of a prescriptive approach to quality assurance is that all laboratories use a single consistent set of guideline. A significant disadvantage is that it does not take into account a laboratory’s ability to produce quality results when determining the frequency of collecting and analyzing quality assessment data. A laboratory with a record of producing high quality results is forced to spend more time and money on quality assessment than is perhaps necessary. At the same time, the frequency of quality assessment may be insufficient for a laboratory with a history of producing results of poor quality.

### 15.4.2 Performance-Based Approach

In a performance-based approach to quality assurance, a laboratory is free to use its experience to determine the best way to gather and monitor quality assessment data. The quality assessment methods remain the same—duplicate samples, blanks, standards, and spike recoveries—because they provide the necessary information about precision and bias. What a laboratory can control is the frequency with which it analyzes quality assessment samples and the conditions it chooses to identify when an analysis falls out of a state of statistical control.

The principal tool for performance-based quality assessment is a **control chart**, which provides a continuous record of quality assessment results. Our fundamental assumption is that if an analysis is under statistical control, individual quality assessment results are randomly distributed around a known mean with a known standard deviation. When an analysis moves out of statistical control, the quality assessment data is influenced by additional sources of error, increasing the standard deviation or changing the mean value.

Control charts were originally developed in the 1920s as a quality assurance tool for the control of manufactured products.^{4} Although there are many types of control charts, two are common in quality assessment programs: a property control chart, in which we record single measurements or the means for several replicate measurements, and a precision control chart, in which we record ranges or standard deviations. In either case, the control chart consists of a line representing the mean value for the measured property or the precision, and two or more boundary lines whose positions are determined by the precision of the measurement process. The position of the data points about the boundary lines determines whether the analysis is in statistical control.

#### Constructing a Property Control Chart

The simplest property control chart is a sequence of points, each representing a single determination of the property we are monitoring. To construct the control chart, we analyze a minimum of 7–15 samples while the system is under statistical control. The center line (*CL*) of the control chart is the average of these *n* samples.

\[CL=\overline X=\dfrac{\sum X_i}{n}\]

The more samples in the original control chart, the easier it is to detect when an analysis is beginning to drift out of statistical control. Building a control chart with an initial run of 30 or more samples is not an unusual choice.

Boundary lines around the center line are determined by the standard deviation, *S*, of the *n* points

\[S=\sqrt \frac{\large{\sum (X_i-\overline X)^2}}{\large{n-1}}\]

The upper and lower warning limits (*UWL* and *LWL*) and the upper and lower control limits (*UCL* and *LCL*) are given by the following equations.

\[UWL= CL + 2S \hspace{30px} LWL = CL - 2S\]

\[UCL= CL + 3S \hspace{30px} LCL = CL - 3S\]

Why these limits? Examine Table 4.12 in Chapter 4 and consider your answer to this question. We will return to this point later in this chapter when we consider how to use a control chart.

Example 15.4

Construct a property control chart using the following spike recovery data (all values are for percentage of spike recovered).

sample: | 1 | 2 | 3 | 4 | 5 |
---|---|---|---|---|---|

result: | 97.3 | 98.1 | 100.3 | 99.5 | 100.9 |

sample: | 6 | 7 | 8 | 9 | 10 |

result: | 98.6 | 96.9 | 99.6 | 101.1 | 100.4 |

sample: | 11 | 12 | 13 | 14 | 15 |

result: | 100.0 | 95.9 | 98.3 | 99.2 | 102.1 |

sample: | 16 | 17 | 18 | 19 | 20 |

result: | 98.5 | 101.7 | 100.4 | 99.1 | 100.3 |

*Solution*

The mean and the standard deviation for the 20 data points are 99.4% and 1.6%, respectively. Using these values, we find that the *UCL* is 104.2%, the *UWL* is 102.6%, the *LWL* is 96.2%, and the* LCL* is 94.6%. To construct the control chart, we plot the data points sequentially and draw horizontal lines for the center line and the four boundary lines. The resulting property control chart is shown in Figure 15.3.

**Figure 15.3** Property control chart for Example 15.4. The warning limits are shown in yellow and the control limits in red.

Practice Exercise 15.3

A control chart is a useful method for monitoring a glucometer’s performance over time. One approach is to use the glucometer to measure the glucose level of a standard solution. An initial analysis of the standard yields a mean value of 249.4 mg/100 mL and a standard deviation of 2.5 mg/100 mL. An analysis of the standard over 20 consecutive days gives the following results.

day: |
1 248.1 |
2 246.0 |
3 247.9 |
4 249.4 |
5 250.9 |
6 249.7 |
7 250.2 |
8 250.3 |
9 247.3 |
10 245.6 |
---|---|---|---|---|---|---|---|---|---|---|

day: result: |
11 246.2 |
12 250.8 |
13 249.0 |
14 254.3 |
15 246.1 |
16 250.8 |
17 248.1 |
18 246.7 |
19 253.5 |
20 251.0 |

Construct a control chart and evaluate the glucometer’s performance.

Click here to review your answer to this exercise.

We also can construct a control chart using the means for a set of replicate determinations on each sample. The mean for the *i*th sample is

\[\overline X_i=\dfrac{\sum_{j}X_{ij} }{n_\ce{rep}}\]

where *X*_{ij} is the *j*th replicate and *n*_{rep} is the number of replicate determinations for each sample.

When using means to construct a property control chart, all samples must have the same number of replicates.

The control chart’s center line is

\[CL=\dfrac{\sum \overline X_i}{n}\]

where *n* is the number of samples used in constructing the control chart. To determine the standard deviation for the warning limits and the control limits, we first calculate the variance for each sample.

\[s_i^2=\dfrac{\sum_{j}(X_{ij}-\overline X_i)^2}{n_\ce{rep}-1}\]

The overall standard deviation, *S*, is the square root of the average variance for the samples used in constructing the control plot.

\[S=\sqrt\frac{\large{\sum s_i^2}}{\large{n}}\]

The resulting warning and control limits are given by the following four equations.

\[UWL=CL+\dfrac{2S}{\sqrt{n_\ce{rep}}}\hspace{5mm}LWL=CL-\dfrac{2S}{\sqrt{n_\ce{rep}}}\]

\[UCL=CL+\dfrac{3S}{\sqrt{n_\ce{rep}}}\hspace{5mm}LCL=CL-\dfrac{3S}{\sqrt{n_\ce{rep}}}\]

#### Constructing a Precision Control Chart

A precision control chart shows how the precision of an analysis changes over time. The most common measure of precision is the range, *R*, between the largest and the smallest results for *n*_{rep} analyses on a sample.

\[R=X_\ce{largest}-X_\ce{smallest}\]

To construct the control chart, we analyze a minimum of 15–20 samples while the system is under statistical control. The center line (*CL*) of the control chart is the average range of these *n* samples.

\[\overline R=\dfrac{\sum R_i}{n}\]

The more samples in the original control chart, the easier it is to detect when an analysis is beginning to drift our of statistical control. Building a control chart with an initial run of 30 or more samples is not an unusual choice.

The upper warning line and the upper control line are given by the following equations

\[UWL=f_\ce{UWL}\times \overline R\]

\[UCL=f_\ce{UCL}\times \overline R\]

where *f*_{UWL} and *f*_{UCL} are statistical factors determined by the number of replicates used in determining the range. Table 15.2 provides representative values for *f*_{UWL} and *f*_{UCL}. Because the range is greater than or equal to zero, there is no lower control limit or lower warning limit.

replicates | f_{UWL} |
f_{UCL} |
---|---|---|

2 | 2.512 | 3.267 |

3 | 2.050 | 2.575 |

4 | 1.855 | 2.282 |

5 | 1.743 | 2.115 |

6 | 1.669 | 2.004 |

Example 15.5

Construct a precision control chart using the following ranges, each determined from a duplicate analysis of a 10.0-ppm calibration standard.

sample: | 1 | 2 | 3 | 4 | 5 |
---|---|---|---|---|---|

result: | 0.36 | 0.09 | 0.11 | 0.06 | 0.25 |

sample: | 6 | 7 | 8 | 9 | 10 |

result: | 0.15 | 0.28 | 0.27 | 0.03 | 0.28 |

sample: | 11 | 12 | 13 | 14 | 15 |

result: | 0.21 | 0.19 | 0.06 | 0.13 | 0.37 |

sample: | 16 | 17 | 18 | 19 | 20 |

result: | 0.01 | 0.19 | 0.39 | 0.05 | 0.05 |

*Solution*

The average range for the duplicate samples is 0.177. Because two replicates were used for each point the *UWL* and *UCL* are

\[UWL = 2.512 × 0.177 = 0.44\]

\[UCL = 3.267 × 0.177 = 0.58\]

The resulting property control chart is shown in Figure 15.4.

**Figure 15.4** Precision control chart for Example 15.5. The warning limits are shown in yellow and the control limits in red.

The precision control chart in Figure 15.4 is strictly valid only for the replicate analysis of identical samples, such as a calibration standard or a standard reference material. Its use for the analysis of nonidentical samples—as often is the case in clinical analyses and environmental analyses—is complicated by the fact that the range usually is not independent of the magnitude of the measurements. For example, Table 15.3 shows the relationship between the average range and the concentration of chromium in 91 water samples. The significant difference in the average range for different concentrations of chromium makes a single precision control chart impossible. As shown in figure 15.5, one solution is to prepare separate precision control charts, each of which covers a range of concentrations for which *R* is approximately constant.

[Cr] (ppb) | number of duplicate samples | R |
---|---|---|

5 to < 10 | 32 | 0.32 |

10 to < 25 | 15 | 0.57 |

25 to < 50 | 16 | 1.12 |

50 to < 150 | 15 | 3.80 |

150 to < 500 | 8 | 5.25 |

> 500 | 5 | 76.0 |

**Figure 15.5** Example showing the use of a precision control chart for samples that span a range of analyte concentrations. The precision control charts are for (a) low concentrations of analyte; (b) intermediate concentrations of analyte; and (c) high concentrations of analyte.

#### Interpreting Control Charts

The purpose of a control chart is to determine if an analysis is in a state of statistical control. We make this determination by examining the location of individual results relative to the warning limits and the control limits, and by examining the distribution of results around the central line. If we assume that the individual results are normally distributed, then the probability of finding a point at any distance from the control limit is determined by the properties of a normal distribution.^{5} We set the upper and lower control limits for a property control chart to *CL* ± 3*S* because 99.74% of a normally distributed population is within three standard deviations of the population’s mean. This means that there is only a 0.26% probability of obtaining a result larger than the *UCL* or smaller than the *LCL*. When a result exceeds a control limit, the most likely explanation is a systematic error in the analysis or a loss of precision. In either case, we assume that the analysis no longer is in a state of statistical control.

**Rule 1.** An analysis is no longer under statistical control if any single point exceeds either the *UCL* or the *LCL*.

By setting the upper and lower warning limits to *CL* ± 2S, we expect that no more than 5% of the results will exceed one of these limits; thus

**Rule 2.** An analysis is no longer under statistical control if two out of three consecutive points are between the *UWL* and the *UCL* or between the *LWL* and the *LCL*.

If an analysis is under statistical control, then we expect a random distribution of results about the center line. The presence of an unlikely pattern in the data is another indication that the analysis is no longer under statistical control.

**Rule 3. **An analysis is no longer under statistical control if seven consecutive results fall completely above or completely below the center line.

**Rule 4.** An analysis is no longer under statistical control if six consecutive results increase or decrease in value.

**Rule 5.** An analysis is no longer under statistical control if 14 consecutive alternate up and down in value.

**Rule 6.** An analysis is no longer under statistical control if there is any obvious nonrandom patter to the results.

Figure 15.6 shows three examples of control charts in which the results show an analysis that has fallen out of statistical control. The same rules apply to precision control charts with the exception that there are no lower warning limits and lower control limits.

**Figure 15.6** Examples of property control charts that show a sequence of results—indicated by the highlighting—that violate (a) rule 3; (b) rule 4; and (c) rule 5.

Exercise 15.4

In Practice Exercise 15.3 you created a property control chart for a glucometer. Examine your property control chart and evaluate the glucometer’s performance. Does your conclusion change if the next three results are 255.6, 253.9, and 255.8 mg/100 mL?

Click here to review your answer to this exercise.

#### Using Control Charts for Quality Assurance

Control charts play an important role in a performance-based program of quality assurance because they provide an easily interpreted picture of the statistical state of an analysis. Quality assessment samples such as blanks, standards, and spike recoveries are monitored with property control charts. A precision control chart is used to monitor duplicate samples.

The first step in using a control chart is to determine the mean value and the standard deviation (or range) for the property being measured while the analysis is under statistical control. These values must be established using the same conditions that will be present during subsequent analyses. Preliminary data is collected both throughout the day and over several days to account for short-term and long-term variability. An initial control chart is prepared using this preliminary data and discrepant points identified using the rules discussed in the previous section. After eliminating questionable points, the control chart is replotted. Once the control chart is in use, the original limits may be adjusted if the number of new data points is at least equivalent to the amount of data used to construct the original control chart. For example, if the original control chart includes 15 points, new limits can be calculated after collecting 15 additional points. The 30 points are pooled together to calculate the new limits. A second modification can be made after collecting an addition 30 points. Another indication that a control chart needs to be modified is when points rarely exceed the warning limits. In this case the new limits can be recalculated using the last 20 points.

Once a control chart is in use, new quality assessment data is added at a rate sufficient to ensure that the analysis remains in statistical control. As with prescriptive approaches to quality assurance, when the analysis falls out of statistical control, all samples analyzed since the last successful verification of statistical control must be reanalyzed. The advantage of a performance-based approach to quality assurance is that a laboratory may use its experience, guided by control charts, to determine the frequency for collecting quality assessment samples. When the system is stable, quality assessment samples can be acquired less frequently.