Laboratory Data Management Guidance
Laboratory Data Management Guidance
Technical Review:
Dr Phil Nethercote
On behalf of the ECA Analytical Quality Control Working Group
Approved by:
Dr Günter Brendelberger
On behalf of the ECA Analytical Quality Control Working Group
Table of Contents
Document Revision History ...................................................................................................................... 5
Scope & Application ................................................................................................................................. 6
Expert Drafting Group .............................................................................................................................. 7
Regulatory References ............................................................................................................................. 8
Overview of Laboratory Data Management & the Analytical Process......................................................... 9
QU involvement/Responsibilities ........................................................................................................... 10
Overview & purpose of trend analysis .................................................................................................... 10
Control Charting Concept .................................................................................................................................. 12
Detecting and Managing OOE results ...................................................................................................... 16
Introduction ....................................................................................................................................................... 16
Unexpected Variation in Replicate Determinations .......................................................................................... 16
Unexpected Results in a Single Test or a Small Set of Tests.............................................................................. 17
Trend Analysis for Statistical Process Control .......................................................................................... 19
Overview............................................................................................................................................................ 19
Control of continuous data................................................................................................................................ 19
Determination of a Trend using Statistical Process Control (SPC) .................................................................... 21
Control of continuous data................................................................................................................................ 21
I-Moving Range (MR) Control Charts ................................................................................................................ 22
The Individuals chart control limits ................................................................................................................... 23
The MR chart control limits ............................................................................................................................... 23
The R Chart control limits .................................................................................................................................. 24
The S Chart control limits .................................................................................................................................. 24
The X-bar chart control limits ............................................................................................................................ 24
Normality assumption ....................................................................................................................................... 25
CuSum & EWMA charts ..................................................................................................................................... 26
CuSum charts ..................................................................................................................................................... 26
EWMA ................................................................................................................................................................ 27
Case 1: σ slope
2
= 0 ............................................................................................................................................. 64
Case 2: σ int,slope
2
≥ 0 .......................................................................................................................................... 65
Case 3: Non-linearity ........................................................................................................................................ 66
Data sets for RCR Examples ............................................................................................................................... 68
Case 1 ................................................................................................................................................................ 68
Case 2 ................................................................................................................................................................ 69
Case 3 ................................................................................................................................................................ 70
This guideline applies primarily to physicochemical -based laboratory testing resulting in continuous data
(variables, for example assay, impurity values, hardness etc., which may be assumed to be normally
distributed) or discrete data (attributes for example, particle counts, identity tests or cosmetic quality
defects derived from AQLs which are not normally distributed). However, for discrete data, it may also be
applicable to the microbiological laboratory.
Laboratory tests are performed on active pharmaceutical ingredients, excipients and other components, in-
process materials, and finished drug products. It is applicable to PAT (Process Analytical Technology) or RTR
(Real Time Release) approaches. This SOP is complementary to, and should be used in conjunction with, the
ECA SOP on OOS Results1
If a number of measurements are made over a short period of time and an anomalous or unexpected value
is observed within these measurements then it is designated OOE (Out of Expectation). An OOE is defined as
a parameter value which lies outside the expected variation of the analytical procedure variation with
respect to either location or dispersion.
A trend can occur in a sequence of time related events, measurements or outputs. Trend analysis refers to
techniques for detecting an underlying pattern of behaviour in a time or batch sequence which would
otherwise be partly or nearly completely hidden by noise. These techniques enable specific behaviours
(OOT; Out of Trend) such as a shift, drift or excessive noise to be detected.
1. Where the expectation is that there will be no trend, for example for production or analytical
process data which are known or assumed to be under statistical control.
or
2. Where the expectation is that there is will be trend; for example in stability testing.
There is a fundamental difference between these two situations in that the variance increases with time in
the second situation.
Therefore in this guideline there are three distinct sections covering OOE and the two types of OOT. Each
section is supported by examples given in the appendices. The methods used in examples are intended to
be advisory as to represent recommended practice but should not be mandatory. Other statistically sound
procedures may be used as alternatives.
1
STANDARD OPERATING PROCEDURE Laboratory Data Management; Out of Specification (OOS) Results, Version 2, 14-
Aug-2012
Regulatory References
9. PIC/S Draft Guidance PI041-1 Good Practices for Data Management and Integrity in Regulated
GMP/GDP Environments; 10th August 2016
Laboratory data quality management processes are a part of the overall Quality Management System as
required by Chapter 1 of EU GMP and the FDA cGMPs as specified in 21 CFR §210 & §211.
Analytical processes and procedures are managed as part of a lifecycle concept. Laboratory data integrity
and security are critical requirements under the GMPs. Such a process is illustrated below.
The purpose of this guidance document is to define the procedures for managing laboratory data which are
Out-of-Expectation (OOE) or Out-of-Trend(OOT). Any confirmed OOE or OOT should trigger a deviation and
appropriate investigation. The investigation should follow the principles laid down in the Out-of-
Specification (OOS) SOP, ECA_AQCWG_SOP 01. This guidance document does not cover the evaluation of
trend data with respect to specification. Process capability is mentioned briefly but the details are a topic
beyond the scope of this document.
The pharmaceutical industry lags far behind many other manufacturing industries in the area of process
evaluation and control. This guidance document is intended to assist in the simple implementation of
trending techniques to meet regulatory requirements particularly in the areas of Product Quality Review
(EU) and Annual Product Review (US).
In 1960, Dr Genichi Taguchi introduced a new definition of "World Class Quality" namely;
On target with Minimum Variance
This contrasts with the traditional Conformance with Specification, previously adopted by the FDA and other
authorities.
Indeed it is not in technical accordance with the principles of Six Sigma which allowed the mean to vary
±1.5σ.
However, his revolutionary definition ensured that the application of statistical process control techniques
was in the forefront of the tools required to achieve this life cycle objective.
QU involvement/Responsibilities
Quality Control testing is considered an integral part of the Company's Quality Unit as explicitly required by
EU GMP. Formal Quality involvement, e.g. by a separate QA function, should be kept to the minimum
consistent with US & EU regulatory expectations and requirements based upon published legislation and
guidelines.
The extent of Quality oversight is very dependent on individual company requirements. Organisation and
nomenclature of Quality Control and Assurance functions and assignment of responsibilities are also highly
company specific. This Guideline does not dictate or recommend specific steps that must be supervised by
specific quality functions other than those required by regulation. Therefore the term Quality Unit (QU) as
used in the revised chapter 1 of EU GMP Guide, is used here.
The initial OOE or OOT investigation, however, should be performed directly under the responsibility of the
competent laboratory.
The approaches set out in this guidance document are dependent on the applicable shape (mathematical
distribution model) of the data. The data types under consideration here are variables and attributes. A
continuous random variable is the one which can take any value over a range of values for example an assay
value or an impurity level. An attribute is an integer where the set of possible values for a discrete random
variable is at most countable for example a cosmetic defect on a tablet or the number of particles in a
solution.
Hence the selection of the appropriate mathematical distribution may be shown as a decision tree. For example
2
Figure 1 Decision tree for the selection of an appropriate mathematical data model based on data shape
For our purposes, the most useful distribution for continuous variables is the Normal or Gaussian
distribution for a population whose properties are well known.
For a true mean value (µ) of zero and a standard deviation (σ) of 1 then the probability distribution is given by
− ( x − µ )2
1
2σ 2
y= e (1.1)
2πσ
and shown graphically in Figure 2. The areas under the curve indicate the probability of values lying ±σ, ±2σ and ±3σ
from the mean. This distribution is the basis for control charting of continuous random variables and stability trending
as discussed later.
2
Adapted and redrawn from a paper by a Prof Aswath Damodaran at the Stern School of Business at New York
University http://people.stern.nyu.edu/adamodar/New_Home_Page/home.htm
68.27%
95.45%
99.73%
For attribute data, the Binomial or Poisson distributions are preferred. If counted defects are to be used, the
Binomial distribution is used. If the data are defects expressed as a % for example then the use of the
Poisson distribution is indicated.
UAL
+3σ
UWL
RESPONSE VARIABLE
+2σ
+1σ
P=95.45%
P=68.27%
P=99.73%
Mean
-1σ
-2σ LWL
-3σ
LAL
TIME VARIABLE
Figure 3 Idealised control chart for a continuous variable under the normal distribution
The approach is based on the idea that, no matter how well a process is designed, there exists a certain
amount of natural variability in output measurements. When the variation in process quality is due to
random causes alone, the process is said to be statistically in-control. If the process variation includes both
random and special causes of variation, the process is said to be statistically out-of-control.
All test results have variation that comes from measurement (system) variation and process performance
variation. There are two types of variation; Common Cause variation, inherent noise, and Special Cause
variation owing to, for example, a process shift, drift or excessive noise .
The control chart is designed to detect the presence of special causes of variation. The normal distribution
may be characterised by two particular parameters; a measure of location (the arithmetic mean or average)
and a measure of dispersion (the standard deviation). If a process is unstable it means that either of these
parameters are changing in an uncontrolled manner (Figure 4 (a)). This would be apparent from a mean and
range control chart for example. The next task would be to bring these two parameters into a state of
statistical control. This would entail ensuring that the mean and the standard deviations were not varying
significantly. This ideal situation is illustrated in (Figure 4 (b)). This would then said to be under statistical
control i.e. no special cause variation and controlled common cause variation. In this state, the process is
amenable to the tools of Statistical Process Control (SPC). However, a stable process may not be statistically
capable of meeting the specification limits. Figure 4 (c) illustrates this showing that the red process albeit
stable is incapable. The desired state is, of course, to arrive at the blue capable state. The method of
calculating process capabilities are briefly described later in this guidance.
INCAPABLE
SPECIFICATION
LIMITS
(a) An unstable process (b) A stable process (c) Stable Processes; Capable
and In Capable
3
Figure 4 Process stability & capability
The question is how are we to judge when a process is in a state of statistical control with respect to time?
3
Redrawn and based on QMS – Process Validation Guidance, GHTF/SG3/N99-10:2004 (Edition 2) Annex A Statistical
methods and tools for process validation [http://www.imdrf.org/documents/doc-ghtf-sg3.asp]
The answer lies in the application of SPC decision rules. These are based on the patterns expected from the
distribution shown in Figure 3. These rules were developed many years ago and the simplest are the four
WECO rules4.
Figure 5 The 4 basic WECO rules for detecting out of trend (OOT) results
More recently, an extended set of 8 rules. the Nelson Rules5, have been proposed. These rules are
incorporated within many standard software control charting applications such as Minitab or SAS JMP for
example. The choice of rule selection is left to the user. It is not recommended to select all rules as this
increases the likelihood of false trends being identified. Quite often, the 4 basic WECO rules are sufficient.
4
Western Electric Company (1956), Statistical Quality Control handbook. (1 ed.), Indianapolis, Indiana: Western Electric
Co or see Montgomery, Douglas C. (2009), Introduction to Statistical Quality Control (6 ed.), Hoboken, New Jersey: John
Wiley & Sons
5
Lloyd S. Nelson, "Technical Aids," Journal of Quality Technology 16(4), 238-239, (October 1984)
Rule 1 One point is more than 3 standard deviations from One sample is grossly out of control
the mean
Rule 2 Nine or more points in a row are on the same side Some prolonged bias exists
of the mean
Rule 3 Six or more points in a row are continually A trend exists. This is directional and the position of the
increasing or decreasing mean and size of the standard deviation do not affect this
rule.
Rule 4 Fourteen or more points in a row alternate in This much oscillation is beyond noise. This is directional
direction, increasing then decreasing. and the position of the mean and size of the standard
deviation do not affect this rule.
Rule 5 Two or three out of three points in a row are more There is a medium tendency for samples to be out of
than 2 standard deviations from the mean in the control.
same direction.
Rule 6 Four (or five) out of five points in a row are more There is a strong tendency for samples to be slightly out of
than 1 standard deviation from the mean in the control.
same direction
Rule 7 Fifteen points in a row are all within 1 standard With 1 standard deviation, greater variation would be
deviation of the mean on either side of the mean expected
Rule 8 Eight points in a row exist with none within 1 Jumping from above to below whilst missing the first
standard deviation of the mean and the points are standard deviation band is rarely random
in both directions from the mean.
Introduction
The terms “Out of Specification Result” and “Out of Trend Result” are well defined, e.g. in the UK Medicines
and Healthcare Products Regulatory Agency (MHRA Guidance “Out Of Specification Investigations”, detailing
the MHRA expectations, second version, issued 2013:
Out-of-Specification (OOS) Result – test result that does not comply with the pre-determined
acceptance criteria (i.e. for example, filed applications, drug master files, approved marketing
submissions, or official compendia or internal acceptance criteria)
Out of Trend (OOT) Result – a stability result that does not follow the expected trend, either in
comparison with other stability batches or with respect to previous results collected during a
stability study. However, trends of starting materials and in-process samples may also yield out of
trend data. The result is not necessarily OOS but does not look like a typical data point. Should be
considered for environmental trend analysis such as for viable and non viable data action limit or
warning limit trends.
This definition is extremely focused on stability studies, however, mentioning environmental trend analysis
indicates that OOT results may also be observed during trend analysis for statistical process control.
However, no formal definition is given for the term “Out of Expectation Result”. In contrast to OOS results
this is not linked to a violation of a formal specification and in contrast to OOT results this is not statistically
deducible from a data base comprehensive enough to allow calculation whether the result belongs to a
population to be expected from the analytical procedure’s uncertainty or not. This might be possible
starting from a number of 30 independent tests.
Usual analytical practice will use a specific number of replicates - that is several discrete measurements -
to provide more accurate results. These may be either replicate injections from the same HPLC sample
preparation, replicate readings or other multiple determinations. This procedure has to be specified in the
written, approved test procedure together with the limits for variability (range and/or RSD) among the
replicates. These could be based upon the process capability of the method as determined during the
method development and its subsequent validation. However, usually companies use a general limit of the
range of Δ ≤ 2.0 % for assays. In case of replicate series of complete tests (full run-throughs of the test
procedure) wider limits for variability among the replicates may be defined.
Any unexpected variation in replicate determinations - either derived by multiple measurements of one
sample preparation or replicate series of complete tests - disqualifies this data set from being used for
further result calculation. E.g. if the range between replicates is limited to Δ ≤ 2.0 % and the two replicates
differ by 2.2 %, data generated from the analysis cannot be used. It is very important that the
documentation accompanying the analysis is very clear about why the data sets have been rejected.
If only one set of data within a bigger data pool is affected - e.g. one out of several samples and reference
samples tested over night using an automated HPLC system - only the directly affected replicates are
considered disqualified, all other data in the series may be further processed to calculate the results of the
other samples.
When unexpected variation in replicate determinations occurs, investigation into the cause is required
similar to an investigation in the case of a non-compliant system suitability (SST) test. Usually this is
reported as a laboratory deviation. The flow of the investigation may follow the proven approach of
investigating an OOS result on a lab scale.
Repeating the test or measurement- preferable using the same sample preparation if appropriate - should
not be performed prior to identifying a hypothesis why the replicates range was higher than expected and
having taken corresponding actions.
Analytical results from one single performance of one test or from a small number of tests obtained over a
short period of time may be considered "Out of Expectation" if
The test result does not fit into the other results of that series, but the number of tests and data
points is not comprehensive enough to allow statistical calculation whether the result belongs to a
population to be expected from the mean and the variability of the overall data set.
The result does not violate a given specification
There is enough evidence and information allowing to anticipate the "expected" result and thus to
allow judgement that the result does not represent the expectations.
Analytical results of the same sample or the same material using another, validated analytical
procedure (e.g. IPC testing of a compounded bulk product, using an UV assay procedure and a later
testing of the filled product using HPLC)
Knowing the theoretical composition of the sample (e.g. samples prepared during galenic
development)
Results of tests of other samples/batches within a campaign or series of experiments(e.g. results of
three out of four batches in one campaign are close to the theoretical assay, one is close to a
specification limit)
To decide, whether a result is really out of expectation or may be considered representing the typical
variability of the procedure applied, data of the analytical validation of the procedure used should be used.
According to the concept of analytical uncertainty usually applied in chemical analysis, the combined
standard uncertainty of the result would be the appropriate performance indicator to help deciding,
whether the result in question really is "unexpected" or simply represents a rare, but still probable value.
As analytical uncertainties of pharmaceutical test procedures are rarely established, a common way to
estimate this range may be used.
In case an assay procedure based on HPLC has a reported (and correctly determined) intermediate precision
of 0.8 %, the expanded analytical uncertainty to be expected in later routine application of the procedure is
1.2 % RSD.
To determine the limits (based on a 95 % confidence level) within which analytical results are representing
the analytical variability of the procedure to be expected and accepted, the following calculation has to be
performed;
In the example, any analytical result falling within a range of ± 2 x 1.2 % = ± 2.4 % of the anticipated result
are representing analytical variability of the procedure on a 95 % confidence level and have to be accepted
as is.
Only results falling outside this range are to be
considered "out of Expectation".
−1σ + 1σ
In this case, data should not be used and accepted
68.2 6%
without previous investigation to determine the cause
for the unexpected discrepancy from the anticipated
result. This investigation should follow the well
− 2σ 95.46% + 2σ established process of laboratory investigations in case
− 3σ 99.74% + 3σ of OOS results.
6
B Renger, Journal Chromatography B, 745 (2000), 167 - 176
As mentioned earlier, all measurements have variation. There are two types of variation.
1. Common Cause variation or noise
2. Special Cause variation such as process shift, drift or excessive noise.
The purpose of a control chart is to detect Special Cause variation. The expectation for a process is that it is
under statistical control i.e. the only component of the variation is the test result noise.
Quality Control (QC) plays an essential role in the Pharmaceutical and Biopharmaceutical industries and
associated processes. A large part of QC focuses on tracking the ongoing performance of a process to spot
problems or identify opportunities for improvement. An ideal quality control system will highlight the
approach of trouble even before it becomes a problem. A number of statistical and graphical techniques
exist for tracking ongoing quality performance.
Under certain circumstances, if not investigated and or corrected, an OOT may lead in time to an OOS and
therefore an identification of an OOT may be an early indicator of a potential future OOS and so facilitate
action being taken designed to reduce the incidence of non-random OOS results.
Thus the generation of trended data is an essential tool in the management of manufacturing and quality
issues. These processes may only be effective where there is a suitable control strategy in place.
A control strategy is a planned set of controls, derived from current product and process understanding, that
ensures process performance and product quality. These controls can include parameters and attributes
related to drug substance and drug product materials and components, facility and equipment operating
conditions, in-process controls, finished product specifications, and the associated methods and frequency
of monitoring and control.
A typical control strategy for a Product Quality and Process Performance life cycle in the pharmaceutical
industry today may consist of the following elements:
An Out of Trent (OOT) result is a non-random event that is identified as test result or pattern of results that
are outside of pre-defined limits. For continuous data evaluation, this guideline recommends using simple
Shewhart type control charts in the first instance. These control charts developed in the 1930s have been
widely applied in engineering and manufacturing industries.
These control chart use data that is collected in an appropriate manner and then applied to the standard or
ideal result based upon historical data. The centre line on any control chart represents the mean (average)
of the values collected during a reference period.
One (or more) line(s) is positioned both above and below the centre line to serve as control limits. These
limits, the Upper Control Limit and the Lower Control Limit (UCL and LCL), provide a range of what is still
acceptable for a result. Control charts are therefore used to determine if the results that are coming in are
within the limits of what is acceptable or if the process is out of control. These upper and lower control
limits must, wherever possible, be based on the values determined for the Proven Acceptable Range (PAR)
and Normal Operating Range (NOR) for a process.
In investigational circumstances it may be required to analyse historical data to see if there have been
special cause variations. In this instance a post mortem CuSum approach is to be recommended
CuSum stands for "cumulative sum." A CuSum chart is related to a standard control chart and is made in
much the same manner, except that the vertical axis is used to plot the cumulative sum of the variability
(differences between successive values) in a process. This CuSum is plotted on the vertical (Y) axis against
time on the horizontal (X) axis.
This type of plot is helpful in spotting a biased process, in which the process misses the calculated mean
value high or misses it low, since repeated misses on one side of the ideal value will force the cumulative
sum away from the ideal value or benchmark value (which may be zero ) which is the ideal low variance (no
variance) objective.
The minimum number of data values from which a suitable statistical mean can be calculated for use in a
CuSum chart is 10 individual values. The maximum number of values to limit variation in the data set, should
be set at 30 to 100 data values.
Measurements of all outcomes and processes will vary over time but variation is often hidden by current
practices in data management, where data is aggregated (averaged) and presented over long time periods
(e.g. by quarter). Plotting data continuously (weekly or monthly) can be very informative. If we do this we
reveal the sources and extent of variation.
When dealing with a quality characteristic that is a variable we want to make sure that the characteristic is
under control.
Shewhart identified two sources of process variation: common cause variation (chance variation) that is
inherent in process, and stable over time, and special cause variation (assignable, or uncontrolled variation),
which is unstable over time - the result of specific events outside the system.
A process that is operating only with common causes of variation is said to be in statistical control. A
process that is operating in the presence of assignable causes is said to be out of control. The eventual goal
of SPC is the elimination of variability in the process.
The control chart was designed so that one could distinguish between common and special causes of
variation within a process and to provide a rule for minimizing the risk of reacting to a special cause when it
is in fact a common cause, and not reacting to a special cause when one is present. It allows visualizing
variations that occur in the central tendency and dispersion of a set of observations
A typical control chart has control limits set at values such that if the process is in control, nearly all points
will lie between the upper control limit (UCL) and the lower control limit (LCL).
A control chart is typically constructed as follows:
UCL = µW + Lσ W
Centre Line =µW (1.2)
UCL = µW − Lσ W
where L = a constant multiplier which will define the distance of the control limit from the centre line
= mean of the sample statistic, W.
= standard deviation of the statistic, W.
When the assignable causes are eliminated and the points plotted are within the control limits, the process
is in state of control. Further improvement can be obtained through changing basic process, system.
Depending on the data than can be collected and on the purpose (detect small shift or large shift,
investigation or continuous process verification), different control charts can be used. The following
flowchart gives an indication of which chart to use when.
Not within the
scope of this
No Are process data Yes guideline
autocorreleated?
Seek statistical
help
X &R CUSUM
X
CUSUM p
CUSUM
c
CUSUM
(individuals) EWMA EWMA
EWMA EWMA np u
X& s MR using p using c, u
MRi = xi − xi −1 (1.3)
When data are collected in subgroups (eg, several determinations on the same batch), the X-bar control
chart for subgroups means is being used. It is usually presented along with R-charts or S-charts. The R-chart
plots subgroup ranges (when subgroup sample size <9), and the S-chart plots subgroup standard deviations
(when subgroup sample size >=9).
Suppose m samples are available, each containing n observations. Let ̅ , ̅ , ⋯ , ̅ be the average of each
sample, then the estimator of the process average is
̅ + ̅ +⋯+ ̅
̿=
Let = |max − min |, the range for group i, i=1, …, m. Then the average range is:
+ + ⋯+
"=
LCL = x − A2 R
where the constant # is tabulated for various sample sizes in Table 2.
UCL = D4 R
Centre Line = R (1.8)
LCL = D3 R
Where " is the sample average range and the constants $% and $& are tabulated for various sample sizes in
Table 2
UCL = B4 s
Centre Line = s (1.9)
LCL = B3 s
Where the constants (% and (& are tabulated for various sample sizes in Table 2
Also the parameters of the X-bar chart can be adapted to include '̅, instead of " .
LCL = x − A3 s
STANDARD
AVERAGES MEAN and RANGE
DEVIATIONS
# of
Observations A2 A3 B3 B4 d2 D3 D4
(n)
2 1.880 2.659 0 3.267 1.128 0 3.267
3 1.023 1.954 0 2.568 1.693 0 2.574
4 0.729 1.628 0 2.266 2.059 0 2.282
5 0.577 1.427 0 2.089 2.326 0 2.114
6 0.483 1.287 0.030 1.970 2.534 0 2.004
7 0.419 1.182 0.118 1.882 2.704 0.076 1.924
8 0.373 1.099 0.185 1.815 2.847 0.136 1.864
9 0.337 1.032 0.239 1.761 2.970 0.184 1.816
10 0.308 0.975 0.284 1.716 3.078 0.223 1.777
11 0.285 0.927 0.321 1.679 3.173 0.256 1.744
12 0.266 0.886 0.354 1.646 3.258 0.283 1.717
13 0.249 0.850 0.382 1.618 3.336 0.307 1.693
14 0.235 0.817 0.405 1.594 3.407 0.328 1.672
15 0.223 0.789 0.428 1.572 3.472 0.347 1.653
16 0.212 0.763 0.448 1.552 3.532 0.363 1.637
17 0.303 0.739 0.466 1.534 3.588 0.378 1.622
18 0.194 0.718 0.482 1.518 3.640 0.391 1.608
19 0.187 0.698 0.497 1.503 3.689 0.403 1.597
20 0.180 0.680 0.51 1.490 3.735 0.415 1.585
21 0.173 0.663 0.523 1.477 3.778 0.425 1.575
22 0.167 0.647 0.534 1.466 3.819 0.434 1.566
23 0.162 0.633 0.545 1.455 3.858 0.443 1.577
24 0.157 0.619 0.555 1.445 3.895 0.451 1.548
25 0.153 0.606 0.565 1.435 3.931 0.459 1.541
Normality assumption
A common assumption when constructing control charts for continuous data (individuals or X-bar) is that
data follows a normal distribution. The normality should be tested before using these charts. A common
way to check for normality is to visually inspect the histogram and the quantile-quantile plot, as well as to
conduct a normality test. The most used normality test is the Shapiro-Wilk test. If data are not normally
distributed, a deeper understanding of the non-normality is necessary: are there outliers, are there trends in
the data, are there two populations or is it another distribution? Often, data might be log-normally
distributed, in which case a logarithmic transformation is necessary in order to normalize the data. Another
common transformation is the reciprocal one, 1/X. The control charts should be constructed on the
transformed data.
CuSum is short for Cumulative Sums. As measurements are taken, the difference between each
measurement and the bench mark value/process target ( ) is calculated, and this is cumulatively summed
up (thus CuSum):
* = +, - − ).
-/
If the processes are in control, measurements do not deviate significantly from the bench mark, so
measurements greater than the bench mark and those less than the bench mark averaged each other out,
and the CuSum value should vary narrowly around the bench mark level. If the processes are out of control,
measurements will more likely to be on one side of the bench mark, so the CuSum value will progressively
depart from that of the bench mark.
CuSum can be used as a ‘post-mortem’ analysis of historical data, that may allow to determine the cause of
unexpected changes in result.
EWMA
Exponentially Weighted Moving Average (EWMA) chart, also referred to as a Geometric Moving Average
(GMA) chart are a good alternative to the Shewart control chart when we want to detect small shifts. It acts
in the same way as a CuSum chart.
Each point on a EWMA chart is the weighted average of all the previous subgroup means, including the
mean of the present subgroup sample. The weights decrease exponentially going backward in time.
0 =1 + 1−1 03
Where 0 < 1 ≤ 1 is a constant and the starting value is the process target:
0) = )
If 7 is close to 0, more weight is given to past observations. If 7 is close to 1, more weight is given to present
information. When 7=1, the EWMA becomes the Individuals control chart. Typical values for 7 are less
than 0.25.
Specification limits are used to evaluate process capability enabling a measure of how well the product
meets customer expectations. Control limits are used to evaluate process stability. Unstable processes
generally lead to failure to meet customer expectations.
Process capability refers to the performance of the process when it is operating in control. Two capability
indices are usually computed: JK and JKL . JK measures the potential capability in the process, if the process
was centred (it does not take into account where the process mean is located relative to the specifications),
while JKL measures the actual capability in a process (process can be off-centre). If a process is centred, then
JK = JKL .
MNO − ONO
JK =
6
STU3V V3UTU
JKQ = and JKR =
%W %W
X"
Where σ is estimated either by when variables control charts are used in the capability studies or by the
YZ
sample standard deviation s.
Typical values for JK and JKL are 0.5 or 1 for not capable processes, 1.33 and 1.67 for capable processes and
>2 for highly capable processes.
An important assumption underlying the interpretation of JK and JKL is that the process output follows a
normal distribution. If data are not normally distributed, one can transform the data to normalize it. Then
work with the transformed data (and specifications!) to compute the indices. Commonly used
transformations are logarithmic, ln(X), or reciprocal, 1/X.
Whenever the measured quantities for one item are not continuous but rather quality characteristics or
count data, control chart for discrete data should be used. Usually, one would classify the inspected item
into “conforming item” or “nonconforming item”. A nonconforming item is a unit of product that does not
satisfy one or more of the specifications of the product (it contains at least one nonconformity). If more
than one defect can be observed on the same unit, one can be interested in the number of nonconformities
(defects) per unit, instead of the fraction nonconforming for a single nonconformity (defect).
One can construct control charts for either fraction of non-conforming (p-chart) or total number of non-
conforming units, if sample size is the same (np-chart)
Suppose m samples of sample size ni are available , let [" be the average sample size:
1
[" = +[
/
If the sample size is the same for each group, then [" = [.
The sample fraction nonconforming for sample i is defined as the ratio of the number of non-conforming
units in the sample i, Di, to the sample size ni.
$
=
[
Suppose m samples are available, than the average fraction nonconforming is:
∑/
̅=
The distribution of the random variable can be obtained from the binomial distribution.
p(1 − p)
LCL = p − 3
n
Depending on the values of ̅ and ni, sometimes the lower control limit LCL<0. In these cases, we set LCL=0
and assume the control chart only has an upper limit.
nP-charts
If the sample sizes for all samples are equal, one can also construct the control chart for the number
nonconforming (np-control chart) instead of the fraction non-conforming.
LCL = n p − 3 n p (1 − p )
Many commercial statistical programmes will produce np control charts with variable control limits based
upon n.
Discussion
If sample size is too large with respect to the number of nonconforming units (eg., 20 nonconforming units
out of 500000), than the p-chart will not work properly because the control limits are inversely proportional
to the sample size. Therefore they became very small and process will look out of control, as data plotted on
the control chart will be out of control limits. If the sample size is the same (or approximately the same),
one could use the individuals control charts instead, where one would plot the number of nonconforming
units. If the sample size is significantly different from one sample point to another, then one could use a
Laney p-chart7.
Over dispersion exists when there is more variation in your data than you would expect based on a binomial
distribution (for defectives) or a Poisson distribution (for defects). Traditional P charts and U charts assume
that your rate of defectives or defects remains constant over time. However, external noise factors, which
are not special causes, normally cause some variation in the rate of defectives or defects over time. Under
dispersion is the opposite of over dispersion. Under dispersion occurs when there is less variation in your
data than you would expect based on a binomial distribution (for defectives) or a Poisson distribution (for
defects). Under dispersion can occur when adjacent subgroups are correlated with each other, also known
as autocorrelation. For example, as a tool wears out, the number of defects may increase. The increase in
defect counts across subgroups can make the subgroups more similar than they would be by chance.
When data exhibit under dispersion, the control limits on a traditional P chart or U chart may be too wide. If
the control limits are too wide, you can overlook special cause variation and mistake it for common cause
variation.
7
David B. Laney Quality Engineering, 14(4), 531–537 (2002) and see also, for example, Chin-liang Hung, M.S
dissertation from Iowa State University, Control Charts for Attributes: Some Variations, 1997,
http://www.public.iastate.edu/~wrstephe/HungCC.pdf
One can construct control charts for either the total number of nonconformities in a unit (c-chart) or the
average number of nonconformities per unit (u-chart)
When we have a constant sample size, n, of inspection units from one sample to another, one can work with
the total number of nonconformities per sample and construct the c-chart. The total number of
nonconformities in a unit is represented on the chart:
a
] ∑`/ x_`
c^=
where xij is the number of defects for inspection unit i in sample j. The total nonconformities in a sample
follow a Poisson distribution.
LCL = c − 3 c
where *̅ is the observed average number of non-conformities in a preliminary sample of m inspection units,
n is the constant sample size and - is the number of defects for inspection unit i:
∑b/ *̅
*̅ =
If LCL yields a negative value, than LCL is fixed to 0.
U-Charts
If the sample size is not constant and can vary from one sample to another, then one should work with the
average number of nonconformities per unit of product instead of total number of nonconformities per
sample and the u-chart is to be used, instead of a c-chart. Let the average number of nonconformities per
unit be
∑a`/c -
" =
[
Where - is the total nonconformities in a sample of [ .
u
UCL = u + 3
n
Center Line = u (1.15)
u
LCL = u − 3
n
Where " represents the observed average number of nonconformities per unit in a preliminary data set of m
inspection units, n is the sample size of the current inspected sample:
∑/ "
"=
Overview
The purpose of this section is to provide guidance for generating, maintaining and monitoring trends of
stability data by establishing trend limits calculated from existing historical stability data for pharmaceutical
products stored at the recommended storage condition.
The purpose of trend analysis for stability data should be to detect if;
• a batch is out of trend with respect to historical batches
and
• one or more observations is out of trend within a batch
Although there are numerous approaches to trending stability data, two different models are presented
here to generate stability trends. These two approaches that may be used are a simple linear regression
model and the more sophisticated Random Coefficients Regression Model. These models are used to
understand the degradation rates over time to support the expiration dating of the product.
To see if a specific batch is out of trend, a comparison of the slope of the batch under study with the slopes
of the historical batches should be performed. A poolability test8 may be used for this comparison, or
improved statistical description of the historical behaviour and detection of an OOT batch can be obtained
by estimating the slope of the historical batches and the new batch via the Random Coefficients Regression
Model (with a fixed effect being the type of batch: historical or under study) and then use contrasts to make
a test whether the difference between the slope of the historical batches and the slope of the new batch is
different from zero or not. If the difference is significantly (at 0.05) different from zero, then the new batch
is considered to be OOT. A minimum number of observations per batch needed for this analysis should be
defined (e.g. 3 or 4 observations to determine a meaningful statistical trend based on product history and
measurement variability, as 2 may not be sufficient; however, 2 time points may be sufficient to highlight an
OOE).
To see if one observation is out-of-trend, a prediction interval for the batch under study should be
constructed (without the observation under study), taking into account the variability from the historical
batches (via a common error model between historical+batch under study). If the observation is outside the
prediction interval, than it is considered as OOT. The data set must include a minimum of 3 lots with at least
4 time points per lot to start this analysis.
8
For example as described in ICH Q1E
Data from multiple configurations maybe combined if there is a technical rationale or if equivalency of the
configurations can be demonstrated.
Generate a preliminary graph of the test result versus storage time. Any data point that appears atypical or
discrepant might be removed from the data set. Any data point removed must be identified and its removal
justified in a final report.
The principles discussed here are in accordance with WHO TRS No. 953; Annex 2 Stability testing of active
pharmaceutical ingredients and finished pharmaceutical products, 2009 and the ICH Guidance Q1A(R2)
Stability Testing of New Drug Substances and Product.
The following papers should also be consulted for more detailed information;
1. Identification of Out-of-Trend Stability Results, A Review of Potential Regulatory Issue and Various
Approaches, Pharma CMC Statistics and Stability Expert Teams, Mary Ann Gorko, Pharmaceutical
Technology, 27 (4), 38–52, 2003
2. Identification of Out-of-Trend Stability Results, Part II PhRMA CMC Statistics, Stability Expert Teams,
Pharmaceutical Technology, 2006
3. Methods for Identifying Out-of-Trend Results in Ongoing Stability Data, Adrijana Torbovska and
Suzana Trajkovic-Jolevska, Pharmaceutical Technology Europe, June, 2013
4. Carter, R. L. and Yang, M. C. K. (1986). “Large Sample Inference in Random Coefficient Regression
Models.” Communication in Statistics Theory and Methods 15(8), 2507-2525
5. Chow, Shein-Chung, Statistical Design and Analysis of Stability Studies, Chapman & Hall/CRC
Biostatistics Series, Boca Raton Fl, 2007
6. Dempster, A. P., Rubin, D. B. and Tsutakawa, R. K. (1981). “Estimation in Covariance Component
Models.” Journal of the American Statistical Association 76, 341 – 353
7. Laird, N. M. and Ware, J. H. (1982). “Random Effects Model for Longitudinal Data.” Biometrics 38,
963 – 974
8. Searle, R. Shayle, Casella, G., McCulloch, C. E., Variance Components, John Wiley & Sons, Inc., New
York, 1992
It is recommended that a Stability Subject Matter Expert (SME) advises on steps to be taken in case of
insufficient and/or inconclusive stability data. SME must have sufficient education, training, and specific
experience to provide such advice. The SMEs are required to have a good understanding of the stability
data, analytical methods, as well as the strength, quality, identity, and purity of the product. In addition, a
professional statistician can provide specific information and advice on statistical problems that arise in
execution of procedures discussed in this Guideline.
Establishing Trend Limits from Stability Data - Simplified Approach Using the Linear
Regression Model
The basic procedure is as follows;
1. Plot the assay test data vs. storage time and fit a regression line using the simple linear least-
squares-regression model. The unit for time is usually months.
2. Consult an SME or a statistician if there are unusual patterns or shifts in the stability graph. The
statistician advises if an investigation is required.
3. Any observation that appears atypical or discrepant may be removed from the data set if it has an
identified root cause. The removal of data must be justified in the trend report.
4. If the graph is obviously not linear, transform the X-axis, for example by taking a square or square
root of the X-axis values. If the graph cannot be linearised, consult a statistician or an SME.
5. Fit a linear regression to the data, and plot the 99% regression and prediction curves for the stability
trend limits.
6. In addition, calculate and plot the 99.5% confidence trend limits
The model
The stability data consists of responses from some method collected over multiple lots over a period of time.
Linear regression is used to analyse the stability data (y-axis) versus time (x-axis). The analysis is used to
understand the relationship between the stability data and time and can be used to predict an expected
stability result over time.
The data consists of j pairs of numbers; the stability response and the time when the data were collected.
Denote this pair as R j , T j , where and R j is the response and T j is associated the time point.
Note that the data may be from all lots on test simply put together and there is no identification of a
response and a time point to a lot during the analysis.
Hence, if we have L lots, there will most likely be L pairs of numbers have time point = 0, one for each of the
L lots. The corresponding response, R j , will be different depending on which lot was analysed.
^ = b + mT
R (1.16)
j j
where R j is the best fit estimate of the regression line, b is the intercept and m is the slope.
Let us assume that there are N data pairs. We can calculate the mean response, R , and the mean time T
from;
1 N
R= ∑ Rj
N j =1
(1.17)
and
1 N
T= ∑Tj
N j =1
(1.18)
The sum of squares of the differences from these means from the actual values, S R and ST are then readily
calculated from
( )
N 2
SR = ∑ R j − R (1.19)
j
and similarly
( )
N 2
ST = ∑ T j − T (1.20)
j
( )( )
N
S RT = ∑ R j − R T j − T (1.21)
j
The slope of the regression line, m, is simply the ratio of the two sums of squares
S RT
m = (1.22)
ST
The intercept of the regression line, b, where the x axis is zero is calculated using m and the two mean values
R and T
b = R − mT (1.23)
The degree of correlation, r, is found from
2
SRT
r= (1.24)
SR ST
Note that this is not a measure of linearity but of correlation.
The errors associated with the slope and the intercept can now be calculated from the mean square error,
MSE,
S R − mS RT
MSE = (1.25)
N −2
giving the standard error of the slope, SEm, and the standard error of the intercept, SEb, in equations (1.26)
and (1.27)
MSE
SEm = (1.26)
SR
T2 1
SEb = MSE + (1.27)
St N
The confidence intervals at 99% confidence for the slope, CIm , and the intercept, CIb , can be calculated from
equations (1.28) and (1.29).
The confidence intervals for both regression and prediction are calculated from
1 (T − T )
2
j
CI REG = ±t(0.01, N − 2) RMSE + (1.30)
N ST
and
ECA _AQCWG_ SOP 02_OOE OOT_v1.1_November 2016_rev10_CBu Page 37 of 70
Analytical
Analytical Quality Control Working Group ECA Quality Control
Group
An ECA Foundation Working Group
( )
2
1 Tj − T
CI PRE = ±t(0.01, N − 2) RMSE 1 + + (1.31)
N ST
Using these values the confidence contours for regression and prediction are calculated for each of the T j
time points.
Upper and lower 99.5% confidence acceptance trend limits (TL) can be calculated from
1
TL = mT + b ± t(0.005, N − 2) RMSE 1 + (1.32)
N
Overview
The general random coefficients regression model is a flexible model that allows for multivariate
inputs and covariates. The discussion below applies the random coefficients regression model to
stability data in which for each lot there exists a single response at each time point. Thus, a
simplified version of the general Random Coefficients Regression Model (RCRM) is considered in
which only an intercept and slope are present in the model.
The model
Assume that the trend limits are to be established based on stability data performed on N lots of
product.
Lot l is tested at the nl time points tl,1,L, tl,nl with corresponding responses yl,1,L, yl,nl The Random
Coefficients Regression (RCR) model can be written (Carter & Yang 1986 reference 4 on page 34)
yl, j = al + bl × tl, j + ε l, j where l = 1,L, N j = 1,L, nl (1.33)
The coefficients al and bl are the lot-specific intercept and slope for the degradation rate of lot l . It
is assumed that the coefficients have a bivariate normal distribution:
a j iid α
~ N ,Σ (1.34)
bj β
The error terms ε l, j are assumed to be independent, identically distribution from a normal
distribution with mean 0 and variance σ 2 . It is further assumed that the error terms are
independent of the coefficients al and bl .
Several properties of the RCR Model make it a suitable model for stability trend data:
On-going stability studies are usually designed to have data collected at fixed time points.
However, it is possible to for time points to be missing, or for certain lots to be still under
study, in which case later time points have not been collected. The model (1) allows
complete flexibility in the collection of data, subject to the minor constraint that at least 3 time
points must be collected per lot.
• The error terms, ε l,2 j , are assumed to have constant variance across lots and time.
The error terms, ε l,2 j , represent the variability of the method. There is no reason to suspect
a priori that the variability of the method depends on the lot tested. Thus, it is not
unreasonable to assume that the method variance is the same across all time points and
stability lots.
Parameter estimation
There are multiple approaches to estimating the parameters of the RCR model. See references
4, 6 and 7 on page 34. Trending analysis may be performed across multiple analysis platforms
for example, SAS, R, JMP, Minitab.
An approach that requires only algebraic calculations and not numerical optimization or iterative
re-weighting schemes is a modified version of the estimation scheme described by Carter and
Yang (4 on page 34).
The estimation of the parameters of the RCR model is performed in three steps:
Step 1) A simple linear regression is fitted to each individual lot of stability data as used in the
simple approach
Step 2) The covariance matrix, Σ , and error variance, σ are estimated using the regression
2
α
Step 3) The mean vector, , is estimated as a weighted average of the individual slopes
β
and intercepts obtained in Step 1, with weights depending on the estimates obtained in Step 2
Once the parameters of the RCR model have been estimated, an approximate prediction interval
can be constructed at any time point.
σˆ 2 =
∑ df ⋅ MSE
j j
(1.37)
∑ df j
Let S be the sample covariance matrix of the estimated intercepts and slopes.
Define
1
M =
N
∑M k (1.38)
The estimated covariance matrix, Σ̂ , for the intercept and slope is given by:
Σˆ = S − σˆ 2 ⋅ M (1.39)
The estimated covariance matrix, Σ̂ , given by equation (1.39) may not be positive definite. Carter
and Yang provide a modification to equation (1.39) to insure that the estimate Σ̂ is positive definite.
An alternate approach is that if either the slope or intercept variance is negative, the estimate along
with the estimated covariance are replaced with 0. This is a standard approach for negative
variance estimates [4.10] and is equivalent to converting a random effect into a fixed effect in the
model.
Ω = ( ∑ Wk )
−1
(1.41)
The estimated mean vector is given by:
aˆ ak
ˆ = Ω ∑ Wk ⋅ (1.42)
b bk
OOT Detected
Initiate a
discrepancy to
evaluate the OOT
Include other
parameters in No Other parameters OK?
evaluation
Yes
New time
T0 only Create a new
Yes point result back
data point? time point
within trend)?
No
Draw linear
regression line
through all time
points
Yes
Intersect acceptance
No Lot is OK
criteria < expiry + 6
months?
Yes
Yes
Slope within
No
expected degradation
rates?
No
Expand discrepancy to
evaluate quality of lot
On many occasions, laboratories are faced with historical data requiring analysis (post mortem) after the
discovery of a problem. One of the questions asked of the data is usually has anything changed and, if so,
when did it happen.
Although the Shewhart chart is sensitive to sudden and/or large changes in a measurement sequence, it is
ineffective in detecting small but persistent departure from a bench mark. This bench mark may be a target
or specification value or, more commonly, in post mortem investigations the mean of the data set. The
method of choice in this situation is to employ post mortem CuSum analysis. This technique was developed
in the 1950s by Imperial Chemical Industries Ltd9 and this technique is also described in an obsolescent
British Standard, BS5703 Part 2 recently replaced an ISO norm10.
This is a simple but powerful technique which is not as widely known as it should be. As the name CuSum
implies it is merely the cumulative sum of differences from a bench mark.
It is important to note that this technique attempts to identify if a special cause variation has occurred and
when it happened not why it happened.
The CuSum is calculated from the successive differences from a bench mark. Assuming this bench mark is
the mean of the data set, X , then, for i data points, the value of the CuSum for the ith data point is given by
S i = S i −1 + ( X i − X ) (1.44)
The last value of the CuSum is always zero.
If a process is under statistical control ie contains no special cause variation, the CuSum from the mean will
only have common cause variation ie noise. Therefore, a plot of this CuSum with respect to time (or batch)
will be a straight line parallel to the X axis.
However if there is a downward slope this would indicate that the process average was below the benchmark and
conversely an upward or positive slope would indicate that process average was above the benchmark. The steeper
9 o
Cumulative Sum Techniques, ICI Monograph N 3, Oliver and Boyd, 1964
10
BS ISO 7870-4:2011; Control charts. Cumulative sum charts
the slope the greater is the difference. Hence the objective is to detect changes in slope thereby partitioning the data
into segments. The key aspect is to determine if a slope change is due to a real effect or merely chance through noise.
The distance between the successive real turning points is called the span.
In an ideal noise free world, the interpretation of the CuSum plot would be trivial as illustrated in Figure 9.
+
0
50
-
10 Xi
Figure 9 Idealised CuSum plot
st th
The start and endpoints on a CuSum from the mean are always zero. between the 1 point and the 10 point the slope
th th
is negative indicating that the process average is less than the mean and also between 30 point and the 50 point.
th th
Between the 10 point and the 30 point the reverse is true.
It is important to recognise that this post mortem technique is not an exact statistical evaluation but rather a method of
indicating where to look for change.
1. Calculate the sum of the squares of the differences between successive values and dividing it by 2(n-1). The
localised standard deviation of the data is the square root of this value. The reason for calculating this
localised standard deviation is to minimise the effects of any special cause variation which would increase the
value and make detection of these special causes less sensitive.
i=n
∑∆ 2
i
The successive differences in values are given by ∆ i = X i − X i +1 and therefore sL = i =1
2 ( n − 1)
2. Find by inspection the absolute value of the maximum CuSum for the data set and note the index number
3. Calculate the test statistic
CuSummax
sL
4. Compare this value with the critical value for the span (Table 3). The span is given by the number of data
points within each region. The first value for the Span is the total number of data points in the CuSum.
5. If this change point is statistically significant, divide the CuSum plot into two groups by drawing two lines from
the maximum CuSum to the extremities of the plot. These are the new baselines.
6. Inspect these regions for the next largest CuSum to be tested.
7. If appropriate, recalculate the CuSum and the localised standard deviation for each region.
8. Repeat steps 1 to 7 until no significant statistically turning points remain.
9. Draw the Manhattan plot for each of the regions identified. There will be n+1 regions from n turning points.
The Manhattan plot value is based upon the mean value for each region identified in the CuSum analysis.
Table 3 Critical values for Post Mortem CuSum analysis. Values derived from the nomogram (Figure 12) of British
Standard BS 5703 Part 2 (1980) which was generated using numerical simulation.
ECA _AQCWG_ SOP 02_OOE OOT_v1.1_November 2016_rev10_CBu Page 46 of 70
Analytical
Analytical Quality Control Working Group ECA Quality Control
Group
An ECA Foundation Working Group
TERM DEFINITION
Acceptance Criteria Numerical limits, ranges, or other suitable measures for acceptance of test results .
Acceptance Limit The maximum amount of carryover of one product or cleaning agent allowed in a batch
or dose.
Acceptance Sampling Inspection used to determine whether a batch conforms or not to visual inspection
acceptance criteria.
Accuracy The closeness of agreement between the value which is accepted either as a
conventional true value or an accepted reference value and the value found.
Action Limit/Action Level A level that, when exceeded, indicates a drift from normal operating conditions. Action
limits are based on design criteria, regulatory/industry standards, and intended use of
the area.
Adverse Trend (AT) A continuing deviation from normal “expected” process, product or quality performance
characteristic, that has potential severity impact on safety, purity, efficacy or quality of
the intended product function.
AQL (Acceptance Quality Quality level that is the worst tolerable process average when a continuing series of lots
Limit) is submitted for acceptance sampling.
AQL Inspection Statistical inspection by attributes based on AQL.
Attribute Data Data that consist of counts (i.e. number of defectives in a lot, pass or fail, yes or no) of
defects or defectives in a lot. Typically counts of defective lots or of defects within lots
are used.
Calibration The set of operations which establish, under specified conditions, the relationship
between values indicated by a measuring instrument or measuring system, or values
represented by a material measure, and the corresponding known values of a reference
standard.
Centre Line (CL) Mean value of the control chart statistic
Control Charts Control charts are a graphical method for comparing information from samples
representing the current state of a process against limits established after consideration
of inherent process variability. Their primary use is to provide a means of evaluating if a
process is or is not in a “state of statistical control”.
Control Limits Control limits are used as criteria for signaling the need for assessment or for judging
whether a set of data does or does not indicate a “state of statistical control”. · Lower
Control Limit (LCL) – Minimum value of the control chart statistic that indicates statistical
control · Centre Line (CL) – Mean value of the control chart statistic · Upper Control Limit
(UCL) – Maximum value of the control chart statistic that indicates statistical control
Critical Process Parameter A process parameter whose variability has an impact on a critical quality attribute and
(CPP) therefore should be monitored or controlled to ensure the process produces the
desired quality.
Critical Quality Attribute A physical, chemical, biological or microbiological property or characteristic that should
(CQA) be within an appropriate limit, range, or distribution to ensure the desired product
quality.
TERM DEFINITION
Defect A departure of a quality characteristic that results in a product, process or service not
satisfying its intended usage requirements.
Failure Mode The manner by which a failure is observed; it generally describes the way the failure
occurs.
Histogram A histogram is a graphic display (bar chart) of the frequency distribution (showing
shape, location and variation on a number line) of a data set. The x-axis is divided into
equal size “bins” or segments of the numerical data. The y-axis is the count of
occurrences of data in each segment of the data. The height of the bar for each
segment is proportional to the frequency of occurrence of sample data falling into
that category. Categories must not overlap and must be equal in size.
Individual (I) Chart The individual chart plots each measurement as a separate data point
Individual Moving Range The I-MR chart is made up of an individual chart and moving range chart.
(I-MR) Chart
In-Process Controls Checks performed during production in order to monitor and, if appropriate, to adjust
the process and/or to ensure that the intermediate or API conforms to its
specifications.
Inspection by Attributes Inspection where the unit is classified as conforming or nonconforming with respect to
a specified requirement or set of specified requirements.
Inspection Level (AQL Relationship between lot or batch size and the sample size.
based inspections only)
Intermediate precision Variation in the measurements made by different operators and/or instruments
and/or times in one laboratory from the same sample.
Key Performance Indicator A measured or calculated attribute (typically characterizing the output of a process
(KPI) step) that is indicative of whether a process step has met its goal. KPIs are routinely
measured and should be trended.
Lifecycle All phases in the life of a product from the initial development through marketing until
the product’s discontinuation
Lower Control Limit (LCL) Minimum value of the control chart statistic that indicates statistical control.
Moving Range (MR) Chart The MR chart plots the difference between two consecutive data points as they come
from the process in a sequential order.
Out of Control Limit Single result that is markedly different from others in a series as confirmed by
statistical evaluation, i.e., one point beyond established control limits
Out of Expectation (OOE) OOE results for the purposes of this document are anomalous, unexpected or unusual
Result findings, that have not been classified as either OOS or OOT, for both qualitative or
quantitative testing.
Out of Specification (OOS) All reportable results not meeting established specifications. A confirmed reportable
value that is outside the acceptable specification criteria as stated in the product or
analyte specification (e.g. CoAs, USP).
Out of Trend (OOT) A test result or pattern of results that are outside of pre-defined limits.
Outlier A single result which is markedly different from others in a series as confirmed by
statistical evaluation.
TERM DEFINITION
Prediction Interval (PI) An interval that contains a future observation with a predetermined confidence usually
95% or 99%.
Process Capability Index A statistical measure of process capability. Process Capability Indices are measured in
terms of proportion of the process output that is within product specification or
tolerance.
Process Capability Process capability can be defined as the natural or inherent behaviour of a stable
process that is in a state of statistical control. It is a statistical estimate of the outcome of
a characteristic from a process that has been demonstrated to be in a state of statistical
control.
Process Parameters The variables or conditions of the manufacturing process or in an in-process test
on which a processing decision is based.
Random Sample A random sample is defined as one in which the entire population has an equal
chance of being selected to ensure representative sampling across the batch.
Range Interval comprising an upper and lower range of an attribute or variable
Real-Time Trend Analysis Trend analysis that is performed on applicable tests prior to lot release and as
soon as practically possible after each test result is produced and analyzed.
Repeatability Variation in measurements obtained with one measuring instrument when used
several times by an operator while measuring the identical characteristic on the
same part or parts from the same sample.
Reproducibility Variation in the measurements made by different operators and/or instruments
and/or times and in different laboratories from the same sample.
Robustness The measure of a method’s capacity to remain unaffected by small, but
deliberate variations in method parameters and provides an indication of its
reliability during normal usage.
Sample A portion of supply chain material collected according to a defined sampling
procedure.
Sampling Inspection The inspection of products using one or more samples (e.g. simple, double,
multiple, sequential, skip-lot, etc.).
Sampling Plan Combination of sample size(s) and associated lot/batch acceptability criteria.
Sampling Size The number of items statistically selected from a defined population. Sample
size is always an integer value.
State of Control A condition in which the set of controls consistently provides assurance of
continued process performance and product quality.
State of Statistical A process is considered to be in a “state of statistical control” if it is affected by
Control random (common) causes only, i.e. if no extraordinary, unexpected, special
causes have entered the system.
Tolerance Intervals (TI) An interval within which a stated proportion of a population will lie at a stated
confidence level.
TERM DEFINITION
Suppose we have a run of 84 individual data points from an in-process control. These data are shown in Table 4
Table 4 Data for MR Shewhart Chart (one more significant figure is given for statistical purposes)
The mean value is 100.79. Each moving range is calculated from equation (1.3) and the average moving range
calculated MR . The control limits and the centre line are found from Equation (1.5) as shown below.
MR 3.29
UCL = x + 3 = 100.79 + 3( ) = 100.79 + 8.76 = 109.55
d2 1.128
(1.45)
Centre Line = x = 100.79
MR 3.29
UCL = x − 3 = 100.79 − 3( ) = 100.79 − 8.76 = 92.03
d2 1.128
The value of d2=1.128 is from Table 2 and MR =3.29 from Figure 11.
It is apparent that WECO Rule 4 failures were noted at point 31 on onwards to point 43 indicating a process shift.
Figure 11 Moving Range plot for data from Table 4 (Minitab 17)
A similar feature is noted in Figure 11 for the Moving Range but at a slightly later time, observations 40 to
43, and, in addition 3 points, which exceed WECO rule 1 at 46, 47 and 49 indicating excessive dispersion.
This would indicate that significant trends were apparent between observations 31 to 49 before the process
return
Suppose that the data shown in Table 4 were actually collected in 21 sub groups of n=4 at each time point Allowing the
mean and range to be calculated for each subgroup. It is now possible to plot the mean and range charts
Subgroup X Bar R
1 99.39% 3.26%
2 102.20% 6.67%
3 102.98% 5.48%
4 99.33% 8.49%
Mean (XBar)
5 101.74% 3.81%
6 99.09% 7.18%
x = 100.79
7 99.08% 3.75%
8 97.99% 4.29%
9 98.12% 2.71%
10 98.11% 3.12%
11 99.90% 5.74%
12 98.49% 12.79%
13 104.10% 9.37%
Range (R)
14 100.36% 4.92%
15 102.20% 7.12% R = 5.74
16 100.11% 6.48%
17 105.29% 2.38%
18 100.68% 4.37%
19 100.75% 6.64%
20 103.39% 6.11%
21 103.31% 5.95%
Table 5 Mean & range subgroup Figure 12 Mean and Range Chart from Table 5
data from Table 4
It is now readily apparent that the range is in control but the mean is not as a subgroups 8, 9, 10 & 12 there
are failures of WECO Rule 3 and WECO Rule 1 at point 17. These findings are consistent with those of the
MR chart for individuals in Appendix 2.
By way of example, assume that we have 50 assay results from batches of a drug substance for an impurity and we wish
to see if there have been any changes. The initial plot and the CuSum from the mean is shown in Figure 13. The
CuSum is calculated from equation (1.48). The mean value for the impurity data is 0.45(4).
Data Set
Firstly we need to calculate the localised standard deviation from step 1 on page 46. This is
illustrated on the next page.
As before sL is calculated from; Therefore this turning point is not statistically significant at 95%
i =n
confidence.
∑∆ 2
i
2.030
sL = i =1
= = 0.166
2 ( n − 1) 2 ( 38 − 1) It may be concluded that there is a statistically significant shift in the
process mean at around batch 12 which results in a lower impurity
and the maximum CuSum occurs at Batch 32
content.
with a value of -0.77 (Figure 15).
CuSummax 0.77
= = 4.7
sL 0.166
Similar principles are followed for discrete data trending as for continuous variables but the control limits are calculated
from different equations. Let us assume that we are carrying out an inspection of incoming materials for defects and
we take a sample size of n=50 units. For each subgroup of 50 items we list the number of defects found and calculate
the fraction defective. There is no advantage in calculating the fraction defective p chart over the np chart as will be
seen.
# of Fraction
Subgroup
defectives defective
1 3 0.06
2 8 0.16
3 3 0.06
4 5 0.10 0.40 20
5 4 0.08
6 10 0.20 0.35
7 10 0.20 UCL=15.1
0.30 15
8 9 0.18 UCL= 0.301
Fraction defective
# of defectives
9 4 0.08 0.25
10 6 0.12
11 9 0.18 0.20 10
12 8 0.16
0.15
13 12 0.24
14 6 0.12 0.10 5
15 8 0.16
16 8 0.16 0.05
17 10 0.20 0.00 0
18 13 0.26 0 5 10 15 20 25
19 9 0.18 Subgroup
20 5 0.10
21 7 0.14 Figure 16 p and np control charts for data from Table 6 (Note the fraction
22 9 0.18 defective data points have been offset as they are overlap the # of defectives
23 5 0.10
data points)
24 3 0.06
25 13 0.26
Table 6 Discrete data for p & np
control charts
The p charts limits for fraction defective are calculated from equation (1.12)
p (1 − p ) 0.150(1 − 0.150)
UCL = p + 3 = 0.150 + 3 = 0.150 + 0.151 = 0.301
n 50 (1.49)
Centre Line = p = 0.150
and for the number defective from equation (1.13);
The calculations may be performed manually in Excel using the equations detailed below or by Statistical
Software Packages (for example, Minitab, JMP, SAS, GraphPad Prism amongst others). Whatever method is
used it should be validated or verified under actual conditions of use..
The example given is illustrated using Excel and JMP. Not all the parameters described in the theory section
on page 35 are illustrated but may be calculated from the tables.
The data consist of 8 time points (Tj) at 0, 3, 6, 9, 12, 18, 24 and 36 months. At each time point the
response, (Rj) is determined in triplicate ie a total (N) of 24 data points. The model we are fitting is
which was define in (1.16)
^ = b + mT
R (1.51)
j j
The data and the preliminary calculations are shown in Table 7. The relevant equations are (1.17) to (1.23).
Response Time
( R − R ) (T − T ) ( R − R )(T − T )
2 2
j Rj Tj j j j j
Hence;
( )
N 2
SR = ∑ R j − R = 36.8896 (1.52)
j
( )
N 2
ST = ∑ T j − T = 3024.00 (1.53)
j
and
( )( )
N
SRT = ∑ R j − R T j − T = −309.1500 (1.54)
j
The slope of the regression line, m, and the intercept, b, are obtained by substitution in equations (1.24) and
(1.23)
S RT −309.1500
m = = = −0.1022 (1.55)
ST 3024.00
and
b = R − mT = 98.00 − (−0.1022 *13.5) = 99.3843 (1.56)
Hence (1.51) becomes the fitted regression model
^ = 99.3843 − 0.1022T
R (1.57)
j j
We can now calculate the errors associated with this regression model from the mean square error, MSE,
equation (1.25);
S R − mS RT 36.8896 − (−0.1022(−309.1500)
MSE = = = 0.2402 (1.58)
N −2 (24 − 2)
hence the root mean square error, RMSE, is found from
The sums of squares required for the ANOVA table is now calculated from the fitted values, R j , and shown
^
in Table 8
Response Time Best Fit Line Residuals Squares for ANOVA
( ) ( R − R)
2 2
j Rj Tj ^ = b + mT R j − R j
R Rj − R j j
j j
The sums of squares required are those due to the regression model, SSReg, and the sum of squares attributable to
the residual error, SSError. Taking the values from Table 8 we can construct the ANOVA table shown in Table 9. The
large F ratio indicates a highly significant regression model.
SOURCE OF VARIATION DF SUM OF SQUARES (SS) MEAN SQUARE (MS) F ratio Probability
( )
N 2 SSReg MS Reg
Due to regression 1 1 SSReg = ∑ R j − R 31.6051 MS Reg = 31.6051 F= 131.5752 0.0000
j =1 dof MSE
( )
N 2 SSRes
Residual Error N-2 22 SSError = ∑ R j − R j 5.2845 MSE = 0.2402
j =1
dof
The 99% confidence contours for regression and prediction are calculated for each of the T j time points
from equations (1.30) and (1.31). The 99.5% confidence acceptance trend limits are calculated from
equation (1.32). The results of these calculations are shown in Table 10 and plotted in Figure 17.
Response
Figure 17 Regression line with confidence contours of regression, prediction and trend limits
The output from JMP shows very similar results. They are not identical because of differences in rounding
and algorithms
Linear Fit
Result = 99.384301 - 0.1022321*Time
Summary of Fit
RSquare 0.856748
RSquare Adj 0.850236
Root Mean Square Error 0.490107
Mean of Response 98.00417
Observations (or Sum Wgts) 24
Analysis of Variance
Source DF Sum of Squares Mean Square F Ratio
Model 1 31.605067 31.6051 131.5752
Error 22 5.284516 0.2402 Prob > F
C. Total 23 36.889583 <.0001*
Parameter Estimates
Term Estimate Std Error t Ratio Prob>|t|
Intercept 99.384301 0.156478 635.13 <.0001*
Time -0.102232 0.008913 -11.47 <.0001*
Linear Regression with Confidence Intervals Linear Regression with Prediction Intervals
Result
Linear
DATA Confidence Intervals Prediction Intervals
Regression
Note: Trend limit based on limited data may significantly change after additional data is collected.
The trends generated by the RCR model and by a more simplified regression model both originate
with a linear model:
yt = a + b ⋅ t + ε t (1.60)
where yt is the response at time t
a is the initial response (t = 0)
b is the rate of degradation and
ε t is an error term associated with yt
The two models differ in the interpretation and assumptions of the parameters appearing in the
equation.
The RCR model does not assume that each lot has the same intercept and slope, but rather that the
intercept and slope come from a bivariate normal distribution. The trend limits are generated using
the estimates determined by fitting the stability data to the bivariate normal distribution.
There is no association of a response to a specific lot in the model assumed by a simple regression.
The data is combined and assumed to come from a single population. There is an implicit
assumption that all lots have the same intercept and slope. A single (constant) intercept and slope
are estimated from the pooled data set. The trend limits are generated using the estimates
determined by fitting the stability data to this simple linear model.
To consider the difference between the two trend limit approaches it is necessary to consider if
there is variability in degradation rates between lots, and if so if there is a correlation between the
degradation rate and the response at product release.
Generally, the two trend limits are similar when all lots have a common degradation rate. The RCR
trend limits tend to better reflect the distribution of the stability data over time when there are
differences in slopes between the lots.
Case 1: σ slope
2
=0
The components of Ω are small and as a first order approximation can be set to zero, additionally
there is no variability between the degradation rates of the lots under study.
The approximate RCR trend limits simplify to
aRCR + bRCR ⋅ t ± z0.995 ⋅ σ int
2
+ σ method
2
(1.61)
On the other hand, if there is no variability in degradation rates, the only sources of variability in the
pooled data set are method variability, σ method
2
and variability in the intercepts, σ int
2
.
Provided that the RCR model and simple regression model provide similar estimates of slope and
intercept, it is expected that the two trend limits will be nearly identical as in the example below:
Case 2: σ int,slope
2
≥0
Differences in degradation rates exist between the various stability lots under study. Consider two
lots that follow a linear model, the difference in response between the two lots at time t is given by
the differences in their slopes and intercepts and whether the differences in intercepts and slopes
have the same or different signs.
The behaviour of the absolute difference in response for small values of t depends on whether the
differences in intercepts and slopes have the same or different signs; the absolute difference in
responses between the two lots may diverge for sufficiently large t and the range of the individual
data points from multiple stability lots to increase with time.
In case 2, the covariance of the intercept and slope is non-negative. There is a probability of 0.5 or
greater that differences in intercepts and slopes have the same sign. Consequently, when
examining many lots of data, it is expected that the range of the individual data points from multiple
stability lots to increase with time.
The coefficient of the t 2 term is positive and the coefficient of the t term is non-negative. Hence,
the width of the trend limits increases monotonically with increasing t.
A statistical model that assumes a constant variance in the data points at each time point tends to
over-estimate the variability at early time points and under-estimates the variability at later time
points.
Since the trend limits in that case is a constant multiple of the standard estimate of error, the range
of the trend limits is expected to be large relative to the data points for T = 0 and small relative to the
range of the data points for later time points. In comparison, RCR trend limits may have a smaller
width than simple regression trend limits at T=0 and to have a wider width for later time points, as in
the example below:
Case 3: Non-linearity
RCR and simple regression approaches of establishing trends assume an underlying linear
degradation rate. When all data is pooled prior to analysis, small departures from non-linearity in
the individual lots can be obscured by the method variability and the variability of the response at
time of release due to the manufacturing process. As the RCR model fits a regression to each lot,
variability of the response at time of release due to the manufacturing process does not contribute
to the variability of a single lot and consequently can not obscure non-linearity in the degradation
rates.
Consequently, the RCR model is more sensitive to departures from linearity. In this case, a square
root transformation of the time axis was sufficient and necessary to linearize the data and provide
trend limits. The graph below shows the RCR trend limits applied to the original data (red line), the
trend RCR trend limits applied to the time-transformed data (blue line), and the simple regression
trend limits (green line).
Case 2
Ca se 2
Lot D a ys R e sponse
1 0 99.97
1 183 99.96
1 365 99.84
1 0 99.95
1 183 99.79
1 365 99.72
1 730 99.39
2 0 99.97
2 183 99.81
2 365 99.57
2 730 99.52
3 0 99.96
3 91 99.95
3 183 99.93
3 274 99.87
3 365 99.81
3 548 99.41
3 730 99.16
4 0 99.97
4 223 99.94
4 365 99.81
4 730 99.34
5 0 99.98
5 183 99.76
5 365 99.73
5 730 99.64
6 0 99.97
6 183 99.82
6 365 99.85
6 730 99.77
7 0 99.96
7 183 99.94
7 365 99.91
7 730 99.85
8 0 99.98
8 91 99.78
8 183 99.95
8 365 99.86
8 730 99.39
9 0 99.98
9 183 99.93
9 730 99.63
10 0 99.89
10 183 99.77
10 365 99.88
10 730 99.48
Case 3
Ca se 3
Lot D a ys R e sponse Lot D a ys R e sponse Lot D a ys R e sponse
1 0 0.20 7 91 0.41 13 1096 0.60
1 183 0.43 7 183 0.44 13 1461 0.70
1 365 0.61 7 274 0.38 14 0 0.13
1 0 0.15 7 365 0.50 14 91 0.24
1 91 0.33 7 548 0.57 14 183 0.36
1 183 0.38 7 730 0.59 14 274 0.41
1 274 0.43 7 913 0.60 14 365 0.37
1 365 0.47 7 1096 0.60 14 548 0.39
2 0 0.17 7 1461 0.80 14 730 0.47
2 91 0.26 8 0 0.14 14 913 0.53
2 183 0.39 8 365 0.41 14 1096 0.63
2 274 0.44 8 730 0.62 14 1461 0.70
2 365 0.39 9 0 0.19 15 0 0.13
2 548 0.45 9 91 0.29 15 91 0.21
2 730 0.50 9 183 0.35 15 183 0.34
2 913 0.57 9 274 0.37 15 274 0.29
2 1096 0.69 9 365 0.52 15 365 0.39
2 1461 0.83 9 548 0.57 15 548 0.38
3 0 0.17 9 730 0.62 15 730 0.47
3 91 0.23 10 0 0.17 15 913 0.50
3 183 0.37 10 91 0.27 15 1096 0.61
3 274 0.41 10 183 0.32 15 1461 0.77
3 365 0.36 10 274 0.47 16 0 0.14
3 548 0.42 10 365 0.50 16 183 0.23
3 730 0.49 10 548 0.55 16 365 0.33
3 913 0.54 10 730 0.66 16 730 0.53
3 1096 0.69 11 0 0.16 16 1096 0.59
3 1461 0.75 11 91 0.24 17 0 0.15
4 0 0.18 11 183 0.41 17 365 0.35
4 91 0.26 11 274 0.43 17 730 0.55
4 183 0.39 11 365 0.62 17 0 0.15
4 274 0.42 11 548 0.55 17 365 0.63
4 365 0.39 11 730 0.69 17 730 0.64
4 548 0.44 12 0 0.23 18 0 0.14
4 730 0.52 12 365 0.46 18 365 0.49
4 913 0.57 12 730 0.49 18 730 0.61
4 1096 0.68 12 1096 0.67 18 365 0.44
4 1461 0.77 12 1461 0.76 18 730 0.55
5 0 0.17 13 0 0.12 18 1096 0.80
5 365 0.35 13 91 0.27 18 1461 0.84
5 730 0.50 13 183 0.35 19 365 0.42
5 1096 0.70 13 274 0.34 19 730 0.68
6 365 0.40 13 365 0.31 19 1096 0.73
6 730 0.57 13 548 0.37 19 1461 0.82
6 1096 0.57 13 730 0.44
6 1461 0.74 13 913 0.57