Control charts are
graphical plots used in production control to determine whether quality and
manufacturing
Manufacturing is the creation or production of goods with the help of equipment, labor, machines, tools, and chemical or biological processing or formulation. It is the essence of the
secondary sector of the economy. The term may refer ...
processes are being
controlled under stable conditions. (ISO 7870-1)
The hourly status is arranged on the graph, and the occurrence of abnormalities is judged based on the presence of data that differs from the conventional trend or deviates from the control limit line.
Control charts are classified into
Shewhart individuals control chart (ISO 7870-2)
and
CUSUM(CUsUM)(or cumulative sum control chart)(ISO 7870-4).
Control charts, also known as Shewhart charts (after
Walter A. Shewhart) or process-behavior charts, are a
statistical process control
Statistical process control (SPC) or statistical quality control (SQC) is the application of statistics, statistical methods to monitor and control the quality of a production process. This helps to ensure that the process operates efficiently, ...
tool used to determine if a
manufacturing
Manufacturing is the creation or production of goods with the help of equipment, labor, machines, tools, and chemical or biological processing or formulation. It is the essence of the
secondary sector of the economy. The term may refer ...
or
business process
A business process, business method, or business function is a collection of related, structured activities or tasks performed by people or equipment in which a specific sequence produces a service or product (that serves a particular business g ...
is in a state of
control. It is more appropriate to say that the control charts are the graphical device for statistical process monitoring (SPM). Traditional control charts are mostly designed to monitor process parameters when the underlying form of the process distributions are known. However, more advanced techniques are available in the 21st century where incoming data streaming can-be monitored even without any knowledge of the underlying process distributions.
Distribution-free control charts are becoming increasingly popular.
Overview
If analysis of the control chart indicates that the process is currently under control (i.e., is stable, with variation only coming from sources common to the process), then no corrections or changes to process control parameters are needed or desired. In addition, data from the process can be used to
predict the future performance of the process. If the chart indicates that the monitored process is not in control, analysis of the chart can help determine the sources of
variation, as this will result in degraded process performance. A process that is stable but operating outside desired (specification) limits (e.g., scrap rates may be in statistical control but above desired limits) needs to be improved through a deliberate effort to understand the causes of current performance and fundamentally improve the process.
The control chart is one of the
seven basic tools of
quality control
Quality control (QC) is a process by which entities review the quality of all factors involved in production. ISO 9000 defines quality control as "a part of quality management focused on fulfilling quality requirements".
This approach plac ...
. Typically control charts are used for
time-series data, also known as
continuous data or variable data. Although they can also be used for data that has logical comparability (i.e. you want to compare samples that were taken all at the same time, or the performance of different individuals); however the type of chart used to do this requires consideration.
History
The control chart was invented by
Walter A. Shewhart working for
Bell Labs
Nokia Bell Labs, commonly referred to as ''Bell Labs'', is an American industrial research and development company owned by Finnish technology company Nokia. With headquarters located in Murray Hill, New Jersey, Murray Hill, New Jersey, the compa ...
in the 1920s. The company's engineers had been seeking to improve the reliability of their
telephony
Telephony ( ) is the field of technology involving the development, application, and deployment of telecommunications services for the purpose of electronic transmission of voice, fax, or data, between distant parties. The history of telephony is ...
transmission systems. Because
amplifier
An amplifier, electronic amplifier or (informally) amp is an electronic device that can increase the magnitude of a signal (a time-varying voltage or current). It is a two-port electronic circuit that uses electric power from a power su ...
s and other equipment had to be buried underground, there was a stronger business need to reduce the frequency of failures and repairs. By 1920, the engineers had already realized the importance of reducing variation in a manufacturing process. Moreover, they had realized that continual process-adjustment in reaction to non-conformance actually increased variation and degraded quality. Shewhart framed the problem in terms of
common- and special-causes of variation and, on May 16, 1924, wrote an internal memo introducing the control chart as a tool for distinguishing between the two. Shewhart's boss, George Edwards, recalled: "Dr. Shewhart prepared a little memorandum only about a page in length. About a third of that page was given over to a simple diagram which we would all recognize today as a schematic control chart. That diagram, and the short text which preceded and followed it set forth all of the essential principles and considerations which are involved in what we know today as process quality control." Shewhart stressed that bringing a production process into a state of
statistical control, where there is only common-cause variation, and keeping it in control, is necessary to predict future output and to manage a process economically.
Shewhart created the basis for the control chart and the concept of a state of statistical control by carefully designed experiments. While Shewhart drew from pure mathematical statistical theories, he understood that data from physical processes typically produce a "
normal distribution
In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is
f(x) = \frac ...
curve" (a
Gaussian distribution
In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real number, real-valued random variable. The general form of its probability density function is
f(x ...
, also commonly referred to as a "
bell curve"). He discovered that observed variation in manufacturing data did not always behave the same way as data in nature (
Brownian motion
Brownian motion is the random motion of particles suspended in a medium (a liquid or a gas). The traditional mathematical formulation of Brownian motion is that of the Wiener process, which is often called Brownian motion, even in mathematical ...
of particles). Shewhart concluded that while every process displays variation, some processes display controlled variation that is natural to the process, while others display uncontrolled variation that is not present in the process causal system at all times.
In 1924, or 1925, Shewhart's innovation came to the attention of
W. Edwards Deming, then working at the
Hawthorne facility. Deming later worked at the
United States Department of Agriculture
The United States Department of Agriculture (USDA) is an executive department of the United States federal government that aims to meet the needs of commercial farming and livestock food production, promotes agricultural trade and producti ...
and became the mathematical advisor to the
United States Census Bureau
The United States Census Bureau, officially the Bureau of the Census, is a principal agency of the Federal statistical system, U.S. federal statistical system, responsible for producing data about the American people and American economy, econ ...
. Over the next half a century, Deming became the foremost champion and proponent of Shewhart's work. After the defeat of
Japan
Japan is an island country in East Asia. Located in the Pacific Ocean off the northeast coast of the Asia, Asian mainland, it is bordered on the west by the Sea of Japan and extends from the Sea of Okhotsk in the north to the East China Sea ...
at the close of
World War II
World War II or the Second World War (1 September 1939 – 2 September 1945) was a World war, global conflict between two coalitions: the Allies of World War II, Allies and the Axis powers. World War II by country, Nearly all of the wo ...
, Deming served as statistical consultant to the
Supreme Commander for the Allied Powers. His ensuing involvement in Japanese life, and long career as an industrial consultant there, spread Shewhart's thinking, and the use of the control chart, widely in Japanese manufacturing industry throughout the 1950s and 1960s.
Bonnie Small worked in an Allentown plant in the 1950s after the
transistor
A transistor is a semiconductor device used to Electronic amplifier, amplify or electronic switch, switch electrical signals and electric power, power. It is one of the basic building blocks of modern electronics. It is composed of semicondu ...
was made. Used Shewhart's methods to improve plant performance in quality control and made up to 5000 control charts. In 1958, ''The Western Electric Statistical Quality Control Handbook'' had appeared from her writings and led to use at AT&T.
Chart details
A control chart consists of:
* Points representing a statistic (e.g., a
mean
A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
, range, proportion) of measurements of a quality characteristic in samples taken from the process at different times (i.e., the data)
* The mean of this statistic using all the samples is calculated (e.g., the mean of the means, mean of the ranges, mean of the proportions) - or for a reference period against which change can be assessed. Similarly a median can be used instead.
* A centre line is drawn at the value of the mean or median of the statistic
* The
standard deviation
In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its Expected value, mean. A low standard Deviation (statistics), deviation indicates that the values tend to be close to the mean ( ...
(e.g., sqrt(variance) of the mean) of the statistic is calculated using all the samples - or again for a reference period against which change can be assessed. in the case of XmR charts, strictly it is an approximation of standard deviation, the does not make the assumption of homogeneity of process over time that the standard deviation makes.
* Upper and lower
control limits (sometimes called "natural process limits") that indicate the threshold at which the process output is considered statistically 'unlikely' and are drawn typically at 3 standard deviations from the center line
The chart may have other optional features, including:
* More restrictive upper and lower warning or control limits, drawn as separate lines, typically two standard deviations above and below the center line. This is regularly used when a process needs tighter controls on variability.
* Division into zones, with the addition of rules governing frequencies of observations in each zone
* Annotation with events of interest, as determined by the Quality Engineer in charge of the process' quality
* Action on special causes
(n.b., there are several rule sets for detection of signal; this is just one set. The rule set should be clearly stated.)
# Any point outside the control limits
# A Run of 7 Points all above or all below the central line - Stop the production
#* Quarantine and 100% check
#* Adjust Process.
#* Check 5 Consecutive samples
#* Continue The Process.
# A Run of 7 Point Up or Down - Instruction as above
Chart usage
If the process is in control (and the process statistic is normal), 99.7300% of all the points will fall between the control limits. Any observations outside the limits, or systematic patterns within, suggest the introduction of a new (and likely unanticipated) source of variation, known as a
special-cause variation. Since increased variation means increased
quality costs, a control chart "signaling" the presence of a special-cause requires immediate investigation.
This makes the control limits very important decision aids. The control limits provide information about the process behavior and have no intrinsic relationship to any
specification
A specification often refers to a set of documented requirements to be satisfied by a material, design, product, or service. A specification is often a type of technical standard.
There are different types of technical or engineering specificati ...
targets or
engineering tolerance. In practice, the process mean (and hence the centre line) may not coincide with the specified value (or target) of the quality characteristic because the process design simply cannot deliver the process characteristic at the desired level.
Control charts limit
specification limits or targets because of the tendency of those involved with the process (e.g., machine operators) to focus on performing to specification when in fact the least-cost course of action is to keep process variation as low as possible. Attempting to make a process whose natural centre is not the same as the target perform to target specification increases process variability and increases costs significantly and is the cause of much inefficiency in operations.
Process capability
The process capability is a measurable property of a Process (engineering), process to the specification, expressed as a process capability index (e.g., Cpk or Cpm) or as a process performance index (e.g., Ppk or Ppm). The output of this measurem ...
studies do examine the relationship between the natural process limits (the control limits) and specifications, however.
The purpose of control charts is to allow simple detection of events that are indicative of an increase in process variability. This simple decision can be difficult where the process characteristic is continuously varying; the control chart provides statistically objective criteria of change. When change is detected and considered good its cause should be identified and possibly become the new way of working, where the change is bad then its cause should be identified and eliminated.
The purpose in adding warning limits or subdividing the control chart into zones is to provide early notification if something is amiss. Instead of immediately launching a process improvement effort to determine whether special causes are present, the Quality Engineer may temporarily increase the rate at which samples are taken from the process output until it is clear that the process is truly in control. Note that with three-sigma limits,
common-cause variations result in signals less than once out of every twenty-two points for skewed processes and about once out of every three hundred seventy (1/370.4) points for normally distributed processes.
The two-sigma warning levels will be reached about once for every twenty-two (1/21.98) plotted points in normally distributed data. (For example, the means of sufficiently large samples drawn from practically any underlying distribution whose variance exists are normally distributed, according to the Central Limit Theorem.)
Choice of limits
Shewhart set ''3-sigma'' (3-standard deviation) limits on the following basis.
*The coarse result of
Chebyshev's inequality that, for any
probability distribution
In probability theory and statistics, a probability distribution is a Function (mathematics), function that gives the probabilities of occurrence of possible events for an Experiment (probability theory), experiment. It is a mathematical descri ...
, the
probability
Probability is a branch of mathematics and statistics concerning events and numerical descriptions of how likely they are to occur. The probability of an event is a number between 0 and 1; the larger the probability, the more likely an e ...
of an outcome greater than ''k''
standard deviation
In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its Expected value, mean. A low standard Deviation (statistics), deviation indicates that the values tend to be close to the mean ( ...
s from the
mean
A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
is at most 1/''k''
2.
*The finer result of the
Vysochanskii–Petunin inequality, that for any
unimodal probability distribution, the
probability
Probability is a branch of mathematics and statistics concerning events and numerical descriptions of how likely they are to occur. The probability of an event is a number between 0 and 1; the larger the probability, the more likely an e ...
of an outcome greater than ''k''
standard deviation
In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its Expected value, mean. A low standard Deviation (statistics), deviation indicates that the values tend to be close to the mean ( ...
s from the
mean
A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
is at most 4/(9''k''
2).
*In the
Normal distribution
In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is
f(x) = \frac ...
, a very common
probability distribution
In probability theory and statistics, a probability distribution is a Function (mathematics), function that gives the probabilities of occurrence of possible events for an Experiment (probability theory), experiment. It is a mathematical descri ...
, 99.7% of the observations occur within three
standard deviation
In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its Expected value, mean. A low standard Deviation (statistics), deviation indicates that the values tend to be close to the mean ( ...
s of the
mean
A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means (or "measures of central tendency") in mathematics, especially in statist ...
(see
Normal distribution
In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is
f(x) = \frac ...
).
Shewhart summarized the conclusions by saying:
''... the fact that the criterion which we happen to use has a fine ancestry in highbrow statistical theorems does not justify its use. Such justification must come from empirical evidence that it works. As the practical engineer might say, the proof of the pudding is in the eating.''
Although he initially experimented with limits based on
probability distribution
In probability theory and statistics, a probability distribution is a Function (mathematics), function that gives the probabilities of occurrence of possible events for an Experiment (probability theory), experiment. It is a mathematical descri ...
s, Shewhart ultimately wrote:
''Some of the earliest attempts to characterize a state of statistical control were inspired by the belief that there existed a special form of frequency function'' f ''and it was early argued that the normal law characterized such a state. When the normal law was found to be inadequate, then generalized functional forms were tried. Today, however, all hopes of finding a unique functional form'' f ''are blasted.''
The control chart is intended as a
heuristic
A heuristic or heuristic technique (''problem solving'', '' mental shortcut'', ''rule of thumb'') is any approach to problem solving that employs a pragmatic method that is not fully optimized, perfected, or rationalized, but is nevertheless ...
.
Deming insisted that it is not a
hypothesis test and is not motivated by the
Neyman–Pearson lemma. He contended that the disjoint nature of
population
Population is a set of humans or other organisms in a given region or area. Governments conduct a census to quantify the resident population size within a given jurisdiction. The term is also applied to non-human animals, microorganisms, and pl ...
and
sampling frame
In statistics, a sampling frame is the source material or device from which a Sampling (statistics), sample is drawn. It is a list of all those within a Statistical population, population who can be sampled, and may include individuals, households ...
in most industrial situations compromised the use of conventional statistical techniques.
Deming's intention was to seek insights into the
cause system of a process ''...under a wide range of unknowable circumstances, future and past....'' He claimed that, under such conditions, ''3-sigma'' limits provided ''... a rational and economic guide to minimum economic loss...'' from the two errors:
#''Ascribe a variation or a mistake to a special cause (assignable cause) when in fact the cause belongs to the system (common cause).'' (Also known as a
Type I error
Type I error, or a false positive, is the erroneous rejection of a true null hypothesis in statistical hypothesis testing. A type II error, or a false negative, is the erroneous failure in bringing about appropriate rejection of a false null hy ...
or False Positive)
#''Ascribe a variation or a mistake to the system (common causes) when in fact the cause was a special cause (assignable cause).'' (Also known as a
Type II error
Type I error, or a false positive, is the erroneous rejection of a true null hypothesis in statistical hypothesis testing. A type II error, or a false negative, is the erroneous failure in bringing about appropriate rejection of a false null hy ...
or False Negative)
Calculation of standard deviation
As for the calculation of control limits, the
standard deviation
In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its Expected value, mean. A low standard Deviation (statistics), deviation indicates that the values tend to be close to the mean ( ...
(error) required is that of the
common-cause variation in the process. Hence, the usual
estimator
In statistics, an estimator is a rule for calculating an estimate of a given quantity based on Sample (statistics), observed data: thus the rule (the estimator), the quantity of interest (the estimand) and its result (the estimate) are distinguish ...
, in terms of sample variance, is not used as this estimates the total squared-error loss from both
common- and special-causes of variation.
An alternative method is to use the relationship between the
range of a sample and its
standard deviation
In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its Expected value, mean. A low standard Deviation (statistics), deviation indicates that the values tend to be close to the mean ( ...
derived by
Leonard H. C. Tippett, as an estimator which tends to be less influenced by the extreme observations which typify
special-causes.
Rules for detecting signals
The most common sets are:
*The
Western Electric rules
*The
Wheeler rules (equivalent to the Western Electric zone tests
)
*The
Nelson rules
There has been particular controversy as to how long a run of observations, all on the same side of the centre line, should count as a signal, with 6, 7, 8 and 9 all being advocated by various writers.
The most important principle for choosing a set of rules is that the choice be made before the data is inspected. Choosing rules once the data have been seen tends to increase the
Type I error
Type I error, or a false positive, is the erroneous rejection of a true null hypothesis in statistical hypothesis testing. A type II error, or a false negative, is the erroneous failure in bringing about appropriate rejection of a false null hy ...
rate owing to
testing effects suggested by the data.
Alternative bases
In 1935, the
British Standards Institution
The British Standards Institution (BSI) is the Standards organization, national standards body of the United Kingdom. BSI produces technical standards on a wide range of products and services and also supplies standards certification services ...
, under the influence of
Egon Pearson
Egon Sharpe Pearson (11 August 1895 – 12 June 1980) was one of three children of Karl Pearson and Maria, née Sharpe, and, like his father, a British statistician.
Career
Pearson was educated at Winchester College and Trinity College ...
and against Shewhart's spirit, adopted control charts, replacing ''3-sigma'' limits with limits based on
percentile
In statistics, a ''k''-th percentile, also known as percentile score or centile, is a score (e.g., a data point) a given percentage ''k'' of all scores in its frequency distribution exists ("exclusive" definition) or a score a given percentage ...
s of the
normal distribution
In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is
f(x) = \frac ...
. This move continues to be represented by
John Oakland and others but has been widely deprecated by writers in the Shewhart–Deming tradition.
Performance of control charts
When a point falls outside the limits established for a given control chart, those responsible for the underlying process are expected to determine whether a special cause has occurred. If one has, it is appropriate to determine if the results with the special cause are better than or worse than results from common causes alone. If worse, then that cause should be eliminated if possible. If better, it may be appropriate to intentionally retain the special cause within the system producing the results.
Even when a process is ''in control'' (that is, no special causes are present in the system), there is approximately a 0.27% probability of a point exceeding ''3-sigma'' control limits. So, even an in control process plotted on a properly constructed control chart will eventually signal the possible presence of a special cause, even though one may not have actually occurred. For a Shewhart control chart using ''3-sigma'' limits, this ''false alarm'' occurs on average once every 1/0.0027 or 370.4 observations. Therefore, the ''in-control average run length'' (or in-control ARL) of a Shewhart chart is 370.4.
Meanwhile, if a special cause does occur, it may not be of sufficient magnitude for the chart to produce an immediate ''alarm condition''. If a special cause occurs, one can describe that cause by measuring the change in the mean and/or variance of the process in question. When those changes are quantified, it is possible to determine the out-of-control ARL for the chart.
It turns out that Shewhart charts are quite good at detecting large changes in the process mean or variance, as their out-of-control ARLs are fairly short in these cases. However, for smaller changes (such as a ''1-'' or ''2-sigma'' change in the mean), the Shewhart chart does not detect these changes efficiently. Other types of control charts have been developed, such as the
EWMA chart, the
CUSUM chart and the real-time contrasts chart, which detect smaller changes more efficiently by making use of information from observations collected prior to the most recent data point.
Many control charts work best for numeric data with Gaussian assumptions. The real-time contrasts chart was proposed to monitor process with complex characteristics, e.g. high-dimensional, mix numerical and categorical, missing-valued, non-Gaussian, non-linear relationship.
Criticisms
Several authors have criticised the control chart on the grounds that it violates the
likelihood principle. However, the principle is itself controversial and supporters of control charts further argue that, in general, it is impossible to specify a
likelihood function
A likelihood function (often simply called the likelihood) measures how well a statistical model explains observed data by calculating the probability of seeing that data under different parameter values of the model. It is constructed from the ...
for a process not in statistical control, especially where knowledge about the
cause system of the process is weak.
Some authors have criticised the use of average run lengths (ARLs) for comparing control chart performance, because that average usually follows a
geometric distribution
In probability theory and statistics, the geometric distribution is either one of two discrete probability distributions:
* The probability distribution of the number X of Bernoulli trials needed to get one success, supported on \mathbb = \;
* T ...
, which has high variability and difficulties.
Some authors have criticized that most control charts focus on numeric data. Nowadays, process data can be much more complex, e.g. non-Gaussian, mix numerical and categorical, or be missing-valued.
Types of charts
†Some practitioners also recommend the use of Individuals charts for attribute data, particularly when the assumptions of either binomially distributed data (p- and np-charts) or Poisson-distributed data (u- and c-charts) are violated.
Two primary justifications are given for this practice. First, normality is not necessary for statistical control, so the Individuals chart may be used with non-normal data.
Second, attribute charts derive the measure of dispersion directly from the mean proportion (by assuming a probability distribution), while Individuals charts derive the measure of dispersion from the data, independent of the mean, making Individuals charts more robust than attributes charts to violations of the assumptions about the distribution of the underlying population.
It is sometimes noted that the substitution of the Individuals chart works best for large counts, when the binomial and
Poisson distribution
In probability theory and statistics, the Poisson distribution () is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time if these events occur with a known const ...
s approximate a normal distribution. i.e. when the number of trials for p- and np-charts or for u- and c-charts.
Critics of this approach argue that control charts should not be used when their underlying assumptions are violated, such as when process data is neither normally distributed nor binomially (or Poisson) distributed. Such processes are not in control and should be improved before the application of control charts. Additionally, application of the charts in the presence of such deviations increases the
type I and type II error rates of the control charts, and may make the chart of little practical use.
See also
*
Analytic and enumerative statistical studies
*
Common cause and special cause
*
Process capability
The process capability is a measurable property of a Process (engineering), process to the specification, expressed as a process capability index (e.g., Cpk or Cpm) or as a process performance index (e.g., Ppk or Ppm). The output of this measurem ...
*
Seven Basic Tools of Quality
*
Six Sigma
Six Sigma (6σ) is a set of techniques and tools for process improvement. It was introduced by American engineer Bill Smith while working at Motorola in 1986.
Six Sigma strategies seek to improve manufacturing quality by identifying and removin ...
*
Statistical process control
Statistical process control (SPC) or statistical quality control (SQC) is the application of statistics, statistical methods to monitor and control the quality of a production process. This helps to ensure that the process operates efficiently, ...
*
Total quality management
Total quality management (TQM) is an organization-wide effort to "install and make a permanent climate where employees continuously improve their ability to provide on-demand products and services that customers will find of particular value." ...
References
Bibliography
*
*
*
*
*
*
*
*
*
*
External links
NIST/SEMATECH e-Handbook of Statistical Methods
{{DEFAULTSORT:Control Chart
Product management
Quality
Quality control tools
Statistical charts and diagrams
Change detection