Professional Documents
Culture Documents
The Desk Reference of Statistical Quality Methods PDF
The Desk Reference of Statistical Quality Methods PDF
qxd
10/18/07
11:38 AM
Page i
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page ii
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page iii
Mark L. Crossley
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page iv
2007033871
No part of this book may be reproduced in any form or by any means, electronic, mechanical, photocopying,
recording, or otherwise, without the prior written permission of the publisher.
Publisher: William A. Tony
Acquisitions Editor: Matt Meinholz
Project Editor: Paul OMara
Production Administrator: Randall Benson
ASQ Mission: The American Society for Quality advances individual, organizational, and community
excellence worldwide through learning, quality improvement, and knowledge exchange.
Attention Bookstores, Wholesalers, Schools, and Corporations: ASQ Quality Press books, videotapes,
audiotapes, and software are available at quantity discounts with bulk purchases for business, educational, or
instructional use. For information, please contact ASQ Quality Press at 800-248-1946, or write to ASQ
Quality Press, P.O. Box 3005, Milwaukee, WI 53201-3005.
To place orders or to request a free copy of the ASQ Quality Press Publications Catalog, including ASQ
membership information, call 800-248-1946. Visit our Web site at www.asq.org or http://www.asq.org/
quality-press.
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page v
Contents
H1317_CH00FM.qxd
vi
10/18/07
11:38 AM
Page vi
Contents
Chart Resolution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
Variables Control Charts: Individuals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
Variables Control Charts: Averages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
What about Subgroup Size for the Variables Control Charts? . . . . . . . . . . . 71
Attribute Chart: p Charts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77
For P 0.10, Poisson Distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
For P > 0.10, Normal Distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
Chi-Square Contingency and Goodness-of-Fit . . . . . . . . . . . . . . . . . . . . . . . . . 83
Chi-Square Contingency Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
Goodness-of-Fit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
Goodness-of-Fit to Poisson . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
Goodness-of-Fit for the Uniform Distribution . . . . . . . . . . . . . . . . . . . . . . . . . 89
Goodness-of-Fit for Any Specified Distribution . . . . . . . . . . . . . . . . . . . . . . . . 89
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90
Chi-Square Control Chart . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
Confidence Interval for the Average . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
Confidence Interval, n 30 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
Confidence Interval, n > 30 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
Confidence Limit, n > 30, Single-Sided Limit . . . . . . . . . . . . . . . . . . . . . . . . 102
Confidence Limit, n 30, Single-Sided Limit . . . . . . . . . . . . . . . . . . . . . . . . 103
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
Confidence Interval for the Proportion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113
Confidence Interval for the Standard Deviation . . . . . . . . . . . . . . . . . . . . . . . 115
Testing Homogeneity of Variances . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 117
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118
Defects/Unit Control Chart . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119
Case Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
Demerit/Unit Control Chart . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 134
Descriptive Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140
Designed Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 141
Normal Probability Plot . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148
Visualization of Two-Factor Interactions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
Fractional Factorial Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 155
Case Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 158
Calculation of Main Effects . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 159
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page vii
Contents
vii
H1317_CH00FM.qxd
viii
10/18/07
11:38 AM
Page viii
Contents
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page ix
Contents
Pre-Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 363
Description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 363
Pre-Control Rules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 364
A Comparison of Pre-Control and SPC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 365
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 367
Process Capability Indices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 369
Cpk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 369
Confidence Interval for Cpk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 371
Cr and Cp Indices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 374
Confidence Interval for Cp and Cr . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 375
Cpm Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 376
Required Sample Cpk to Achieve a Desired Cpk . . . . . . . . . . . . . . . . . . . . . . . 380
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 381
Regression and Correlation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 383
Linear Least Square Fit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 383
Confidence for the Estimated Expected Mean Value . . . . . . . . . . . . . . . . . . . 384
Assessing the Model: Drawing Inferences about 1 . . . . . . . . . . . . . . . . . . . . 385
Testing 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 386
Measuring the Strength of the Linear Relationship . . . . . . . . . . . . . . . . . . 387
Testing the Coefficient of Correlation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 387
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 389
Reliability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 391
Exponential Distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 395
Determination of Failure Rate, . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 395
Confidence Interval for from Time-Censored (Type I) Data . . . . . . . . . . 398
Confidence Interval for from Failure-Censored (Type II) Data . . . . . . . 399
Success Run Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 401
Reliability Modeling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 402
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 407
Sequential Simplex Optimization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 409
Theoretical Optimum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 409
Strategies for Optimization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 410
Shotgun Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 410
Approach of a Single Factor at a Time . . . . . . . . . . . . . . . . . . . . . . . . . . . . 410
Basic Simplex Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 411
Simplex Calculations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 412
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 418
Sequential Simplex Optimization, Variable Size . . . . . . . . . . . . . . . . . . . . . . . 419
Short-Run Attribute Control Chart . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 429
Case Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 430
Short-Run p Chart . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 434
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 436
ix
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page x
Contents
For X . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 439
For R . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 440
Case Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 440
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 448
Short-Run Individual/Moving Range Control Chart . . . . . . . . . . . . . . . . . . . 449
Derivation of the Plotting Characteristic for the Location Statistic . . . . . . . . 449
Derivation of the Plotting Characteristic for the Variation Statistic . . . . . . . . 450
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page xi
SPC (Basic)
Reliability
Process Improvement
Descriptive Methods
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page xii
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page xiii
Preface
This desk reference provides the quality practitioner with a single resource that illustrates
in a practical manner exactly how to execute specific statistical methods frequently used
in the quality sciences. This reference is not intended to be a rigorously theoretical treatise
on the subject; rather, it briefly describes how to use a tool or technique. A plethora of topics have been arranged in alphabetical order, ranging from Accelerated Reliability Testing
to Zone Format Control Chart. At the end of most topics is a bibliography. All of the examples in this reference are based on data that are purely hypothetical and fictitious in nature.
This desk reference does not assume that the reader has already been exposed to the details
of a specific topic but perhaps has some familiarity with the topic and wants more information regarding its application. This reference is accessible for the average quality practitioner, who will need a minimal prior understanding of the techniques discussed to benefit from them. Each topic is presented in a stand-alone fashion with, in most cases, several
examples detailing computational steps and application comments.
Descriptive Methods
These tools and techniques are used to measure selected parameters such as the central tendency, variation, probabilities, data presentation methods, and other single-value responses
such as measurement error and process capability indices. Topics in this category answer
such questions as: What size sample should I take to determine the average of a process?
How much error do I have using my current measurement system? How large of a sample
do I take to be 90 percent confident that the true Cpk is not less than 1.20? If the measured
kurtosis of my data distribution is +0.75, is the distribution considered normal? What is the
hypergeometric distribution, and when do I use it? Will my error for the proportion be
affected if my sample size represents a significant proportion of my population?
xiii
H1317_CH00FM.qxd
xiv
10/18/07
11:38 AM
Page xiv
Preface
Process Improvement
Certain techniques can be used to improve processes by reducing variation or optimizing
operating conditions. These techniques provide the means to improve processes by
addressing such questions as: Can I use design of experiments on a continuous basis for
process improvement? How do I use design of experiment to reduce variation?
Reliability
This section provides analysis of life data and applications related to the understanding of
product reliability. These topics address such issues as: I have calculated the MTBF from
12 failure times; how do I determine the 90 percent confidence interval? What is the bathtub curve, and how does failure rate relate to the age of a product? How can I graphically
estimate the parameters of the Weibull distribution?
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page xv
Preface
xv
H1317_CH00FM.qxd
xvi
10/18/07
11:38 AM
Page xvi
Preface
While this reference does not exhaust the list of applicable techniques and tools to improve
quality, it is hoped that those included will provide the reader with a more in-depth understanding on their use.
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page xvii
H1317_CH00FM.qxd
10/18/07
11:38 AM
Page xviii
H1317_CH01.qxd
10/15/07
5:43 PM
Page 1
Accelerated Reliability
Testing
The objective of accelerated testing is to compress the time to failure. The only way to
accelerate this time is to increase the stress on the system or unit under study. Failure
occurs when the stress exceeds the strength of the system. The amount of stress and
strength can be thought of in terms of a distribution. Where the two overlap, we have
unreliability.
Applying additional stress (above the normal operating conditions) simply accelerates
the aging process. The modes of the stress can be:
1.
2.
3.
4.
Thermal
Humidity
Vibrational
Chemical
The most commonly used stress accelerator is thermal, where we increase the temperature
above the normal operating condition. This is referred to as high-temperature operating life
(HTOL). The increased temperature shifts the distribution of stress/strength closer together
and decreases the time to failure. As shown in Figure 1, there will always be some unreliability when operating in a normal condition.
Strength
distribution
Stress
distribution
Unreliability
H1317_CH01.qxd
10/15/07
5:43 PM
Page 2
As the stress on the system increases, the unreliability increases. See Figure 2.
Strength
distribution
Stress
distribution
Unreliability
Figure 2 Unreliability increases due to increased stress.
We will assume that the thermally accelerated failure rate follows the Arrhenius relationship. The Arrhenius relationship is important not only in temperature acceleration but
also in other thermodynamic phenomena.
The acceleration factor A is the ratio of the normal life compared to the life in an accelerated stress environment.
A=
tU
tS
(1)
Ea
K BT
(2)
H1317_CH01.qxd
10/15/07
5:43 PM
Page 3
where a catastrophic event (failure) occurs. The rate of these reactions is inversely proportional to the temperature.
t2 T1
=
t1 T2
(3)
where:
t1 = rate at temperature T1
t2 = rate at temperature T2
Combining equations (1), (2), and (3):
Ea 1
t2
K B T2 T1
AT = = e
t1
(4)
T1 = stressed temperature
T2 = use temperature
Example:
A device is normally used at 35C. If we subject the device to a temperature of 100C, how
long will we have to test to simulate nine years of service at 35C? Assume that the energy
of activation Ea = 0.63 Ev.
AT = e
E 1
1
a
K B Tuse Tstress
0.63a
1
1
AT = e(7310.87)(0.000565)
AT = e4.13
= 62.2
The test time to simulate nine years (78,840 hours) is
Test time =
H1317_CH01.qxd
10/15/07
5:43 PM
Page 4
We would test the unit at 100C for 52.8 days. This is equivalent to nine years at the normal operating temperature of 35C.
In this example we were given the energy of activation Ea = 0.63. How would we
determine this had it not been given?
Example:
To calculate the energy of activation Ea, we determine the mean time between failure
(MTBF) at two elevated temperatures and use the following relationship:
MTBF1
Ln
MTBF2
Ea = K B
,
1 1
T T
1
2
where:
T2 = the greater test temperature
T1 = the lesser test temperature
MTBF1 = the MTBF @ T1
MTBF2 = the MTBF @ T2.
Note: Both T1 and T2 are greater than the normal-use environment temperature.
A device is tested at 190C (T1) and at 235C (T2). The respective MTBFs are 18,000
cycles and 7000 cycles. What is the energy of activation Ea?
MTBF1
Ln
MTBF2
Ea = K B
1 1
T T
1
2
Ea = 8.6173 105
18, 000
Ln
7000
1
1
463.15 508.15
H1317_CH01.qxd
10/15/07
5:43 PM
Page 5
AT = e
AT = e
Ea 1 1
K B T2 T1
1
1
0.43
AT = e( 4989.96 )( 0.00125)
AT = e6.23
AT = 507.8
MTBF @ 20C = (MTBF @ 190C)(507.8) = (18,000)(507.8) = 9,140,400 cycles
E
a
K B
1
1
T
T
use
stress
1
0.65
1
AT = e(7542.9)(0.00061)
AT = e4.60
AT = 99.5
The acceleration factor for humidity is given by
m
AH = stress ,
Ruse
H1317_CH01.qxd
10/15/07
5:43 PM
Page 6
2.66
AH = 10.6
The combined acceleration factor is the product of the two acceleration factors:
ATH = (AT)(AH) = (99.5)(10.6) = 1054.7.
To simulate 10 years of service under normal operating conditions (35 percent RH, 21C),
we subject the device to a temperature of 85C and a relative humidity of 85 percent for
10
= 0.0095 years or (0.0095)(35) = 3 1 2 days.
1054.7
Ea
, where Ea is the activation energy and KB is Blotzmanns
KB
320.5 338.9
Slope = 9668.8
Slope =
H1317_CH01.qxd
10/15/07
5:43 PM
Page 7
5000
1000
500
MTBF
300
100
60
50
10
Figure 3 Arrhenius plot.
Ea
K B
1 1
T2 T1
AT = e
1
1
0.83
0
00086173
294
.
2
320
.2
.0
0
AT = 14.27
H1317_CH01.qxd
10/15/07
5:43 PM
Page 8
Bibliography
Condra, L. W. 1993. Reliability Improvement with Design of Experiments. New York:
Marcel Dekker.
Moen, R. D., T. W. Nolan, and L. P. Provost. 1991. Improving Quality through Planned
Experimentation. New York: McGraw-Hill.
Nelson, W. 1990. Accelerated Testing. New York: John Wiley & Sons.
Peck, D. S. 1986. Comprehensive Model for Humidity Testing Correlation. International Reliability Physics Symposium, pp. 4450.
H1317_CH02.qxd
10/15/07
10:57 AM
Page 9
U = USL Z ,
where: L = the minimum process average allowed
U = the maximum process average allowed
Z = the upper 100(1 ) percentage point of the standard normal distribution
= the maximum percentage nonconforming allowed for the process.
If a specified level for a type I error or an alpha risk is given, then the upper control
limit (UCL) and the lower control limit (LCL) are
H1317_CH02.qxd
10
10/15/07
10:57 AM
Page 10
UCL = U +
Z
,
n
UCL = USL Z +
Z
,
n
UCL = USL Z ,
n
and
LCL = L
Z
,
n
LCL = LSL + Z +
Z
,
n
LCL = LSL + Z ,
n
where: LCL = lower control limit (modified)
UCL = upper control limit (modified)
LSL = lower specification limit
USL = upper specification limit
= process standard deviation
n = sample size
= maximum proportion nonconforming allowed
= probability of a type I error, alpha risk.
Example:
The product specification is 10.00 4.0, = .0025
Maximum proportion nonconforming, = 0.0010
Data:
X:
R:
#1
#2
#3
#4
#5
#6
#7
#8
#9
#10
9.5
10.5
11.2
9.0
8.0
10.4
9.4
7.9
10.4
11.5
11.2
9.6
8.8
9.2
9.8
8.0
10.5
10.1
10.4
9.2
10.1
11.1
9.8
10.7
8.9
9.5
10.0
11.6
9.9
10.9
12.3
11.3
10.4
9.3
8.9
10.2
9.8
10.2
10.6
8.7
10.68
1.9
8.95
1.8
10.05
1.3
10.43
1.3
10.00
2.7
11.10
2.4
9.7
1.5
9.8
1.9
10.05
2.2
8.93
2.5
H1317_CH02.qxd
10/15/07
10:57 AM
Page 11
R
d2
Values of d2 can be found in Table A.7 in the appendix.
Estimate of sigma, =
= 0.95
Z
UCL = USL Z
n
= 0.00010 Z = 3.71 from standard normal distribution
= 0.0025
Z = 2.81 n = 4 = 0.95 USL = 14.0
Z
2.81
UCL = USL Z UCL = 14.0 3.71
0.95 UCL = 11.81
2
n
Z
LCL = LSL + Z
n
= 0.00010
= 0.0025
2.81
LCL = LSL + Z LCL = 6.0 + 3.71
0.95 LCL = 8.19
2
n
The completed modified control chart is illustrated in Figure 1.
Specification: 10 4
12.00
11.00
10.00
9.00
Lower reject limit = 8.19
8.00
1
10
11
H1317_CH02.qxd
12
10/15/07
10:57 AM
Page 12
Z + Z
n=
.
Z p1 Z p 2
The value chosen for n should be the next greatest integer.
The upper and lower control limits are designated as UACL and LACL, respectively,
and are found from
Z
1
2
n
n
Z
.
LACL = LSL + Z p or LACL = LSL+ Z p +
1
2
n
n
Two formulas are required for each UACL and LACL because the solution for n
almost never will yield an integer value. The conservative action would be to select the
lower value of the UACL and the higher value of the LACL.
As an example of the ACC, we will develop a chart using the data from the previous case
(the modified control limits example), adding the additional parameters of , , p1, and p2.
USL = 14.0
LSL = 6.0
= 0.95
p1 = 0.0005, Zp1 = 3.29
H1317_CH02.qxd
10/15/07
10:57 AM
Page 13
13
2
Z + Z
2.33 + 1.65
n=
n = 3.29 2.17 n = 12.6.
Z p1 Z p 2
Therefore, n = 13.
Z
2.33
n
13
and
Z
1.65
UACL = USL Z p +
UACL = 14.0 2.17 +
0.95 UACL = 11.50
13
n
Use 11.36 for UACL (the lower of the two UACLs).
Z
2.33
n
13
and
Z
1.65
LACL = LSL + Z p +
LACL = 6.0 + 2.17 +
0.95 LACL = 8.49
13
n
For LACL use the greater of the two values, LACL = 8.51.
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Freund, R. A. 1957. Acceptance Control Charts. Industrial Quality Control 14, no. 4:
1321.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Schilling, E. G. 1982. Acceptance Sampling in Quality Control. New York: Marcel
Dekker.
H1317_CH02.qxd
10/15/07
10:57 AM
Page 14
H1317_CH03.qxd
10/18/07
11:40 AM
Page 15
Acceptance Sampling
for Attributes
Lot acceptance sampling plans are used when deciding to accept or reject a lot of material,
either as an incoming lot or as one being released from manufacturing.
For example, you are responsible for the screening and acceptance of materials in
receiving inspection and need to decide about the acceptance of a shipment of parts. The
incoming lot has 2500 units. Assume that there are no variables for the inspection and that
the criterion for inspection is simply a functional test: either the parts perform or they do
not. There are several options from which to choose:
1. Perform no inspection and trust that the supplier has done a good job. The supplier
is ISO 9000 certified.
2. Inspect 100 percent of the units.
3. Sample 10 percent of the lot and accept it if no defective units are found.
4. Derive a statistically based sampling plan that defines all the risk factors for a given
level of quality and perform a sampling inspection.
In this module, the following topics will be discussed:
1.
2.
3.
4.
Sampling plans for attributes are completely defined by their performance relative to
seven characteristics or operational parameters:
Acceptable quality limit (AQL): That percent defective that, for purposes of
acceptance sampling, will be considered acceptable,
that is, good. Lots at the AQL percent defective
will be accepted by the plan the predominant
amount of the time (90 percent to 99 percent of the
time). An AQL must have a defining alpha risk.
Manufacturers risk, :
H1317_CH03.qxd
16
10/18/07
11:40 AM
Page 16
Consumers risk, :
Lot size, N:
Sample size, n:
Acceptance number, Ac or C:
Note: The first four characteristics define the performance of the plan (the operating
characteristic curve).
= 10.0%
H1317_CH03.qxd
10/18/07
11:40 AM
Page 17
17
1.0
.9
.8
.7
Pa
.6
.5
.4
.3
.2
.1
20
10
15
Percent defective in lot
25
30
One useful approach for creating an attribute sampling plan is the Poisson Unity Value
method. This method uses an infinite lot size and is based on the four characteristics of
AQL, RQL, , and .
Example:
Develop an attribute sampling plan that will satisfy the conditions of AQL = 4.0%,
RQL = 16.0%, = 5%, and = 10%.
Step 1. Divide the RQL by the AQL.
RQL 16.0
=
= 4.000 = discrimination ratio
AQL 4.0
Step 2. Look up the discrimination ratio by choosing the closest value available in the column provided for = 0.05 and = 0.10. The closest value in the table for this case is 4.057.
Having located the closest discrimination ratio, go to the extreme left column to find the
C = 4 value, and then go to the extreme right to locate the np1 = 1.970 value.
C
0
1
2
3
4
5
6
= 0.05
= 0.10
44.890
10.946
6.509
4.890
4.057
3.549
3.206
np1
0.052
0.355
0.818
1.366
1.970
2.613
3.386
H1317_CH03.qxd
18
10/18/07
11:40 AM
Page 18
The maximum number of defective units in the sample is C = 4. If the critical number of
defective C is exceeded, reject the lot; otherwise, accept the lot.
Step 3. Calculate the sample size n.
The np1 value is the product of the sample size n and the AQL (AQL = p1).
np
1.97
In this example, np1 = 1.97 and n = 1 ; therefore, n =
= 49.25 49.
p1
0.04
The sample plan is n = 49 and C = 4. The four plan criteria of AQL = 4.0%, RQL = 16.0%,
= 5%, and = 10% will be satisfied by this plan.
Additional points to assist in the construction of the OC curve can be calculated using
Table A.3 in the appendix.
Sample calculation:
In the previous example, C = 4 and n = 49. Looking in the table at the column headed
Pa = 0.75 and the row C = 4, we find the value 3.369. This value represents the product
of the lot percent defective and the sample size that will yield a 75 percent probability
of acceptance Pa.
Dividing this value by the sample size n = 49 gives the percent defective that corresponds to a Pa of 0.75, or 75 percent.
np = 3.369 for Pa = 0.75, where n = 49 and C = 4
3.369
p=
= 0.069
49
Other values for P at various Pas can be determined. The other associated values for
the example sample plan of C = 4 and n = 49 and the graph of the OC curve follow.
Percent
probability of
acceptance Pa
100
95
90
75
50
25
10
5
Percent
defective
for lot
0
4.02
4.96
6.90
9.53
12.80
16.31
18.68
In the original calculation for the sample plan, the = 0.05 for an AQL of 4.00 percent was used to give a sample plan of n = 49.25 using an integer sample size of n = 49.
This rounding off results in the Pa table for Pa = 95% to be 4.02 rather than the 4.0 percent AQL as specified. Also, the Poisson Unity Value of 4.000 could not be found exactly,
and the default was 4.057. This same rounding off of the Poisson Unity Value also results
in the p = 16.31 for a Pa = 10% where the original plan was established for a p = 16.0%
for the RQL at Pa = 10%.
H1317_CH03.qxd
10/18/07
11:40 AM
Page 19
19
Probability of acceptance, Pa
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0.0
0
Percent
probability of
acceptance Pa
100
95
90
75
50
25
10
5
10 12 14 16 18 20 22
Percent lot defective
Percent
defective
for lot
0
4.02
4.96
6.90
9.53
12.80
16.31
18.68
Calculation of Poisson Unity Values and np1 for given C, AQL, and RQL values:
Given = 0.05, = 0.10, and C = 7, calculate the Poisson Unity Value and np1.
Poisson Unity Value =
X 2 ,2 c + 2
1
and np1 = X 2 1 , 2 c + 2
X 2 1 , 2 c + 2
2
H1317_CH03.qxd
20
10/18/07
11:40 AM
Page 20
The levels of inspection range from the nondiscriminating set of special inspection
levels of S-1, S-2, S-3, and S-4 to the general inspection levels that are more discriminating. The default level of inspection is general level II.
to 8
to 15
to 25
to 50
to 90
to 150
to 280
to 500
to 1200
to 3200
to 10,000
to 35,000
to 150,000
to 500,000
and over
S-1
S-2
S-3
S-4
A
A
A
A
B
B
B
B
C
C
C
C
D
D
D
A
A
A
B
B
B
C
C
C
D
D
D
E
E
E
A
A
B
B
C
C
D
D
E
E
F
F
G
G
H
A
A
B
C
C
D
E
E
F
G
G
H
J
J
K
II
III
A
A
B
C
C
D
E
F
G
H
J
K
L
M
N
A
B
C
D
D
E
G
H
J
K
L
M
N
P
Q
B
C
D
E
E
F
H
J
K
L
M
N
P
Q
R
Adapted from ANSI/ASQ Z1.4-2003 Sampling Procedures and Tables for Inspection by Attributes (Milwaukee,
WI: ASQ Quality Press), 10. Used with permission.
H1317_CH03.qxd
10/18/07
11:40 AM
Page 21
21
Start
Preceding 10 lots
accepted, with
Total nonconforming
less than limit number
(optional), and
Production steady,
and
Approved by
responsible authority
Reduced
2 of 5 or fewer
consecutive lots
are not accepted
Normal
Tightened
5 consecutive
lots accepted
5 lots not accepted
while on tightened
inspection
Discontinue
inspection
under Z1.4
The following example illustrates the use of the MIL-STD-105E sampling plan.
A shipment of 1000 bolts has been received. The AQL = 4.0%, and the level of inspection is general level II.
Step 1. Look up the appropriate sampling letter in Table 2.
The correct sample letter is J.
Step 2. In Table 3, look up the sample size for letter J, then go across the top heading to
AQL = 4.0 to locate the Ac and Re criteria.
The sample size for letter J is n = 80, and the acceptance criterion is that the lot is accepted
if seven or fewer defective units are found and is rejected if more than seven defective
units are found.
Determine the sample size and acceptance criteria for the following:
A. Lot size N = 250, level = S-4, AQL = 6.5%
B. Lot size N = 1800, level II, AQL = 0.65%
C. Lot size N = 1100, level I, AQL = 0.65%
H1317_CH03.qxd
22
10/18/07
11:40 AM
Page 22
Adapted from ANSI/ASQ Z1.4-2003 Sampling Procedures and Tables for Inspection by Attributes (Milwaukee, WI:
ASQ Quality Press), 10. Used with permission.
Ac
Re
Use the first sampling plan below the arrow. If sample size equals, or exceeds, lot size, carry out 100 percent inspection.
Use the first sampling plan above the arrow.
Acceptance number.
Rejection number.
Adapted from ANSI/ASQ Z1.4-2003 Sampling Procedures and Tables for Inspection by Attributes (Milwaukee, WI:
ASQ Quality Press), 11. Used with permission.
H1317_CH03.qxd
10/18/07
11:40 AM
Page 23
23
Quality of submitted product (p, in percent nonconforming for AQLs 10; in nonconformities per hundred units for AQLs > 10)
Note: Figures on curves are acceptance quality limits (AQLs) for normal inspection.
Tabulated values for operating characteristic curves for single sampling plans
Acceptance quality limits (normal inspection)
Adapted from ANSI/ASQ Z1.4-2003 Sampling Procedures and Tables for Inspection by Attributes (Milwaukee, WI:
ASQ Quality Press), 47. Used with permission.
The actual operating characteristic can be found in Table 4. The subject standards have OC
curves for each code letter. The operating characteristic data are presented both graphically
and tabularly.
H1317_CH03.qxd
24
10/18/07
11:40 AM
Page 24
Start
Inspect i
successive
units
Replace
defective
with
nondefective
Defective
found
No defective
found
Randomly sample
fraction f units
Defective
found
No defective
found
CSP-1 Plan
The CSP-1 plan was first introduced by Harold F. Dodge in 1943 (Figure 2). Initially, 100
percent inspection is required until i consecutive units have been found to be satisfactory.
Upon reaching the clearing number i, the 100 percent inspection decreases to a fraction f
of the units. Discovery of a defective unit during the frequency inspection results in 100
percent inspection until i consecutive units are found, in which case interval inspection of
every fth unit resumes. Table values for various combinations of i and f for designated
average outgoing quality limits (AOQL) can be found in Table 5.
Application example:
1. An electronic assembly line produces 6500 units per day. What CSP-1 plan will
yield a 0.33 percent AOQL?
Referring to the CSP-1 table, there are several choices. One plan would be to inspect
100 percent until 335 consecutive nondefective units are found, and then inspect every
tenth unit. If any defective units are found during the initial 100 percent inspection, the
defective unit would be replaced with a good unit, and the accumulation count would be
restarted until 335 consecutive good units are found. After that, any tenth sample found
defective will result in the reinstatement of the 100 percent inspection. The AOQL for this
plan is 0.33 percent nonconforming.
.018
1540
2550
3340
3960
4950
6050
7390
9110
11730
14320
17420
0.010
1/2
1/3
1/4
1/5
1/7
1/10
1/15
1/25
1/50
1/100
1/200
840
1390
1820
2160
2700
3300
4030
4970
6400
7810
9500
0.015
.033
600
1000
1310
1550
1940
2370
2890
3570
4590
5600
6810
0.025
.046
375
620
810
965
1205
1470
1800
2215
2855
3485
4235
0.040
.074
245
405
530
630
790
965
1180
1450
1870
2305
2760
0.065
.113
194
321
420
498
623
762
930
1147
1477
1820
2178
0.10
.143
25
84
140
182
217
270
335
410
500
640
790
950
0.25
.33
AQL (%)
140
232
303
360
450
550
672
828
1067
1302
1583
0.15
.198
53
87
113
135
168
207
255
315
400
500
590
0.40
.53
36
59
76
91
113
138
170
210
270
330
400
0.65
.79
23
38
49
58
73
89
108
134
175
215
255
1.0
1.22
15
25
32
38
47
57
70
86
110
135
165
1.5
1.90
10
16
21
25
31
38
46
57
72
89
106
2.5
2.90
6
10
13
15
18
22
27
33
42
52
62
4.0
4.94
5
7
9
11
13
16
19
23
29
36
43
6.5
7.12
3
5
6
7
8
10
12
14
18
22
26
10.0
11.46
11:40 AM
AOQL (%)
10/18/07
H1317_CH03.qxd
Page 25
H1317_CH03.qxd
26
10/18/07
11:40 AM
Page 26
The average number of units inspected in a 100 percent screening sequence following
the occurrence of a defect is equal to
u=
1 Qi
,
PQ i
1 Qi
1 .992 335
,
u
=
, u = 1718.
PQ i
(.008 )(.992 335 )
The average number of units passed under CSP-1 plans before a defective unit v is
found is given by
1
f = 1 / 10 = 0.1 and p = .008
fp
1
v=
v = 1250.
(0.1)(.008 )
v=
The average fraction of the total manufactured units inspected (AFI) in the long run is
given by
u + fv u = 1718 f = 0.10
u + v v = 1250
1718 + (0.1)(1250 )
AFI =
AFI = 0.621.
1718 + 1250
AFI =
The average fraction of manufactured units passed under the CSP-1 plan is given by
Pa =
v
Pa = 0.42 or 42%.
u+v
When Pa is plotted as a function of p, the OC curve for the subject plan is derived. In
a traditional lot acceptance sampling plan, Pa equals the probability of accepting lots that
are p proportional defective.
The OC curve for CSP-1 plans gives the probability or proportion of units passed
under the subject plan as a function of process proportion defective.
H1317_CH03.qxd
10/18/07
11:40 AM
Page 27
27
0.01 0.99
u=
1 Qi
PQ i
v=
47
1
fp
Pa =
1000
v
u+v
0.9551
1 Qi
PQ i
u=
1 .99 38
(.01)(.99 38 )
u = 47
v=
1
fp
v=
1
(.10 )(.01)
v = 1000
Pa =
v
u+v
Pa =
1000
47 + 1000
Pa = 0.9551
0.01
0.02
0.03
0.04
0.05
0.06
0.07
0.08
0.09
0.10
0.11
0.12
0.13
0.14
0.15
0.99
0.98
0.97
0.96
0.95
0.94
0.93
0.92
0.91
0.90
0.89
0.88
0.87
0.86
0.85
1 Qi
PQ i
47
58
73
93
120
158
211
285
389
538
753
1064
1521
2195
3200
u=
1
fp
1000
500
333
250
200
167
143
125
111
100
91
83
77
71
67
v=
v
u+v
0.9551
0.8961
0.8202
0.7289
0.6250
0.5138
0.4040
0.3049
0.2220
0.1567
0.1078
0.0724
0.0481
0.0313
0.0205
Pa =
H1317_CH03.qxd
28
10/18/07
11:40 AM
Page 28
Plot Pa as a function of p:
100
Proportion accepted, Pa
90
i = 38
f = 1/10
80
70
60
50
40
30
20
10
0
0
5 6 7 8 9 10 11 12 13 14 15
Process percent defective, %
P (1 f ) Q i
.
f + (1 f ) Q i
For the example given, where i = 38, f = 1/10, and P = 0.03 (process is 3 percent defective), the AOQ is
0.03(1 0.10 )(0.97 )38
= 0.0222
0.10 + (1 0.10 )(0.97 )38
%AOQ = 2.22%.
AOQ =
H1317_CH03.qxd
10/18/07
11:40 AM
Page 29
29
i = 38
f = 1/10
AOQL = 2.90%
% AOQ
%P
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
0
0
5 6 7 8 9 10 11 12 13 14 15
Process percent defective, %
CSP-2 Plan
As a modification to the original CSP-1 plan, the CSP-2 was developed by Dodge and Torrey (1951). Under the CSP-2 mode, 100 percent inspection is performed until i units are
found to be defect free. If a defect is found, the inspection switches to a fraction f of the
units. If a second defective is found during the sampling inspection of the i units, then 100
percent is resumed until ith consecutive units are found to be defect free. If no defectives
are found during the sampling inspection, then the frequency inspection of a fraction f of
the units is resumed.
The following table gives approximate values for i as a function of the AOQL.
AOQL, %
0.5
1.0
2.0
3.0
4.0
5.0
1/2
1/3
1/4
1/5
1/7
1/10
1/15
1/25
1/50
86
140
175
200
250
290
350
450
540
43
70
85
100
125
148
175
215
270
21
35
43
48
62
73
88
105
135
14
23
28
33
40
47
57
70
86
10
16
21
24
30
36
44
53
64
9
13
17
19
24
28
34
42
52
Example:
A 1.0 percent AOQL plan is wanted with a clearing cycle of n = 100. Because 100 units
are inspected and found to be defect free, every fifth unit is now inspected. During this fre-
H1317_CH03.qxd
30
10/18/07
11:40 AM
Page 30
quency inspection, a defective is found. If during the next 100 consecutive units a second
defective is found, 100 percent inspection is required until 100 consecutive units are found
defect free, in which case the frequency inspection of every fifth unit is reinstated.
The values for initial clearing sample i and the fractional sample frequency f are given
for several AOQLs in the following table.
AOQL, %
0.53
0.79
1.22
1.90
2.90
4.94
7.12
11.46
1/2
1/3
1/4
1/5
1/7
1/10
1/15
1/25
1/50
80
128
162
190
230
275
330
395
490
54
86
109
127
155
185
220
265
330
35
55
70
81
99
118
140
170
210
23
36
45
52
64
76
90
109
134
15
24
30
35
42
50
59
71
88
9
14
18
20
25
29
35
42
52
7
10
12
14
17
20
24
29
36
4
7
8
9
11
13
15
18
22
Effects of Lot Size N, Sample Size n, and Acceptance Criteria C on the OC Curve
Traditional acceptance sampling plans such as ANSI Z1.4 and Z1.9 specify sampling plans
as a function of the lot size. While lot size influences the overall OC curve, it is not the
most contributing factor with respect to the shape of the OC curve. Other factors such as
the sample size and the acceptance criteria C (the number of defective units, at which the
lot will be accepted) are equally or more contributory.
Consider the following OC curves, where the lot size, sample size, and acceptance criteria are varied.
Case I: Constant C = 1 and constant n = 100 with variable lot size N = 200 to 1000
1.0
.9
.8
Pa
.7
.6
.5
N = 1000
.4
.3
.2
.1
N = 200
3
4
% Defective, D
10/18/07
11:40 AM
Page 31
1.0
.9
.8
C=4
Pa
.7
.6
.5
.4
.3
.2
C=0
.1
3
4
% Defective, D
1.0
.9
.8
.7
Pa
H1317_CH03.qxd
n = 50
.6
.5
.4
.3
.2
n = 200
.1
3
4
% Defective, D
31
H1317_CH03.qxd
32
10/18/07
11:40 AM
Page 32
Comparatively, we can see that changes in the lot size have a relatively small effect on
the overall shape of the OC curve. All calculations are based on the hypergeometric distribution incorporating both the lot size N and the sample size n.
k1
100.
k2
For example, if the cost of inspection is $0.50 and the cost of a defective unit received
by the consumer is $60.00, the suggested AQL is
AQL =
0.50
100 = 0.83%.
60
+6000
Lot value if 100% inspected
+4000
+2000
0
2000
4000
6000
8000
0
AQL = 0.83%
10
Lot percent defective
15
20
H1317_CH03.qxd
10/18/07
11:40 AM
Page 33
33
Bibliography
ASQ Quality Press. 2003. Sampling Procedures and Tables for Inspection by Attributes,
ANSI/ASQ Z1.4-2003. Milwaukee, WI: ASQ Quality Press.
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Dodge, H. F., and M. N. Torrey. 1951. Additional Continuous Sampling Inspection
Plans. Industrial Quality Control 7, no. 5 (March): 712.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Schilling, E. G. 1982. Acceptance Sampling in Quality Control. New York: Marcel Dekker.
H1317_CH03.qxd
10/18/07
11:40 AM
Page 34
H1317_CH04.qxd
10/18/07
11:42 AM
Page 35
Acceptance Sampling
for Variables
Lot acceptance plans for attribute data are based on acceptance criteria where a specified
number of nonconforming units and a specified sample size are given. If this number of
nonconforming units is exceeded in a sample, a decision is made to reject the lot from
which the samples were taken. Such inspection plans are predefined under MIL-STD105E (ANSI/ASQ Z1.4-2003). In acceptance sampling for attributes, the unit of inspection
either conforms to a requirement or it does not. The decision criteria for each individual
inspection are based on attributes. In many cases, the unit of inspection is measured as a
direct variable, such as weight, diameter, and angle or viscosity. If this measurement is
compared to a specification requirement, then the observation is converted to an attribute.
If the variable observation is not treated as an attribute, then the acceptance criteria can be
based on the average of the observations and a variation statistic such as the standard deviation as related to the specification requirement. The following inspection plans are those
covered by MIL-STD-414, sampling plans for variable data.
Single Specification Limit, Form I, Variability Unknown, Standard Deviation Method
Levels of inspection determine the sample size as a function of the submitted lot. Levels
of inspection are directly related to the ratio of the acceptable quality limit (AQL) and the
repeatable quality level (RQL). The closer the RQL is to the AQL, the more discriminating the sample plan. Levels available are
S3 S4 I II III
Increasing discrimination .
For a given lot size and level of inspection, the sample size is determined and designated as a letter B through P.
35
H1317_CH04.qxd
36
10/18/07
11:42 AM
Page 36
Lot size
S3
S4
II
III
Letter
2
to 8
9
to 15
16
to 25
26
to 50
51
to 90
91
to 150
151
to 280
281
to 400
401
to 500
501
to 1200
1201
to 3200
3201
to 10,000
10,001 to 35,000
35,001 to 150,000
150,001 to 500,000
500,001
B
B
B
B
B
B
B
C
C
D
E
F
G
H
H
H
B
B
B
B
B
C
D
E
E
F
G
H
I
J
K
K
B
B
B
C
D
E
F
G
G
H
I
J
K
L
M
N
B
B
C
D
E
F
G
H
I
J
K
L
M
N
P
P
C
D
E
F
G
H
I
J
J
K
L
M
N
P
P
P
B
C
D
E
F
G
H
I
J
K
L
M
N
P
Sample size
3
4
5
7
10
15
20
25
35
50
75
100
150
200
S3
8
15
25
50
90
150
280
400
500
1200
3200
10,000
35,000
150,000
500,000
B
B
B
B
B
B
B
C
C
D
E
F
G
H
H
H
Sample
S4
B
B
B
B
B
C
D
E
E
F
G
H
I
J
K
K
size E =
II
III
Letter
B
B
B
C
D
E
F
G
G
H
I
J
K
L
M
N
B
B
C
D
E
F
G
H
I
J
K
L
M
N
P
P
C
D
E
F
G
H
I
J
J
K
L
M
N
P
P
P
B
C
D
E
F
G
H
I
J
K
L
M
N
P
Sample size
3
4
5
7
10
15
20
25
35
50
75
100
150
200
Step 2. Select the sample and calculate the average and sample standard deviation.
Seven samples are chosen and measured. The characteristic for this example is the diameter of the rods.
0.503
0.502
0.503
0.504
0.505
0.501
0.503
X = 0.503
S = 0.001
H1317_CH04.qxd
10/18/07
11:42 AM
Page 37
37
Step 3. Select an AQL and look up the appropriate k value from Table 1.
The specification for this example is a minimum diameter of 0.500 or LSL of 0.500, and
the selected AQL is 2.5 percent.
k for sample size letter E (sample size n = 7) and a 2.5 percent AQL is 1.33.
Step 4. Calculate the critical decision factor (CDF).
For lower specifications:
CDFL = X kS
CDFU = X + kS
CDFL = X kS
CDFL = 0.503 (1.33)(0.001) CDFL = 0.502
Step 5. Confront the CDF with the acceptance criteria.
For lower specifications:
If CDFL is less than the lower specification, reject the lot.
If CDFL is greater than the lower specification, accept the lot.
For upper specifications:
If CDFU is greater than the upper specification, reject the lot.
If CDFU is less than the upper specification, accept the lot.
For this example, CDFL is greater than the lower specification; therefore, the lot is
accepted.
Problem:
A lot of 200 units has been received. The specification is 25.0 maximum. Perform a level II,
AQL = 1.00 inspection using the single specification, form 1 (k method).
Randomly select the sample from the 200 values listed below.
21
19
16
18
20
16
20
18
21
20
20
23
22
19
22
18
24
20
15
23
22
12
20
23
21
20
16
20
17
20
23
23
22
22
20
20
21
22
22
22
17
20
23
21
17
23
22
19
22
20
20
25
21
21
18
21
23
18
19
24
22
21
17
20
21
27
19
19
21
14
21
17
14
16
15
15
22
21
21
17
21
22
18
24
19
19
15
20
16
18
14
18
20
20
19
20
14
17
23
19
19
23
24
18
17
22
21
24
24
21
22
19
21
15
22
23
19
17
20
23
19
21
18
20
19
22
18
19
22
19
19
22
20
21
19
21
20
23
14
22
20
20
24
22
22
26
20
20
18
19
25
18
14
18
21
23
23
20
20
20
18
23
17
18
16
19
23
17
19
20
19
17
19
19
18
21
17
15
21
20
25
19
20
21
18
19
17
19
20
22
Z P2 Z + Z P1 Z
Z Z
and
Z + Z
n=
Z P1 Z P2
k2
1 +
2
19
20
22
18
24
23
20
16
21
19
38
2.53
2.58
2.61
2.65
2.71
2.77
2.80
2.84
2.85
.10
2.42
2.47
2.50
2.54
2.60
2.66
2.69
2.73
2.73
.15
.10
2.24
2.32
2.36
2.40
2.45
2.50
2.55
2.58
2.61
2.26
.25
.15
2.00
2.11
2.20
2.24
2.26
2.31
2.35
2.41
2.43
2.47
2.47
.40
.25
1.88
1.98
2.06
2.11
2.14
2.18
2.22
2.27
2.29
2.33
2.33
.65
.40
1.45
1.53
1.62
1.72
1.79
1.82
1.85
1.89
1.93
1.98
2.00
2.03
2.04
1.50
1.00
AQL (Tightened)
1.65
1.75
1.84
1.91
1.96
1.98
2.03
2.08
2.12
2.14
2.18
2.18
1.00
.65
AQL (Normal)
1.34
1.40
1.50
1.58
1.65
1.69
1.72
1.76
1.80
1.84
1.86
1.89
1.89
2.50
1.50
All AQL values are in percent nonconforming. T denotes plan used on tightened inspection.
3
4
5
7
10
15
20
25
35
50
75
100
150
200
Sample
size
1.12
1.17
1.24
1.33
1.41
1.47
1.51
1.53
1.57
1.61
1.65
1.67
1.70
1.70
4.00
2.50
0.96
1.01
1.07
1.15
1.23
1.30
1.33
1.35
1.39
1.42
1.46
1.48
1.51
1.51
6.50
4.00
0.77
0.81
0.87
0.96
1.03
1.09
1.12
1.14
1.18
1.21
1.24
1.26
1.29
1.29
10.0
6.50
0.57
0.62
0.68
0.76
0.83
0.89
0.92
0.94
0.97
1.00
1.03
1.05
1.07
1.07
10.0
11:42 AM
B
C
D
E
F
G
H
I
J
K
L
M
N
P
Sample
letter
10/18/07
H1317_CH04.qxd
Page 38
H1317_CH04.qxd
10/18/07
11:42 AM
Page 39
where: P1 = AQL
P2 = RQL
39
The equation for n assumes that is unknown. The calculations for n and k assume a
normal distribution and are independent of the lot size.
The criteria for acceptance of single-sided specifications for either an LSL or a USL
are as follows:
Manufacturers risk = 5%
Consumers risk = 10%
Look up the following Z scores corresponding to the AQL, RQL, , and values:
Z P1 = Z.04 = 1.75
Z P2 = Z.16 = 0.99
Z = Z.05 = 1.65
Z = Z.10 = 1.28
k=
Z P2 Z + Z P1 Z
Z + Z
Z + Z
n=
Z P1 Z P2
k=
(0.99)(1.65) + (1.75)(1.28)
= 1.32
1.65 + 1.28
2
2
k2
1.65 + 1.28 1 + (1.32) = 27.81 = 28
+
=
1
1.75 0.99
2
2
R
,
d2
H1317_CH04.qxd
40
10/18/07
11:42 AM
Page 40
The traditional d2 assumes that the number of ranges used in the determination of R is
large. If the number of subgroups is small, then the d2 value should be adjusted. This
adjustment results in a d 2* value.
0.2778
d2* = d2 1 +
m(nr 1)
n 1
,
0.9(nr 1)
n 1
= 15.
0.9(nr 1)
For this example, 15 subgroups of 3 each would have been chosen, and the d 2* value
is given by
0.2778
d2* = d2 1 +
m(nr 1)
0.2778
d2* = 1.693 1 +
d * = 1.708.
15( 3 1) 2
Z P2 Z + Z P1 Z
Z + Z
k=
H1317_CH04.qxd
10/18/07
11:42 AM
Page 41
41
Sample size n:
2
Z + Z k 2
n=
1 + 2
Z P1 Z P2
2
1.65 + 1.28 (1.44)2
n=
1+
1.96 1.04
2
n = 20.66 = 21
n 1
0.9(nr 1)
m=
11
0.9(5 1)
m = 3.
Three subgroups of subgroup sample size 5 are chosen from the lot.
The appropriate value for d 2* is given by
0.2778
,
d 2* = d 2 1 +
m(nr 1)
0.2778
,
d 2* = 2.326 1 +
6(5 1)
d 2* = 2.353.
Acceptance criteria:
Three samples of five each are chosen, and the average range and grand average are
calculated.
If X + R > 1245, reject the lot; otherwise, accept the lot.
2.380
Bibliography
ASQ Quality Press. 2003. Sampling Procedures and Tables for Inspection by Attributes,
ANSI/ASQ Z1.4-2003. Milwaukee, WI: ASQ Quality Press.
Besterfeld, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Schilling, E. G. 1982. Acceptance Sampling in Quality Control. New York: Marcel
Dekker.
H1317_CH04.qxd
10/18/07
11:42 AM
Page 42
H1317_CH05.qxd
10/17/07
2:03 PM
Page 43
Shewhart control charts are used to detect changes in processes. Changes can occur in both
the average of a process and the variation of the response variable of a process. The simplest control chart for variables data is the individual/moving range control chart. This
chart is relatively insensitive to small changes and subject to distorted results when the
process is significantly nonnormal with respect to the distribution of the data. While all
control charts are robust with respect to nonnormal distributions, the individual/moving
range chart is the one that is most affected by nonnormality of the data. This is due to the
lack of the effect of the central limit theorem, which states that the distribution of averages
will be more normally distributed than the individuals from which they came.
Control charts can be made more sensitive to small process changes by increasing the
sample size (sometimes referred to as subgroup sample size). For sample sizes of 2 to 10,
we may use the average/range control chart.
The central tendency or location of the process is monitored by tracking the average
of individual subgroup averages and their behavior relative to a set of control limits based
on the overall process average plus or minus three standard deviations (for the variation of
averages, not for the variation of individuals).
The variation of the process is monitored by tracking the range of individual subgroups and their behavior relative to a set of control limits.
The control limits for both the averages and the ranges are based on three standard
deviations and will, therefore, reflect the 99.7 percent probability limits. Changes in these
parameters will be noted when certain rules of statistical conduct are violated.
If we elect to have even greater sensitivity for detecting process changes, the choice
for control charts is the average/standard deviation control chart. For this chart, the subgroup sample size may be from 2 to 25. See the module entitled Average/Standard Deviation Chart for a more detailed discussion of these control charts.
The steps for the construction and interpretation of the average/range control chart are
the same as for all variables control charts. The following example illustrates the steps for
establishing an average/range control chart.
Step 1. Collect historical data.
A minimum of 25 subgroup samples of n = 4 to n = 5 are taken over a period of time that
will serve as a baseline period. For our example, measurements will be made approximately every three hours using a subgroup sample size of n = 5 with a total of 16 samples
(k = 16). The total number of actual individual observations will be 5 16 (n k), or 80.
43
H1317_CH05.qxd
44
10/17/07
2:03 PM
Page 44
For each sample of data collected, record the date, time of sample, the individual
observations, the average of the subgroup sample, and the range of the subgroup sample.
The range of a subgroup is calculated by taking the difference between the smallest
and largest values within a subgroup. Ranges are always positive.
Sample #1
Date: 10/2/94
Time: 7:00 AM
Average:
Range:
26.1
31.2
25.2
31.4
24.5
24.3
17.3
26.9
26.6
21.1
26.0
10.9
23.1
8.3
27.7
6.9
23.2
9.6
Sample #6
Date: 10/3/94
Time: 10:30 AM
Sample #7
Date: 10/3/94
Time: 1:10 PM
Sample #8
Date: 10/3/94
Time: 4:20 PM
25.2
22.4
20.9
26.6
27.6
19.7
25.4
23.0
26.5
21.9
20.1
24.9
27.9
28.8
26.4
25.5
23.6
25.2
25.7
23.6
24.5
6.7
23.3
6.8
25.6
8.7
24.7
2.1
Sample #10
Date: 10/4/94
Time: 11:20 AM
Sample #11
Date: 10/4/94
Time: 2:09 PM
Sample #12
Date: 10/4/94
Time: 5:15 PM
27.5
28.5
25.2
28.7
28.0
23.7
19.6
32.6
20.9
23.4
25.4
25.6
27.6
18.6
31.3
21.1
27.4
33.1
27.2
26.3
27.6
3.5
24.0
13.0
25.7
12.7
27.0
12.0
Sample #13
Date: 10/5/94
Time: 7:10 AM
Average:
Range:
Sample #4
Date: 10/2/94
Time: 5:00 PM
19.5
19.9
25.0
23.1
27.8
Sample #9
Date: 10/4/94
Time: 8:10 AM
Average:
Range:
Sample #3
Date: 10/2/94
Time: 2:00 PM
22.4
31.8
20.9
23.6
31.4
Sample #5
Date: 10/3/94
Time: 7:15 AM
Average:
Range:
Sample #2
Date: 10/2/94
Time: 11:00 AM
Sample #14
Date: 10/5/94
Time: 10:00 AM
Sample #15
Date: 10/5/94
Time: 1:00 PM
Sample #16
Date: 10/5/94
Time: 4:25 PM
26.3
22.4
19.5
23.5
21.1
25.9
27.2
22.7
18.5
23.5
27.4
24.8
25.8
25.5
27.1
22.4
19.5
26.1
24.3
25.2
22.6
6.8
23.6
8.7
26.1
2.6
23.5
6.6
H1317_CH05.qxd
10/17/07
2:03 PM
Page 45
45
X1 + X 2 + X 3 + . . . + X k
,
k
R1 + R2 + R3 + . . . + Rk
,
k
Step 3. Determine the control limits for the location and variation statistics.
Sometimes referred to as the grand average, the average of the averages X represents the
best overall location statistic for the process. This statistic is monitored by charting the
individual averages X and their relationship to the grand average X. Control limits are
defined as X 3Sx, and the probability of finding individual averages of subgroups inside these limits is 99.7 percent. The 0.3 percent probability of getting an average outside is so small that when averages outside these limits are obtained, it
is assumed that the process average has changed.
In a similar manner, we will calculate the three standard deviation control limits for
the distribution of the ranges R. Any individual range falling outside the upper control limits (UCL) or lower control limits (LCL) will serve as evidence that the variation of the
process has changed.
Changes in the process will be monitored in the following two areas:
1. Location
2. Variation
H1317_CH05.qxd
46
10/17/07
2:03 PM
Page 46
3 S X = A2 R
where A2 = a factor whose value depends on the subgroup sample size. Control chart factors can be found in Table A.7 of the appendix. For our example, the subgroup sample size
is 5; therefore, A2 = 0.577.
3 S x = A2 R = 0.577 7.81 = 4.51
UCL = 24.89 + 4.51 = 29.40
The LCL is calculated in a similar manner, except that rather than add the three standard deviations (4.51) to the grand average (24.89), we will subtract.
LCL x = X A2 R = 24.89 4.51 = 20.39
There is a 99.7 percent probability that a sample of five observations chosen from this
process will yield an average value between 20.39 and 29.40, provided that the process
average remains 24.89 and the standard deviation of averages for subgroup size n = 5
remains 1.50 (one-third of the three standard deviations, which is estimated to be 4.51).
Since we are using the average of the subgroup ranges R to estimate the three standard
deviations, we need a statistical check to determine if each of the individual ranges used
to calculate the average range is appropriate or statistically acceptable. We check the individual ranges by comparing them to a UCL and LCL for ranges.
UCL R = D 4 R = 2.114 7.81 = 16.51
where the value D4 depends on the subgroup sample size. For our case, D4 = 2.114.
LCL R = D3 R
For a subgroup size n = 5, there is no defined value for D3; therefore, there is no LCL for
the range. There is no LCL for the range until we have a subgroup sample size of n 7.
With our example, we expect 99.7 percent of the time to find subgroup ranges between
0 and 16.51.
Step 4. Construct the control chart and plot the data.
Appropriate plotting scales are determined. Control limits for both the averages and the
ranges are drawn using a broken line, and the grand average or center line for the averages
chart is drawn using a solid line. The completed chart using the historical data follows. The
vertical wavy line separates the history from the future data. The control limits are derived
using only the historical data. If there is supporting evidence in the future that there has
been a real and continuing change in the process, the control limits may be recalculated
using the recent historical data.
Step 5. Continue plotting data, looking for changes in the future.
H1317_CH05.qxd
10/17/07
2:03 PM
Page 47
47
History
UCL = 29.40
30
Average
= 24.89
25
20
LCL = 20.39
15
1
10
15
20
Average
25
30
UCL = 16.51
15
10
0
1
10
15
Range
20
25
30
35
H1317_CH05.qxd
48
10/17/07
2:03 PM
Page 48
There are several rules, or patterns of data as they appear on the control chart, that indicate a lack of control or give us evidence that the process has changed. The rules serve as
an aid in interpreting the control charts. Some of the rules are as follows:
Rule 1. A single point outside the UCL or LCL.
W. A. Shewhart originally devised this rule in 1931, and it was defined as
Criterion I. In a normal distribution where no change in the average has
occurred, this rule would be invoked 0.3 percent of the time, or 1 in 333
observations. This is the rate at which a false alarm would be seen. If the
range chart indicates no change, then we assume that a point outside the
average control limit indicates a change in the process average.
Rule 2. Seven points in a row on the same side of the average or centerline with no
points outside the control limits.
The probability of the pattern of no process change occurring is
approximately the same as getting seven consecutive heads in the toss of a
fair coin. This probability is 1 in 128, or .78 percent. This likelihood is so
remote that if it occurs, we assume that there has been a process change.
Rule 3. Seven consecutive points steadily increasing or decreasing.
This is a test for a trend or drift in the process. A small trend in a process
will be indicated by this rule violation before rule 1 can react. Causes for a
trend can be improvement in skills over time or tool wear.
Rule 4. Two out of three points greater than the two standard deviation limits.
Any two points between the two and three standard deviation limit lines
with the third point being anywhere is an indication of process change. The
probability of getting a false alarm is the same as rule 1.
Rule 5. Four out of five points greater than the one standard deviation limit.
Similar to rule 4 above, the fifth point can be located anywhere.
Rule 1 through rule 3 utilize only the control limits based on 3 standard deviations
and are traditionally the major ones used in statistical process control (SPC). However, if
additional limits based on 1 and 2 standard deviations are chosen, the sensitivity of the
control chart can be increased by invoking the additional rules. For a more detailed discussion, see the module entitled SPC Chart Interpretation.
Assume that the following additional data points are collected in the future after establishing the control limits and operating characteristics based on the historical data. Use
them to determine whether you have evidence that the process has changed.
H1317_CH05.qxd
10/17/07
2:03 PM
Page 49
49
Sample #17
Date: 10/6/94
Time: 8:00 AM
Sample #18
Date: 10/6/94
Time: 11:15 AM
Sample #19
Date: 10/6/94
Time: 2:08 PM
Sample #20
Date: 10/6/94
Time: 4:48 PM
21.8
25.0
17.0
24.4
21.1
19.7
22.1
21.7
24.9
26.2
22.9
24.1
19.7
28.2
23.7
18.7
23.3
23.5
25.2
28.8
Sample #21
Date: 10/7/94
Time: 6:48 AM
Sample #22
Date: 10/7/94
Time: 10:25 AM
Sample #23
Date: 10/7/94
Time: 2:00 PM
Sample #24
Date: 10/7/94
Time: 5:12 PM
23.3
19.4
24.6
30.2
30.4
21.8
21.7
28.3
23.0
27.5
21.2
18.9
27.8
23.9
24.6
27.6
27.7
30.3
21.7
25.0
Sample #25
Date: 10/8/94
Time: 7:45 AM
Sample #26
Date: 10/8/94
Time: 11:15 AM
Sample #27
Date: 10/8/94
Time: 1:55 PM
Sample #28
Date: 10/8/94
Time: 4:48 PM
18.6
23.6
29.1
30.7
22.2
27.2
24.8
24.0
28.2
23.2
22.8
24.4
21.1
28.7
26.7
15.2
25.9
24.1
20.2
24.7
Sample #29
Date: 10/9/94
Time: 8:10 AM
Sample #30
Date: 10/9/94
Time: 11:09 AM
Sample #31
Date: 10/9/94
Time: 2:20 PM
31.5
28.0
24.7
27.2
25.9
22.7
17.8
29.5
25.4
28.3
25.7
21.5
28.0
25.5
26.2
With this new information, do you feel that the process average and/or standard deviation have changed?
H1317_CH05.qxd
10/17/07
50
2:03 PM
Page 50
History
UCL = 29.40
30
Average
= 24.89
25
20
LCL = 20.39
15
1
10
15
20
Average
25
30
35
UCL = 16.51
15
10
0
1
10
15
Range
20
25
30
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Wheeler, D. J., and D. S. Chambers. 1992. Understanding Statistical Process Control.
2nd edition. Knoxville, TN: SPC Press.
H1317_CH06.qxd
10/18/07
12:53 PM
Page 51
Control limits:
For X : X A
For R : D1 (Lower control limit)
D2 (Upper control limit)
All the factors, d2, A, Dl, D2, and so on, are listed in the following table:
Subgroup size n
d2
D1
D2
2
3
4
5
6
7
8
9
10
2.196
4.130
6.094
8.074
10.060
12.051
14.044
16.038
18.034
1.752
3.660
5.615
7.589
9.572
11.561
13.553
15.546
17.541
1.128
1.693
2.059
2.326
2.534
2.704
2.847
2.970
3.078
2.121
1.732
1.500
1.342
1.225
1.134
1.061
1.000
0.949
0
0
0
0
0
0.205
0.387
0.546
0.687
3.686
4.358
4.698
4.918
5.078
5.203
5.307
5.394
5.469
51
H1317_CH06.qxd
52
10/18/07
12:53 PM
Page 52
The process variation can be estimated from either the sample range R or the sample standard deviation S:
R
R
R
d2
=
,
n1 + n2 + + nk k
=
(1)
g1S1 + g2 S2 + + gk Sk
.
h1 + h2 + + hk
(2)
Sample #
1
2
3
4
5
6
7
8
9
10
Data
Subgroup size n
R
d2
6
4
5
4
5
2
4
4
5
6
12.2
10.5
10.8
10.25
12.4
12.0
13.25
12.75
12.6
13.0
3
1
5
2
3
2
3
5
4
3
1.169
0.577
1.924
0.957
1.140
1.414
1.258
2.630
1.673
1.265
1.18
0.49
2.15
0.97
1.29
1.77
1.46
2.43
1.72
1.18
n1 X 1 + n2 X 2 + + nk X k
n1 + n2 + + nk
X=
RCL = d2 , nk ( )
H1317_CH06.qxd
10/18/07
12:53 PM
Page 53
53
Each individual subgroup will have a unique variation statistic that is dependent on the
subgroup sample size. In order to determine RCL, we will need an estimate of the process
variation estimated from the sample ranges:
R
R
R
d2
=
n1 + n2 + + nk k
=
g1S1 + g2 S2 + + gk Sk
h1 + h2 + + hk
Note: Differences in the two estimates are due to rounding error. We will use the estimate
for sigma based on the range method.
Centerlines for the ranges, RCL = d2 , nk ( )
= 1.436
Subgroup number
Subgroup size n
d2
RCL
1
2
3
4
5
6
7
8
9
10
6
4
5
4
5
2
4
4
5
6
2.534
2.059
1.693
2.059
1.693
1.128
2.059
2.059
1.693
2.534
3.64
2.96
2.43
2.96
2.43
1.62
2.96
2.96
2.43
3.64
X = 12.03
= 1.4366
H1317_CH06.qxd
54
10/18/07
12:53 PM
Page 54
Subgroup number
Subgroup size n
LCL
UCL
1
2
3
4
5
6
7
8
9
10
6
4
5
4
5
2
4
4
5
6
1.225
1.500
1.342
1.500
1.342
2.121
1.500
1.500
1.342
1.225
10.27
9.88
10.10
9.88
10.10
8.98
9.88
9.88
10.10
10.27
13.78
14.18
13.96
14.18
13.96
15.08
14.18
14.18
13.96
13.78
= 1.436.
Subgroup number
Subgroup size n
D2
UCL
1
2
3
4
5
6
7
8
9
10
6
4
5
4
5
2
4
4
5
6
5.078
4.698
4.918
4.698
4.918
3.686
4.698
4.698
4.918
5.078
7.29
6.75
7.06
6.75
7.06
5.29
6.75
6.75
7.06
7.29
H1317_CH06.qxd
10/18/07
12:53 PM
Page 55
55
History
15
14
13
12
11
10
9
8
7
6
1
10
11
12
10
11
12
Subgroup number
8
7
6
5
4
3
2
1
1
Subgroup number
Step 5. Continue collecting data and look for signs of a process change.
All of the traditional rules are applicable when using a variables size average/range control chart.
Data to the left of the wavy vertical line represent the historical data from which the
location statistic average and the estimate of sigma were derived. Future data points will be
compared with the historical reference values. Subsequent values for the range center line
RCL, UCLR/LCLR, UCLX, and LCLX will be determined and values of R and X plotted.
Consider the following set of data for subgroups 11 and 12:
Subgroup
number
11
12
Data
10,
11,
14,
15,
9, 10, 10, 9,
9, 10, 10
13, 15, 13, 15,
12, 14, 15, 14
10
Dl
D2
D2
9.78
1.000
0.546
5.394
2.970
14.00
0.949
0.687
5.469
3.078
H1317_CH06.qxd
56
10/18/07
12:53 PM
Page 56
X = 12.03
= 1.4366
= 1.436
UCLR = D2
UCLR = (5.394)(1.436) = 7.75
LCLR = D1
= 1.436
Average = 9.78
Ran
nge = 2
X = 12.03
= 1.4366
= 1.436
UCLR = D2
UCLR = (5.496)(1.436) = 7.85
LCLR = D1
= 1.436
Average = 14.00
Raange = 3
H1317_CH06.qxd
10/18/07
12:53 PM
Page 57
57
The completed plot can be seen in the following figure. Notice that both samples #11
and #12 are out of control on the location chart and in control on the variation chart.
15
14
13
12
11
10
9
8
7
6
1
10
11
12
10
11
12
8
7
6
5
4
3
2
1
H1317_CH06.qxd
10/18/07
12:53 PM
Page 58
H1317_CH07.qxd
10/17/07
2:07 PM
Page 59
Average/Standard Deviation
Control Chart
The traditional Shewhart control charts used to monitor variables are the individual/moving
range, average/range, and the average/standard deviation. The individual/moving range chart
is used when the observations are treated as individuals. In order to increase the sensitivity
of the chart to detect a process change, the average/range control chart is used. The process
variation is measured using the range. The average range R is used to monitor the process
variation and is also used to estimate the variation of the averages using the relationship
3Sx = A2 R,
where A2 is a constant whose value depends on the subgroup sample size.
This relationship is effective as an estimate of the standard deviation for the distribution of averages provided that the subgroup sample size is 10 or less. Sample subgroup
sizes greater than 10 render this relationship less efficient, and the following alternative
estimate must be used:
3Sx = A3S .
Increasing the sample size also increases the sensitivity of the control chart. The
average/standard deviation chart may be used for any subgroup sample size of n 2.
The five steps for establishing the average/standard deviation control chart are as follows:
1.
2.
3.
4.
5.
H1317_CH07.qxd
60
X:
S:
10/17/07
2:07 PM
Page 60
49.7
46.7
46.5
46.2
52.0
46.3
47.0
51.8
47.0
53.0
48.2
51.4
45.9
52.3
50.0
52.9
48.3
45.7
43.4
51.3
47.0
50.5
49.3
48.9
48.6
53.4
53.1
50.4
55.1
43.8
49.8
47.9
50.9
48.7
53.2
57.3
52.3
49.4
48.3
48.2
56.5
47.3
52.4
52.9
51.4
54.0
56.5
52.6
50.9
46.8
43.9
45.4
44.1
48.8
48.1
2.4
49.7
2.9
49.5
3.1
50.8
4.0
51.0
2.9
49.2
4.5
10
11
12
47.3
50.0
51.7
56.3
53.0
48.3
47.8
49.7
52.6
49.6
48.6
49.3
51.5
50.8
49.4
51.1
51.5
50.4
51.8
52.6
45.4
51.1
48.0
51.3
52.1
47.2
51.3
51.9
48.2
50.4
51.2
51.8
50.5
50.9
47.4
47.7
48.2
46.5
45.5
47.7
50.9
49.9
53.5
45.5
51.7
49.4
46.3
52.9
53.5
55.3
50.7
46.8
51.7
50.5
50.7
2.9
50.2
1.1
50.1
2.5
50.0
1.8
48.8
2.8
50.8
2.9
S=
Step 3. Determine control limits for both the location and variation statistic.
A. Control limits for the location statistic:
The location or central tendency of this process will be monitored by the movement of the averages about the grand average. Each of the individual averages varies. Upper control limits
(UCLs) and lower control limits (LCLs) for these individual averages will be determined based
on the grand average 3 standard deviations (standard deviation of the averages, not of the individual observations). The three standard deviations will be determined from
3Sx = A3S .
The value of A3 is dependent on the subgroup sample size n. For this case, n = 9 and
A3 = 1.032 (see Table A.7 in the appendix).
3 S x = (1.032 ) ( 2.8 ) = 2.89 = 2.9
The UCL and LCL for the distribution of averages are, respectively,
UCL = 49.9 + 2.9 = 52.8
LCL = 49.9 2.9 = 47.0.
Since these limits are based on 3 standard deviations, 99.7 percent of the averages
are expected to fall between 47.0 and 52.8.
H1317_CH07.qxd
10/17/07
2:07 PM
Page 61
61
UCL s = B4 S
LCL s = B3S
Approximately 99.7 percent of the sample standard deviations will fall within the limits of 1.1 to 4.9.
Step 4. Construct the control chart.
The averages, control limits, and data values are plotted on the chart. Control limits are
normally drawn using a broken line, and averages are drawn using a solid line.
History
UCL = 52.4
X = 49.5
50
X
X
X
X
X
X
LCL = 46.6
45
40
Average
UCL = 4.9
S = 2.8
4
X
2
LCL = 1.1
1
0
1
7
8
9 10 11 12
Standard deviation
Note: The vertical wavy line is drawn only to show the period from which the chart
parameters were determined (averages and control limits).
H1317_CH07.qxd
62
10/17/07
2:07 PM
Page 62
Step 5. Continue data collection and plotting, looking for signs of a process change.
X:
S:
13
14
15
16
17
18
46.9
53.3
47.5
52.5
50.7
47.9
50.1
54.9
54.5
52.4
54.0
51.5
51.8
50.6
52.7
50.7
55.9
55.5
52.3
49.4
50.3
51.3
56.5
50.3
52.4
52.9
51.4
51.9
50.7
51.2
53.5
53.0
51.2
53.2
53.8
52.4
50.4
49.7
51.5
51.1
48.0
51.3
52.1
47.2
51.3
50.2
46.7
46.5
52.6
55.9
52.0
53.4
48.9
49.2
50.9
3.1
52.8
2.3
51.9
2.1
52.3
1.1
50.3
1.7
50.6
3.1
History
UCL = 52.4
X
X
50
X
X
X X
X
X
X
LCL = 46.6
45
40
Average
UCL = 4.9
S = 2.8
4
X
X
X
2
X
LCL = 1.1 1
X
X
X
X
X
0
1
7 8 9 10 11 12 13 14 15 16 17 18
Standard deviation
H1317_CH07.qxd
10/17/07
2:07 PM
Page 63
63
Sample #14 has an average greater than the UCL for averages, indicating that the
process, relative to the historical characterization, has experienced a change. It is suspected that the process average has increased. This is an example of the violation of one
of the SPC detection rules (a single point outside the UCL or LCL). The process variation
remains stable with no indication of change.
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Wheeler, D. J., and D. S. Chambers. 1992. Understanding Statistical Process Control.
2nd edition. Knoxville, TN: SPC Press.
H1317_CH07.qxd
10/17/07
2:07 PM
Page 64
H1317_CH08.qxd
10/15/07
11:47 AM
Page 65
Chart Resolution
The ability of a control chart to resolve the difference between its historical characterization (the statistics used to establish the chart) and future process data is a measure of the
power or resolution of the chart. Several interrelated factors affect this resolution:
1.
2.
3.
4.
Two families of control charts will be reviewed for their resolving power as related to
these factors:
1. Control charts for variables
A. Individuals
B. Averages
2. Control charts for attributes
A. Proportion (defective)
B. Nonconformities (defects)
Consider an individuals control chart developed from a process that has an average X = 35.00
The control limits are calculated upon X 3S, where 3S is estimated by 2.66 MR .
3S = (2.66)(9.0) = 23.94 24.00
S = 8.00
UCL = 59.00
LCL = 11.00
X = 35.00
S = 8.0
65
H1317_CH08.qxd
66
10/15/07
11:47 AM
Page 66
59.0 35.0
= 3.00.
8.00
Looking up this value in a standard normal distribution, the probability of getting a value
greater than 59.00 is 0.00135, or a percent probability of 0.135 percent. The probability of
getting a value less than 11.00 is also 0.00135, for a combined probability of 0.0027.
1
= 370
0.0027
There is a probability that 1 sample in 370 will exceed the control limit when there has
been no process change. The value of 370 is also called the average run length (ARL).
Once points of interest such as the upper control limits (UCLs) and the lower control
limits (LCLs) have been determined, the probability of exceeding either or both can be
determined assuming any future process average.
For example, if the UCL and LCL for an individual chart are 59.00 and 11.00, respectively, and the standard deviation of individuals is 8.00, what is the probability of getting
a single point above the UCL if the process average shifts from 35.00 to 43.00 (an increase
of + 1.5)?
UCL = 59.00
X = 47.00
S = 8.0
Z=
59.00 47.00
= 1.50
8.00
H1317_CH08.qxd
10/15/07
11:47 AM
Page 67
Chart Resolution
67
The probability of Z = 1.50 is 0.0668, or 6.68 percent, and the ARL = 15. This means
that if a control chart is established with a UCL and an LCL of 59.00 and 11.00, respectively, and the process average shifts from 35.00 to 47.00, there is a probability of .0668
that a single point will exceed the UCL. One sample in 15 will exceed the UCL.
Calculations can be performed in a similar manner to determine the probability of
detection with various degrees of process shift (in units of ). Table 1 lists several degrees
of process change, the corresponding probability of exceeding a control limit, and the
ARL. A plot of Table 1 is seen in Figure 1.
Table 1 Probability of exceeding control limit as a function of shift.
Amount of shift,
Probability of detection
ARL
0.00
0.20
0.40
0.60
0.80
1.00
1.20
1.40
1.60
1.80
2.00
2.20
2.40
.00135
.00256
.00466
.00820
.01390
.02275
.03593
.05480
.08076
.11507
.15866
.21186
.27425
741
391
215
122
72
44
28
18
12
9
6
5
4
Probability of exceeding
a control limit
0.30
0.25
0.20
0.15
0.10
0.05
0.00
.2
.6
1.0
1.4
1.8
2.2
H1317_CH08.qxd
68
10/15/07
11:47 AM
Page 68
To this point, all the probabilities have been based on exceeding a limit with a single
sample. We can determine the probability of exceeding a control limit within any number
of samples k using the relationship
Pk = i = 1 (1 Pk =1 ) ,
k
where:
Example:
A process has been defined by an average of 26.5 and a standard deviation of 3.8. What is
the probability of detecting a shift in the average to 25.5 within 10 samples after the shift
has occurred?
UCL = X + 3S = 26.5 + 3(3.8) = 37.9
LCL = X 3S = 26.53(3.8) = 15.1
LCL = 15.1
X = 25.5
S = 3.8
25.5 15.1
= 2.74.
3.8
P10 = 1 (1 0.00307 )
10
Pk =10 = 0.0302.
H1317_CH08.qxd
10/15/07
11:47 AM
Page 69
Chart Resolution
69
k=
shift
10
15
20
0.2
0.4
0.6
0.8
1.0
1.2
1.4
1.6
1.8
2.0
2.2
2.4
.0026
.0047
.0082
.0139
.0228
.0359
.0548
.0808
.1151
.1587
.2119
.2743
.0051
.0093
.0163
.0276
.0450
.0706
.1066
.1550
.2169
.2922
.3788
.4733
.0077
.0139
.0244
.0411
.0667
.1040
.1556
.2232
.3070
.4045
.5104
.6177
.0102
.0185
.0324
.0544
.0879
.1362
.2018
.2860
.3868
.4989
.6142
.7226
.0129
.0231
.0403
.0676
.1087
.1672
.2456
.3436
.4573
.5784
.6959
.7986
.0253
.0456
.0790
.1306
.2056
.3064
.4308
.5692
.7055
.8223
.9075
.9595
.0377
.0677
.1162
.1894
.2919
.4224
.5706
.7172
.8402
.9251
.9719
.9918
.0500
.0892
.1518
.2442
.3689
.5190
.6761
.8144
.9133
.9684
.9915
.9984
Table 2 gives the probability of a single point being outside a control limit within
k samples of a process shift of units.
1. From the SPC relationship 3Sx = A2R, where A2 = a constant dependent on the sub
group sample size n and R = the average range
process average X = 15.00, and the average range R = 4.8. What is the standard deviation
of the averages?
3 S x = A2 R 3 S x = ( 0.577 ) ( 4.8 ) = 2.77
Sx =
2.77
= 0.92
3
Example of application 2:
The standard deviation for the individuals of a process is S = 8.55. What is the standard
deviation for the distribution of averages of n = 7 from this process?
H1317_CH08.qxd
70
10/15/07
11:47 AM
Page 70
n
S = 8.55 n = 7
Sx =
Sx =
8.55
= 3.23
7
Having the standard deviation for the distribution of averages, we can calculate the
probability of a single point exceeding a control limit in one sample or the probability of
a single point exceeding a limit within k samples of a process change.
Example:
What is the probability of an average of n = 6 exceeding the UCL within eight samples
after the process shifts from the historical average of X = 45.00 to 47.50? The UCL is 50.0,
and the LCL is 40.00.
6 S x = UCL LCL = 10.0 S x = 1.67
UCL = 50.00
X = 47.50
The Z-score is
Z=
50.00 47.50
= 1.50 .
1.67
The probability of getting a Z = 1.50 is 0.0668. The probability of exceeding the UCL
on a single sample average is 0.0668.
If the probability of exceeding the UCL on a single sample is 0.0668, then the probability of exceeding the UCL within eight samples after the process shift is
Pk=i = 1 (1 Pk=l)k,
where:
k=8
Pk = 1 = 0.0668
P8 = 1 (1 0.0668)8
P8 = 0.4248.
H1317_CH08.qxd
10/15/07
11:47 AM
Page 71
Chart Resolution
71
Table 3 Probability of exceeding a limit for averages of subgroup size n and shift .
Subgroup n
Shift
0.2
0.4
0.6
0.8
1.0
1.2
1.4
1.6
1.8
2.0
2.2
2.4
2
.0033
.0075
.0157
.0308
.0564
.0963
.1539
.2306
.3249
.4321
.5445
.6534
.0040
.0105
.0249
.0531
.1023
.1782
.2825
.4095
.5468
.6788
.7913
.8765
.0047
.0139
.0359
.0808
.1587
.2743
.4207
.5793
.7257
.8413
.9192
.9641
.0054
.0176
.0487
.1131
.2228
.3762
.5525
.7188
.8478
.9298
.9727
.9911
.0060
.0216
.0628
.1488
.2907
.4756
.6660
.8210
.9206
.9713
.9916
.9980
.0069
.0267
.0804
.1913
.3655
.4265
.7625
.8931
.9618
.9893
.9977
.9996
.0076
.0311
.0969
.2313
.4327
.6538
.8316
.9364
.9817
.9960
.9994
.9999
9
.0081
.0358
.1148
.2741
.5000
.7259
.8852
.9642
.9919
.9987
.9998
.9999
10
.0089
.0412
.1347
.3186
.5641
.7865
.9232
.9803
.9965
.9996
.9999
.9999
Table 3 gives the probability of exceeding a control limit for various subgroup sample
sizes and various degrees of process shift in units of in a single sample.
What about Subgroup Size for the Variables Control Charts?
The average/range control chart is used only with subgroup sample sizes of 2 to 10. For
subgroups larger than 10, we use the average/standard deviation control chart. However,
the average/standard deviation control chart can be used with subgroup sample sizes of 2
to 10. The question arises, How large should my subgroup sample size be, and how often
should I take a sample? The usual response is that most people use five, and you should
take a sample every hour. While you could take a guess, there is actually a better way to
make these decisions. It depends on three factors:
1. How large of a shift do you want to detect?
2. How soon do you want to detect this shift after it occurs?
3. What is the probability that you want in detecting the shift?
Consider the following scenario:
A gypsum board manufacturer tests the edge hardness of its boards prior to shipment.
An individual/moving range control chart has been established (subgroup size is n = l), and
samples are obtained every hour. The process average is 10.0, and the average moving
range is 1.128.
If the process average shifts from 10.0 to 11.0, what is the probability you will detect
the shift with one sample after the shift occurs? What is the probability of detection within
10 samples after the shift? Detection is indicated by a single point falling above or below
the control limits. Assume a sample is taken each hour.
Determine the probability of detecting the shift with one subgroup (detection in
one hour).
H1317_CH08.qxd
72
10/15/07
11:47 AM
Page 72
The control limits for our average/range chart are determined by X 2.66 MR or l0.0
(2.66)(1.128).
LCL = 10.0 3.0 = 7.0
2.66 MR = 3
Therefore =
2.66(1.128)
= 1.00
3
If we shift the process from its original average of 10.0 to a new average of 11.0 and
sigma is 1.0, we will experience a one-sigma shift.
UCL = 13
X = 10.0
=1
Original process.
13 11
= 2.00.
1
Looking up the Z-score in a standard normal distribution table (Table A.1 in the appendix), we find the probability of exceeding the UCL is 0.02275. We will call this P1, the
probability of detection in one sample after the shift occurs.
We now shift the process up by one standard deviation, keep the UCL and the standard
deviation the same, and recalculate the probability of detection (detection in one sample
after the shift).
UCL = 13
X = 11.0
=1
H1317_CH08.qxd
10/15/07
11:47 AM
Page 73
Chart Resolution
73
This is a 110 percent improvement, but we can improve this. We can increase k to a
larger value such as 15.
Pk = 1 (1 P1)k
ln Pk 1
ln(1 P1 )
k=
1.897
ln (0.15)
=
= 82.6 83 samples
ln (1 0.02275) 0.023
P1 = 0.02275 Px = 0.85
In other words, the probability of detecting a one-sigma shift within 83 samples after
the shift is 85 percent. Making this detection would take 83 hours, or about 3 1/2 days
a long time. If we want to detect this shift quicker, we could sample more frequently but
we would still sample 83 times. If we sample every 7 minutes, we will detect the shift in
11.8 hours, or about half a day. This probability of detecting a one-sigma shift applies to
any process regardless of the value of the process average and the process sigma. The
probability of detecting a one-sigma shift (up or down) within 83 samples is 85 percent
using an individual control chart.
We could also consider using an average/range control chart with a subgroup size of
n = 5. The upper and lower control limits will now be based on the standard deviation for
the distribution of averages (n = 5) rather than the standard deviation of individuals (sometimes referred to as sigma). Using the central limit theorem, we know that the standard
deviation of averages is
SX =
.
n
H1317_CH08.qxd
74
10/15/07
11:47 AM
Page 74
Recall that in our example, sigma for the process was 1.0 and the process average was
10.0. Our control limits will be based on three standard deviations for the distribution of
averages, or
SX =
1
= 0.45.
5
UCL = 11.35
X = 10.0
Sx = .045
We use the standard deviation for averages because we are using averages (of n = 5)
to detect our shift of one sigma. We now shift our process up one sigma from 10.0 to 11.0,
keeping the UCL at 11.35.
The probability of detecting this shift in one subgroup P1 is determined by calculating
the Z-score.
Z=
11.35 11.0
= 0.78
0.445
H1317_CH08.qxd
10/15/07
11:47 AM
Page 75
Chart Resolution
75
This is quite an improvement. To judge our increase in efficiency, lets look at the
actual workload. With the individual control chart, we had a 20.56 percent probability of
detecting a one-sigma shift. With the average control chart where n = 5, we had a probability of detection of 91.42 percent. To make a fair comparison, we should evaluate the
probability of detection within 50 individual samples since 10 samples of 5 each equates
to 50 individual observations.
Pk = 1 (1 P1)k
where: P1 = 0.2275 and k = 50
Pk = 1 (1 0.2275)50 = 0.06836, or 68.36%
Everything being equal, the average control chart is more efficient than the individual
control chart. Fifty observations as individuals gives a probability of detection of 68.36
percent. Fifty observations taken as 10 samples of 5 each gives a 91.42 percent probability of detection.
Illustrative example 1:
The current process average is 18.5, and the process standard deviation is 2.8. You would
like to detect a shift of 1.3 units within 24 hours after the shift with a 90 percent probability of detection. Design a control chart that will accomplish this objective.
The amount of shift in units of the process standard deviation (sigma) is
1.3
= 0.46.
2.8
We have two unknowns in our problem: the number of samples in our subgroup size
n and the number of samples for detection k. We must fix one and determine the other.
There are two ways to accomplish our objective. For a fixed probability of detection Pk we
can:
1. Decrease P1 and increase the number of samples k
2. Increase P1 and decrease the number of samples k
Lets consider using an average control chart where the subgroup sample size is n = 5.
UCL = 22.25
X = 18.5
= 2.8
Sx = 1.25
H1317_CH08.qxd
76
10/15/07
11:47 AM
Page 76
Shifting the average up by 1.3 units and keeping the UCL at 22.25, we have:
UCL = 22.25
X = 19.8
Sx = 1.25
The probability of detection in one sample (of n = 5) is determined from the Z-score:
22.25 19.8
= 1.96
1.25
1.96 = 0.025.
Z=
ln ( Pk 1) ln (0.90 1) 2.303
= 92.
=
=
.025
ln(1 P1 )
ln(1 .025)
There is a 90 percent probability of detection of the 0.45 sigma shift within 92 samples
of n = 5. This would require taking a subgroup of n = 5 approximately every 15 minutes.
Illustrative example 2:
The process average is 45.0 and the standard deviation is 3.7. You have established an
average/range control chart with n = 8. What is the probability of detecting a shift from
45.0 to 46.3 within 12 samples after the shift?
UCL = 48.93
X = 45.0
Sx = 1.31
SX =
Prior to shift, initial condition.
3.7
=
=1
n 2.83
H1317_CH08.qxd
10/15/07
11:47 AM
Page 77
Chart Resolution
77
UCL = 48.93
X = 46.3
Sx = 1.31
After shift.
II. k =
ln ( Pk 1)
ln(1 P1 )
III. P1 = 1 ( Pk 1)
Pk = the probability of detecting the shift within k samples after the shift occurs
P1 = the probability of detecting the shift within one sample (or subgroup) after the
shift occurs
k = the number of samples (or subgroups)
H1317_CH08.qxd
78
10/15/07
11:47 AM
Page 78
P 3
UCL = 0.295
LCL = 0.065.
If the process average proportion P shifts, we may calculate the probability of exceeding a control limit. For proportions less than or equal to 0.10, we will use the Poisson distribution to estimate probabilities, and for proportions greater than 0.10, we will use the
normal distribution as an approximation to the binomial.
For P <
0.10, Poisson Distribution
Sample Calculations
Assume that the process average has shifted from the historical average of 0.10 to 0.08.
What is the probability of getting a single point below the LCL, or 0.065?
Using the Poisson model, we are determining the probability of getting less than
(0.065)(100) defective (or to the nearest integer, seven, defective) in a sample chosen from
a process that has an average number defective np = (100)(.08) = 8.0.
Using the Poisson relationship, we can determine the probability of getting seven or
fewer defective units in a sample of 100 chosen from a process that is 0.08 proportional
defective. This is done by obtaining the probability of getting exactly zero defectives plus
exactly one defective plus exactly two defectives, and so on, until the cumulative probability of getting seven defectives has been reached. This cumulative probability represents
the probability of getting seven or fewer defective units in a sample of n = 100 and represents the probability of not exceeding the UCL. The probability of exceeding the UCL is
1.00 minus this probability.
The Poisson formula is:
Px =
np
( np)x ,
x!
np
( np)x ,
x!
Px = 3 =
[(100 )( 0.05)]
[(100 )( 0..05)]3 , P
3!
x =3
e 5 (5)3
, Px = 3 = 0.1404.
3!
H1317_CH08.qxd
10/15/07
11:47 AM
Page 79
Chart Resolution
79
The probability of getting three or fewer defective units from a sample of n = 100 chosen from a process that is 0.05 proportional defective is the sum of all the individual probabilities from zero defective through three defective:
e 5 ( 5 )
e 5 ( 5 )
e 5 ( 5 )
e 5 ( 5 ) ,
+
+
+
0
1!
2!
3!
0
Px 3 =
UCL = 0.295
P = 0.250
SP = 0.0433
Sp =
P 1 P
n
SP =
0.25 (1 0.25 )
100
Calculation of the standard deviation was based on a binomial distribution for attribute
data. The probability that a given sample proportion will exceed the UCL is determined by
using the normal distribution.
Z=
0.295 0.250
= 1.04
0.0433
H1317_CH08.qxd
80
10/15/07
11:47 AM
Page 80
Additional values for various process shifts have been made for the control chart based on
a sample size of n = 100 and a historical process proportional defective of P = 0.18. Table 4
lists the probabilities of exceeding a limit for those cases. All probabilities for shifts to p 0.10
are based on the Poisson, and probabilities for shifts to P > 0.10 are based on the normal
distribution.
Table 4 Resolution of an n = 100, P = 0.18 p chart.
P
0.01
0.02
0.03
0.04
0.05
0.06
0.07
0.08
0.09
0.10
0.11
0.12
0.13
0.14
0.15
0.16
0.17
0.18
0.19
Probability Probability
< LCL
> UCL
1.000
0.995
0.988
0.949
0.867
0.744
0.599
0.453
0.324
0.220
0.075
0.045
0.027
0.018
0.009
0.005
0.003
0.001
0.001
Total
Probability Probability Probability
Total
probability
P
< LCL
> UCL
probability
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.001
0.004
1.000
0.995
0.988
0.949
0.867
0.744
0.599
0.453
0.324
0.220
0.075
0.045
0.027
0.018
0.009
0.005
0.003
0.002
0.005
0.20
0.21
0.22
0.23
0.24
0.25
0.26
0.27
0.28
0.29
0.30
0.31
0.32
0.33
0.34
0.35
0.36
0.37
0.38
0.001
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.009
0.184
0.035
0.061
0.099
0.149
0.213
0.287
0.369
0.456
0.543
0.627
0.704
0.772
0.829
0.876
0.912
0.940
0.960
Probability of exceeding
a control limit
1.0
n = 100
P = 0.18
.9
.8
.7
.6
.5
.4
.3
.2
.1
.05
.10
.15
.20
.25
.30
New process proportion defective
.35
.40
0.010
0.184
0.035
0.061
0.099
0.149
0.213
0.287
0.369
0.456
0.543
0.627
0.704
0.772
0.829
0.876
0.912
0.940
0.960
H1317_CH08.qxd
10/15/07
11:47 AM
Page 81
Chart Resolution
81
A plot of total probability as a function of P from the data in Table 4 can be seen in
Figure 2. This plot is called an OCC and completely defines the resolution or power for
the control chart.
Bibliography
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Wheeler, D. J. 1995. Advanced Topics in Statistical Process Control. Knoxville, TN: SPC
Press.
Wheeler, D. J., and D. S. Chambers. 1992. Understanding Statistical Process Control.
2nd edition. Knoxville, TN: SPC Press.
H1317_CH08.qxd
10/15/07
11:47 AM
Page 82
H1317_CH09.qxd
10/15/07
11:55 AM
Page 83
Chi-Square Contingency
and Goodness-of-Fit
Chi-Square Contingency Tables
The chi-square distribution is frequently used in statistics to test for the independence of
two factors. Data are arranged in a tabular fashion with rows and columns. In the following example, we want to test for the independence of the two factors of dealership and type
of vehicle sold.
Example 1:
Four dealerships are evaluated by the number of three different types of vehicles sold in
one month. The objective is to determine if there is a dependency on the types of vehicles
sold as a function of the dealership.
We will construct a contingency table consisting of four columns (dealerships) and
three rows (types of vehicles). The response is the number of vehicles sold in one month.
Dealership
Type of vehicle
Row totals
Truck
Compact
Midsize
26
18
33
38
24
45
75
32
68
53
28
49
192
102
195
Column totals
77
107
175
130
489
(Grand total)
The null hypothesis assumes that the distribution of the number of a particular type of car
sold is independent of the dealership. That is,
H0: assumes no real differences among the dealerships with respect to distribution of sales.
The expected frequency of sales within each class is given by:
expected value =
H1317_CH09.qxd
84
10/15/07
11:55 AM
Page 84
Dealership
Type of vehicle J.R.s Auto Crazy Harrys
Abes
Truck
Compact
Midsize
26(30.2)
18(16.1)
33(30.7)
75(68.7)
32(36.5)
68(69.8)
Column totals
77
38(42.0)
24(22.3)
45(42.7)
107
Reliable Roscos
53(51.0)
28(27.1)
49(51.8)
175
130
Row totals
192
102
195
489
(Grand total)
E
30.2
42.0
68.7
51.0
16.1
22.3
36.5
27.1
30.7
42.7
69.8
51.8
OE
4.2
4.0
6.3
2.0
1.9
1.7
4.5
0.9
2.3
2.3
1.8
2.8
(O E)2
17.64
16.00
6.30
4.00
3.61
2.89
20.25
0.81
5.29
5.29
3.24
7.84
(O E)2/E
0.58
0.38
0.58
0.08
0.22
0.13
0.55
0.03
0.17
0.12
0.05
0.15
H1317_CH09.qxd
10/15/07
11:55 AM
Page 85
85
If we chose a risk of 5 percent, then the critical chi-square is 26.0.05 = 12.6. Since the
calculated chi-square is not greater than the critical chi-square, we cannot reject the
hypothesis that the types of vehicles sold is independent of the dealership. The two factors
are independent of each other at the level of significance or risk of 5 percent.
Example 2:
A paint tester wishes to determine whether there is a relationship between the type of UV
inhibitor and the shade of paint with respect to the length of time required to cause a test
specimen to lose 15 percent of its original gloss. One hundred and fifty samples are tested
using UV inhibitor types A and B with both a light and dark color of paint. A summary of
the test data in each of the four categories is shown in the following table. Do the data provide sufficient evidence to indicate a relationship between UV inhibitor type and shade of
paint at a significance of = 0.10?
Type of UV inhibitor
Shade of paint
Type 101
Type 320
Row totals
Light
Dark
47
21
32
50
79
71
Column totals
68
82
150
(Grand total)
79 68
= 35.8
150
71 68
= 32.2
150
79 82
= 43.2
150
71 82
= 38.8
150
Type of UV inhibitor
Shade of paint
Light
Dark
Column totals
Type 101
Type 320
Row totals
47(35.8)
21(32.2)
32(43.2)
50(38.8)
79
71
68
82
150
(Grand total)
H1317_CH09.qxd
10/15/07
86
11:55 AM
Page 86
Calculation of 2:
2 =
32.2
43.2
38.8
= 13.53
Calculation of 2Critical:
Degrees of freedom = ( r 1) ( c 1) = ( 2 1) ( 2 1) = 1
Significance or risk = = 0.10
2 1 , 0.10 = 2.71
Since the chi-square calculated (13.53) is greater than the chi-square critical (2.71), we
reject the hypothesis that the type of UV inhibitor and shade of paint are independent with
respect to the time required to lose 15 percent of the original gloss.
Performance of the type of inhibitor is related to the shade of paint.
Note: There is an assumption that the sample size is sufficiently large such that each
category will have at least five observations. This may require that some categories be
combined to give a minimum count of five.
Goodness-of-Fit
A goodness-of-fit test is a statistical test to determine the likelihood that sample data have
been generated from a population that conforms to a particular distribution such as a normal, binomial, Poisson, uniform, and so on.
The objective is to compare the sample distribution with the expected distribution. The
null hypothesis would be that the sample data come from a specified distribution with a
defined mean and standard deviation. The alternative hypothesis would be that the sample
data come from some other type of distribution.
Goodness-of-Fit to Poisson
The following data represent the number of defects found on printed circuit boards.
Number of defects found:
3 4 5
H1317_CH09.qxd
10/15/07
11:55 AM
Page 87
87
The probability of getting exactly zero defects based on the Poisson distribution where
= 1.53 is
Px =0 =
e x
= 0.217.
x!
In a sample of n = 146, we would expect to find (146)(0.217) = 31.6 boards with zero
defects.
For these data the observed number of boards with zero defects was 32, and the
expected value is 31.6.
The probability of finding exactly one defect is given by:
e1.53 (1.53)
e x
= 0.331.
=
= Px =1 =
x!
1
1
Px =0
In a sample of n = 146, we would expect to find (146)(0.331) or 48.3 boards with one
defect. For these data the observed number of boards with one defect was 44, and the
expected value is 48.3.
The probability of finding exactly two defects is given by:
e1.53 (1.53)
e x
= 0.253.
= Px =1 =
x!
2
2
Px =0 =
In a sample of n = 146, we would expect to find (0.253)(146) or 36.9 boards with two
defects.
For these data the observed number of boards with one defect was 41, and the
expected value is 36.9.
The probability of finding exactly three defects is given by:
e1.53 (1.53)
e x
= 0.129.
= Px =1 =
x!
6
3
Px =0 =
In a sample of n = 146, we would expect to find (0.129)(146) or 18.8 boards with three
defects.
For these data the observed number of boards with one defect was 22, and the
expected value is 18.8.
The probability of finding exactly four defects is given by:
e1.53 (1.53)
e x
= 0.049.
= Px =1 =
x!
24
4
Px =0 =
In a sample of n = 146, we would expect to find (0.049)(146) or 7.2 boards with four defects.
For these data the observed number of boards with one defect was 4, and the expected
value is 7.2.
H1317_CH09.qxd
10/15/07
88
11:55 AM
Page 88
Px =0 =
In a sample of n = 146, we would expect to find (0.015)(146) or 2.2 boards with five
defects.
For these data the observed number of boards with one defect was 3, and the expected
value is 2.2.
The following is a summary of the observed frequency (from the data) and the
expected frequency (for a Poisson distribution where = 1.52). The chi-square value is
calculated the same as in the example for the contingency table.
Number
of defects
Frequency
observed, O
Frequency
expected, E
OE
(O E)2
(O E)2/E
0
1
2
3
4
5
32
44
41
22
4
3
31.6
48.3
36.9
18.8
7.2
2.2
0.4
4.3
4.1
3.2
3.2
0.8
0.16
18.49
16.81
10.24
10.24
0.64
0.01
0.38
0.46
0.54
1.42
0.29
The last row has fewer than five observations in the expected column; therefore, we
will combine this row with the previous row to increase the count to > 5 as required.
Number
of defects
Frequency
observed, O
Frequency
expected, E
OE
(O E)2
(O E)2/E
0
1
2
3
4 and 5
32
44
41
22
7
31.6
48.3
36.9
18.8
9.4
0.4
4.3
4.1
3.2
2.4
0.16
18.49
16.81
10.24
5.76
0.01
0.38
0.46
0.54
0.61
2 = 2.00
The critical chi-square value from the table is 23,0.0.10 = 6.25. Since 2.00 is less than the
critical value of 6.25, we cannot reject the null hypothesis that the data come from a Poisson distribution.
H1317_CH09.qxd
10/15/07
11:55 AM
Page 89
89
72772262
26509623
17503581
56704709
24052919
32549642
58136745
34286999
84544014
13103479
25178095
87717396
92733162
17194985
46878096
The objective is to confirm a level of significance of 5 percent that the distribution of integers is uniform.
Observed
frequency, O
Expected
frequency, E
0
1
2
3
4
5
6
7
8
9
15
19
19
11
20
13
13
20
14
16
16
16
16
16
16
16
16
16
16
16
Total:
160
Integer
OE
1
3
3
5
4
3
3
4
2
0
(O E )2
(O E )2/E
1
9
9
25
16
9
9
16
4
0
0.0625
0.5625
0.5625
1.5625
1.0000
0.5625
0.5625
1.0000
0.2500
0.0000
2 = 6.125
If the frequency were uniform, there would be the same frequency for all integer values, 16.
The critical chi-square value is determined with a significance of = 0.05 and the
degrees of freedom of:
df = number of classes (adjusted) 1 number of estimated parameters
df = 10 1 0 = 9.
Note that no parameters are estimated from the sample data.
Critical chi-square, 29,0.05 = 16.9. Since the calculated chi-square is less than the critical chi-square, we cannot reject the hypothesis that the distribution is uniform.
H1317_CH09.qxd
90
10/15/07
11:55 AM
Page 90
examined for the frequency of the use of the words thou, withal, wouldst, thine, and hadst.
The frequency of the words per 1000 words of text is determined and compared with the
suspect manuscript.
These data provide an opportunity to examine the true distribution to the suspect distribution. In this example the observed rate is that of the suspect, and the expected is that
of the true Shakespeare works.
The following computation reveals the truth about the suspect manuscript.
Word
Thou
Withal
Wouldst
Thine
Hadst
Observed rate
(suspect)
Expected rate
(true Shakespeare)
OE
(O E )2
40
18
58
13
44
58
24
56
19
31
18
6
2
6
13
324
36
4
36
169
(O E )2/E
5.58
1.50
0.07
1.89
5.45
Setting the level of significance at 5 percent and the degrees of freedom at 4, the critical
chi-square value is
2 4 , 0.05 = 9.49.
The calculated chi-square is greater than the critical chi-square; therefore, we conclude
that the suspect manuscript is not a work of William Shakespeare with a risk of 5 percent.
Bibliography
Walpole, R. E. 1972. Probability and Statistics for Engineers and Scientists. 5th edition.
Englewood Cliffs, NJ: Prentice Hall.
H1317_CH10.qxd
10/15/07
12:04 PM
Page 91
,
Exp
Exp
Exp
1
2
k
where: Obs (observed) = the actual observed frequency of occurrence for strata or
class k
Exp (expected) = the expected frequency for strata or class k (can be a count
or a percent).
91
H1317_CH10.qxd
10/15/07
92
12:04 PM
Page 92
The data for 15 lot analyses are shown in the following table:
Sieve size
Date
20
28
35
48
65
100
150
180
200
11/1
11/2
11/3
11/4
11/5
11/6
11/7
11/8
11/9
11/10
11/11
11/12
11/13
11/14
11/15
2
1
4
3
2
3
4
2
4
2
3
5
1
3
3
9
11
10
10
8
9
12
9
12
10
7
8
11
14
10
17
13
15
18
9
15
19
17
15
18
15
17
16
14
17
17
22
21
20
18
22
19
22
19
21
20
22
20
21
19
19
20
16
17
19
18
15
18
16
17
18
16
17
18
17
13
11
10
12
16
13
10
11
12
14
12
13
11
12
11
11
9
12
9
10
11
10
9
10
11
13
9
8
7
9
9
10
7
9
15
8
7
9
8
5
9
5
7
9
10
3
3
5
2
3
1
4
3
4
2
3
5
3
2
4
2.80
10.00
15.67
20.20
17.40
12.07
9.87
8.87
3.13
X:
Percent retained
40
30
20
10
0
20 28 35 48 65 100 150 180 200
Sieve size
Particle size distribution, average
21 = 1.2990
H1317_CH10.qxd
10/15/07
12:04 PM
Page 93
93
The remaining lot analysis results have the chi-square values calculated as in the previous example. A summary of all 15 chi-square values is as follows:
Date
Chi-square
11/1
11/2
11/3
11/4
11/5
11/6
11/7
11/8
11/9
11/10
11/11
11/12
11/13
11/14
11/15
1.299
2.582
2.990
0.859
9.378
2.060
3.017
0.821
1.456
3.150
1.966
5.469
2.124
3.089
0.765
The control limits for the chi-square control chart are based on the following relationships:
k = number of classes or strata in which data are reported. For this example, k = 9.
Center line = k 1 = 8.
25
UCL = 20.0
20
15
10
5
LCL = 0.0
0
1
10
15
H1317_CH10.qxd
94
10/15/07
12:04 PM
Page 94
The control limits are based on 3S, which means there is a 99.7 percent probability that
samples chosen from the population with the historical characterization will have a
99.7 percent probability of having a calculated chi-square less than 20.0 (for this example).
The first 15 samples are compared to a parent distribution defined by an expected percent retention on sieve #20 of 2.8 percent, sieve #28 of 10.0 percent, sieve #35 of 15.56
percent, and so on. The greater the deviation from the expected distribution, the greater the
chi-square value.
Additional data are taken, and signs of a process (distribution) change will be indicated if the chi-square calculated exceeds the calculated UCL of 20.0.
Calculate the chi-square values for the next five lots of material and see if there is evidence of a process (distribution) change.
Sieve size
Date
20
28
11/16
11/17
11/18
11/19
11/20
3
2
1
4
5
9
7
11
9
11
17
14
15
12
16
Avg. (1620)
Avg. (115)
3.0
2.8
9.4
10.0
14.8
15.67
Percent retained
40
35
48
65
100
150
180
200
11
14
10
15
14
12
9
11
12
11
21
23
19
17
22
13
16
15
17
14
6
9
11
9
6
8
6
7
5
1
13.6
20.2
11.0
17.4
20.4
12.07
15.0
9.87
30
20
10
0
20 28 35 48 65 100 150 180 200
Sieve size
Particle size distribution, average
8.2
8.87
5.4
3.13
H1317_CH10.qxd
10/15/07
12:04 PM
Page 95
95
40
Percent retained
11/1611/20
30
20
10
0
20 28 35 48 65 100 150 180 200
Sieve size
Particle size distribution, average
Date
11/16
11/17
11/18
11/19
11/20
Chi-square
22.20
22.75
20.73
12.77
16.76
The new additional chi-square values are plotted on the original chart. The results can
be seen in the following figure.
H1317_CH10.qxd
96
10/15/07
12:04 PM
Page 96
History
25
UCL = 20.0
20
15
10
5
LCL = 0.0
0
1
10
15
20
The chart has detected this change in the base or historical distribution.
An additional four lots are analyzed with the following results. Continue to calculate
the chi-square values and plot the data.
Sieve size
Date
20
28
35
11/21
11/22
11/23
11/24
2
3
1
3
11
9
12
12
12
11
10
13
Avg. (2124)
Avg. (1620)
Avg. (115)
2.3
3.0
2.8
11.0
9.4
10.0
11.5
14.8
15.67
48
65
100
150
180
200
17
16
19
17
15
18
14
15
15
13
17
14
13
11
14
15
10
14
10
9
3
5
3
2
17.3
13.6
20.2
15.5
11.0
17.4
14.8
20.4
12.07
13.3
15.0
9.87
10.8
8.2
8.87
3.3
5.4
3.13
H1317_CH10.qxd
10/15/07
12:04 PM
Page 97
97
A significant change in the distribution of the data in samples 1620 has caused the
chart to be out of control. The distribution of data in samples 2124 is not sufficient to
cause a change in the control chart pattern relative to the historical data contained in samples 115.
History
25
UCL = 20.0
20
15
10
5
LCL = 0.0
0
1
10
Percent retained
40
15
30
20
10
0
20 28 35 48 65 100 150 180 200
Sieve size
Particle size distribution, average
20
25
H1317_CH10.qxd
98
10/15/07
12:04 PM
Page 98
40
Percent retained
11/1611/20
30
20
10
0
20 28 35 48 65 100 150 180 200
Sieve size
Particle size distribution, average
Percent retained
40
30
20
10
0
20 28 35 48 65 100 150 180 200
Sieve size
Particle size distribution, average
Bibliography
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Shewhart, W. A. 1980. Economic Control of Quality of Manufactured Product. Milwaukee, WI: ASQC Quality Press.
H1317_CH11.qxd
10/15/07
12:25 PM
Page 99
Two of the most used descriptive statistics are the average X and the standard deviation S.
As with all statistics, these statistics are derived from samples taken from a larger population or universe and, therefore, are only estimates of the true parameter. The true average
is actually the mean and is abbreviated . All estimates are subject to error with respect to
the degree to which they represent the truth. This discussion focuses on the confidence of
the average as it represents the truth or parameter of the mean. Confidence is related to the
following:
1. Sample size n
2. Error of the estimate E
Confidence in the average can be stated in two specific ways:
1. Two-sided confidence intervals
2. Single-sided confidence limits
An example of the first case, the two-sided confidence interval, can be expressed in
statements such as, I dont know the true average, but I am 90 percent confident it is
between 12.4 and 15.5.
An example of a single-sided confidence limit would be, I do not know the true average, but I am 95 percent confident that it is less than 34.6 or I am 90 percent confident
that the true average is greater than 123.0.
Confidence can be thought of as a degree of certainty, and the opposite of confidence
is risk or degree of uncertainty. The sum total of confidence and risk is 100 percent:
Confidence + Risk = 100%.
The risk as associated with confidence is abbreviated .
Calculations for confidence are based on probability distributions. For sample sizes
less than 30, we use the t-distribution, and for sample sizes equal to or greater than 30, we
use the normal distribution.
The following example illustrates confidence calculations.
99
H1317_CH11.qxd
100
10/15/07
12:25 PM
Page 100
Confidence Interval, n 30
Example:
What is the 90 percent confidence interval for the average, where n = 15, S = 1.2, and the
X = sample average
t/2, n1 = a factor based on sample size and a confidence of 1 a.
S
is the standard error, and the quantity
n
t/2,n1 represents the number of standard error to add and subtract to the estimate average
to define the confidence interval. In selecting levels of confidence, we typically choose
values of 99.9, 99.0, 95.0, and 90.0.
For informational purposes, the quantity
S
E = t / 2,n 1
E = 1.761 0.31 E = 0.55
n
Step 4. Determine the confidence interval.
Confidence interval = Point estimate error
Confidence interval = 25.0 0.55
Confidence interval = 24.45 to 25.55
An appropriate statement for this case would be: I do not know the true average
(mean), but I am 90 percent confident that it is between 24.45 and 25.55.
H1317_CH11.qxd
10/15/07
12:25 PM
Page 101
101
Supplemental problem:
Determine the 95 percent confidence interval when n = 22, S = 4.5, and X = 45.00.
Confidence C
Risk
/2
Z/2
0.999
0.995
0.990
0.950
0.900
0.001
0.005
0.010
0.050
0.100
0.0005
0.0025
0.0050
0.0250
0.0500
3.29
2.80
2.58
1.96
1.65
Example:
A sample n = 150 yields an average X = 80.00 and standard deviation S = 4.50. What is the
99 percent confidence interval for this point estimate?
Step 1. Determine the Z/2 value.
The chosen level of confidence is 99.0 percent, or 0.990; the risk is 0.010; and /2 is
0.0050. Looking in the body of a standard normal distribution, we find that the nearest
Z value is 2.58.
Step 2. Calculate the standard error.
The standard error is given by
,
n
where: = true standard deviation
n = sample size.
Our best estimate of is S, and in this case S = 4.50. Therefore, our estimate of the
standard error is
4.50
= 0.37.
150
H1317_CH11.qxd
102
10/15/07
12:25 PM
Page 102
S
n
4.50
E = (2.58)
E = 0.95
150
Step 4. Determine the confidence interval.
Confidence interval = Point estimate error
Confidence interval = 80.00 0.95
Confidence interval = 79.05 to 80.95
Risk
0.999
0.995
0.990
0.950
0.900
0.001
0.005
0.010
0.050
0.100
3.09
2.58
2.33
1.65
1.28
H1317_CH11.qxd
10/15/07
12:25 PM
Page 103
103
3
= 0.35
75.0
3.0
S
E = Z
E = 1.65
E = 0.57
n
75.0
Step 4. Calculate the upper 95 percent confidence limit.
Upper 95% confidence limit = Point estimate + Error
Upper 95% confidence limit = 65.00 + 0.57
Upper 95% confidence limit = 65.57
An appropriate statement for this example would be: I do not know the true average
(mean), but I am 95 percent confident that it is not greater than 65.57.
Supplemental problem:
What is the 90 percent lower confidence limit given a standard deviation of 12.00, an average of 150, and a sample size of 200?
What is the lower 90 percent confidence limit where n = 25, S = 10.0, and X = 230.0?
Step 1. Determine the t value.
Risk = 0.10, and n = 25 (we look up n 1 or 24 degrees of freedom df ). The t value from
the t-table distribution is 1.318.
Step 2. Calculate the standard error.
S
10
=
= 2.00
n
25
Step 3. Calculate the error of the estimate E.
10
S
E = t ,n 1
E = 1.318
E = 2.64
n
25
H1317_CH11.qxd
104
10/15/07
12:25 PM
Page 104
Bibliography
Dovich, R. A. 1992. Quality Engineering Statistics. Milwaukee, WI: ASQC Quality Press.
Gryna, F. M., R.C. H. Chua, and J. A. De Feo. 2007. Jurans Quality Planning and Analysis
for Enterprise Quality. 5th edition. New York: McGraw-Hill.
Petruccelli, J. D., B. Nandram, and M. Chen. 1999. Applied Statistics for Engineers and
Scientists. Upper Saddle River, NJ: Prentice Hall.
H1317_CH12.qxd
10/15/07
12:30 PM
Page 105
105
H1317_CH12.qxd
106
10/15/07
12:30 PM
Page 106
The degree to which the sample estimate reflects the truth is defined by our level of
confidence. All statistics are subject to error.
The error E (or sometimes referred to as standard error SE) for the proportion P is calculated as follows:
E=
p (1 p ) ,
n
where: p = proportion
n = sample size.
In order to express a level of confidence in the estimate of the proportion, we add and
subtract a specific number of standard errors to the point estimate. The greater the level of
confidence, the more standard errors we add and subtract. The number of standard errors
is Z/2, where is the risk.
When the error is added and subtracted to the estimate, we are determining a two-sided
limit or confidence interval. If all of the error is applied to one side of the estimate, we
refer to a single-sided confidence limit and use Z.
Confidence and risk must total 100 percent. If we are 95 percent confident, we are
assuming a risk of 5 percent. If we are 90 percent confident, we are assuming a risk of
10 percent.
% Confidence + % Risk = 100%
% Risk = 100% % Confidence
Confidence and risk are expressed as a proportion rather than as a percentage. If we
are 90 percent confident, we are assuming a 10 percent risk that the true proportion will
not be inside the limits defined by the point estimate of Z/2E:
p Z/2
where:
p (1 p ) ,
n
p = proportion
n = sample size
Z/2 = factor dependent on risk assumed.
H1317_CH12.qxd
10/15/07
12:30 PM
Page 107
107
Confidence C
Risk
/2
Z /2
.999
.990
.950
.900
0.001
0.010
0.050
0.100
0.0005
0.0050
0.0250
0.0500
3.291
2.576
1.960
1.645
Example 1:
Calculate the 90 percent confidence interval for the previous example, where n = 150
and p = 0.18. The total population is 100,000; therefore, n < 0.05N.
From Table 1 for a level of confidence of 90 percent, the appropriate Z/2 is 1.645, and
using the relationship
p Z /2
p (1 p )
0.18 (1 0.18 )
0.18 1.645
0.18 0.05
150
n
15
= 0.176
85
0.176(1 0.176 )
,
85
H1317_CH12.qxd
108
10/15/07
12:30 PM
Page 108
6
= 0.040
150
p (1 p )
=
n
0.040 (1 0.040 )
150
E = 0.016
99 percent confidence interval = 0.040 2.576(0.014)
= 0.040 0.041
The upper 99 percent confidence limit is 0.040 + 0.041 = 0.081.
The lower 99 percent confidence limit is 0.040 0.041 = 0.001 0.000.
When a negative lower control limit results, default to zero.
Case 2: Two-sided confidence interval where the sample size is large relative to the population, n > 0.05N
When the sample size n is greater than 0.05N, a correction factor must be applied to
the standard error. This correction factor is called the finite population correction factor.
The finite population factor is given by
N n
n
= 1 .
N 1
N
The standard error must be multiplied by this correction factor anytime the sample size
n is greater than 0.05N.
H1317_CH12.qxd
10/15/07
12:30 PM
Page 109
109
p (1 p )
n .
1
n
N
p (1 p )
n
1 .
n
N
Example 1:
What is the 90 percent confidence interval for the proportional defective where the sample size n = 50, the population N = 350, and the number defective np = 12?
12
= 0.24
50
= 1.645
p=
Z /2
E=
p (1 p )
n
E=
1
n
N
0.24 (1 0.24 )
50
1
50
350
E = 0.056
90% confidence interval = 0.24 1.645(0.056)
= 0.24 0.092
The 90% upper confidence limit is 0.24 + 0.092 = 0.332.
The 90% lower confidence limit is 0.24 0.092 = 0.148.
Example 2:
A small engine-repair company serviced 6800 customers last quarter. A survey was sent to
the entire customer base, and of the 1500 who responded, 60 indicated that they felt the
length of time for engine repairs was excessively long. What is the 95 percent confidence
interval for the resulting percent of dissatisfied customers?
N = 6800
n = 1500
np = 60
p = 0.04
C = 95%
The sample size is 22 percent of the population, which exceeds the criteria of n 0.05 N;
therefore, we must apply the finite population correction factor to the error calculation.
H1317_CH12.qxd
110
10/15/07
12:30 PM
Page 110
Z / 2 = 1.960
E=
p (1 p )
n
1
E=
n
N
0.04 (1 0.04 )
1500
1
E = 0.004
1500
6800
Confidence C
0.999
0.990
0.950
0.900
Risk
0.001
0.010
0.050
0.100
3.090
2.326
1.645
1.282
H1317_CH12.qxd
10/15/07
12:30 PM
Page 111
111
Example 1:
A sample of 90 individuals are interviewed during the week as they are leaving the Le
Cash, Le Penty restaurant. During the same week, 3500 customers exited the restaurant.
Of those interviewed, 28 percent felt that the service was a little too slow. What is the
upper 90 percent confidence limit for the estimated percent of dissatisfied customers?
Since n < 0.05N, the finite population correction factor is not required.
p + Z
p (1 p )
n
0.28 + 1.282
0.28(1 0.28)
90
36
= 0.09
400
Z = 1.645
p (1 p )
0.090 (1 0.090 )
n
400
1
1 0.09 1.645
n
N
400
1500
N = 1800
n = 200
np = 12
C = 95%
N = 50,500
n = 160
np = 42
C = 99%
H1317_CH12.qxd
112
10/15/07
12:30 PM
Page 112
( Z / 2 ) 2 p (1 p ) ,
( E)2
p(1 p )
2
E p(1 p )
Z + N
/2
H1317_CH12.qxd
10/15/07
12:30 PM
Page 113
n=
113
( Z / 2 ) 2 p (1 p )
(1.960 ) 2 0.15 (1 0.15 )
n=
n = 544
( E)2
( 0.03 ) 2
To test whether the finite population factor is required, we determine the percentage
of the population the sample represents:
544
= 0.17
3200
Since 0.17 0.05, we should recalculate using the correction factor version:
n=
p(1 p)
2
E p(1 p)
Z + N
/2
, n=
0.15(1 0.15)
n = 465.
2
0.03 + 0.15(1 0.15)
1.960
3200
A sample size of 465 should be taken. We can be 95 percent confident that the resulting proportion will be within 0.03 of the true proportion.
Example 2:
What is the required sample size n for the two-sided confidence interval given
Confidence C = .99
Error E = 0.05
Should the finite population correction factor be applied for this case?
Bibliography
Dovich, R. A. 1992. Quality Engineering Statistics. Milwaukee, WI: ASQC Quality Press.
Gryna, F. M., R.C.H. Chua, and J. A. De Feo. 2007. Jurans Quality Planning and Analysis
for Enterprise Quality. 5th edition. New York: McGraw-Hill.
Petruccelli, J. D., B. Nandram, and M. Chen. 1999. Applied Statistics for Engineers and
Scientists. Upper Saddle River, NJ: Prentice Hall.
H1317_CH12.qxd
10/15/07
12:30 PM
Page 114
H1317_CH13.qxd
10/15/07
12:38 PM
Page 115
Confidence Interval
for the Standard Deviation
The standard deviation is one of several variation statistics (those descriptive statistics that
measure variation). Others include the range and maximum absolute deviation. The square
of the standard deviation is the variance. For each continuous and discrete distribution,
there is an associated variance and standard deviation. For the normal continuous distribution, we can determine the sample standard deviation from the relationship
S=
( X X ) 2 ,
n 1
X = average of individuals
n = sample size
= sum.
The standard deviation can be described as a standard way of measuring the deviation
(variation) of the individuals from their average.
As with all statistics, the standard deviation is subject to error. The error for the sample standard deviation is calculated using the chi-square distribution,
where: n = sample size
s = sample standard deviation
= population standard deviation (unknown).
Cumulative probabilities for the chi-square distribution can be determined as a function of the area associated with various tail areas designated as /2 for the upper portion
of the distribution and 1 /2 for the lower portion of the distribution. The probability of
risk is .
Derive the confidence interval for the standard deviation by
X 2 1 / 2 <
( n 1) s 2
< X 2 /2 .
2
H1317_CH13.qxd
116
10/15/07
12:38 PM
Page 116
21 / 2
2 / 2
Taking the reciprocals reverses the direction of the inequalities. If we reverse the order
of the terms, the result is now
(n 1)s2
(n 1)s2 .
2
<
<
2 / 2
21 / 2
This expression represents the 100(1 ) percent confidence interval for the variance.
Taking the square root of the three terms gives the confidence interval for the standard
deviation.
(n 1)s2
(n 1)s2
<
<
2 / 2
21 / 2
Example of confidence interval calculation:
As part of a study for variation reduction in an injection molding operation, the quality
engineer has taken a sample of 30 parts (all of which have the same target value) and determined the standard deviation s to be 0.0358. What is the 90 percent confidence interval for
the standard deviation?
n = 30
s = 0.0358
C = .90
= .10
(n 1)s2
(n 1)s2
<
<
2 a / 2
21 a / 2
(29)(0.0358)2
(29)(0.0358)2
<
<
20.95, 29
20.05, 29
Looking up the chi-square value in a chi-square table using n 1 degrees of freedom,
we find
2 0.05, 29 = 42.6 and 2 0.95, 29 = 17.7.
(29 )(0.0358 )2
(29 )(0.0358 )2
<<
42.6
17.7
0.0295 < < 0.0458
We are 90 percent confident that the true standard deviation is between 0.0295 and
0.0458.
H1317_CH13.qxd
10/15/07
12:38 PM
Page 117
117
Notice that the interval is nonsymmetrical with respect to the point estimate of sigma
s = 0.0358.
The difference between the lower 90 percent limit and the point estimate is 0.0043, where
the difference between the point estimate and the upper 90 percent confidence limit is 0.0058.
K 2 k 1 =
k (n 1)ln
S2
(n 1)
k
ln s
k +1
C = 1+
,
3k (n 1)
where: k = number of populations or samples
n = sample size (same for all k samples).
Example:
The objective is to establish a short-run SPC control chart based on the deviation from
nominal for several products. The specific characteristic will be the melt index.
The following table gives five measurements (sample size n = 5) for each of nine products (k = 9).
Sample number
Product
S2
ln S2
A
B
C
D
E
F
G
H
I
123
48
228
245
67
155
88
101
132
127
50
230
247
70
150
89
113
132
132
55
225
249
73
143
81
96
137
125
46
229
239
65
160
76
115
138
126
50
222
241
71
162
92
106
130
3.362
3.347
3.271
4.147
3.194
7.714
6.535
7.981
3.493
11.3
11.2
10.7
17.2
10.2
59.5
42.7
63.7
12.2
2.425
2.416
2.370
2.845
2.322
4.086
3.754
4.154
2.501
H1317_CH13.qxd
118
10/15/07
12:38 PM
Page 118
S2
= 26.522
k
k +1
9 +1
C = 1+
C = 1+
= 1.0926
3 k ( n 1)
27 ( 5 1)
ln S
where k = 9
= 26.873
n=5
k (n 1)ln
K 2 k 1 =
S2
k (n 1) ln s
C
[ 9(4)ln(26.522) (4)(26.873)] = 9.624
K 2 k 1 =
1.0926
K 28 = 9.624 vs 28,0.05 = 15.51
Since the calculated K 2 does not exceed 28,0.05, homogeneity (constant or equal variance) is a tenable hypothesis. A deviation from nominal control chart would be plausible.
Bibliography
Burr, E. W. 1974. Applied Statistical Methods. New York: Academic Press.
Dovich, R. A. 1992. Quality Engineering Statistics. Milwaukee, WI: ASQC Quality Press.
Petruccelli, J. D., B. Nandram, and M. Chen. 1999. Applied Statistics for Engineers and
Scientists. Upper Saddle River, NJ: Prentice Hall.
H1317_CH14.qxd
10/18/07
11:46 AM
Page 119
Examples:
Thirty-five CDs were inspected, and three defects were observed. The average number of
defects per unit is
u=
3
= 0.086 .
35
4
= 0.069
58
119
H1317_CH14.qxd
120
10/18/07
11:46 AM
Page 120
Normal variation for the average number of defects per unit u is defined as 3 standard deviations about the average defects per unit. The standard deviation for defects is
based on the Poisson distribution and is dependent on the average defects per unit and the
sample size.
The limits of this normal variation define the upper control limits (UCLs) and lower control limits (LCLs) for the u chart. Processes that are well behaved and statistically stable will
have 99.7 percent of their data within these limits. That is to say that the UCL and LCL for
the u chart are determined by
UCL = u + 3
u
u
.
and LCL = u 3
n
n
Note: The control limits are different for different sample sizes.
Example:
The average number of nonconformities per unit u for customers checking out of a hotel
has been determined to be 0.055. This value was determined from exit surveys during a
three-month period. If the daily sample size is 25, what are the UCL and LCL for this sample size?
UCL = u + 3
u
0.055
UCL = 0.055 + 3
= 0.196
n
25
LCL = u 3
u
0.055
LCL = 0.055 3
= 0.086 = 0.000
n
25
Since the LCL is negative, we default to a zero control limit. The following steps for
producing a u chart are the same as with any control chart for attributes:
Step 1. Collect historical data to provide a basis of process characterization.
Control charts are used to detect changes in processes. Before we can detect a change, we
need to know how the process was performing in the past. Typically, 25 samples k of size
n will be collected. The sample size, or subgroup sample size as it is frequently called,
should be sufficiently large as to contain at least two or three defects. For each sample, the
number of defects per unit will be determined.
Step 2. Determine a location statistic using all of the historical data.
For the u chart, this location statistic is the average defects per unit of inspection or sampling. This statistic is determined by dividing the sum of all defects found by the sum of
all samples taken to characterize the process.
H1317_CH14.qxd
10/18/07
11:46 AM
Page 121
u=
121
c1 + c 2 + c 3 + . . . c k
n1 + n 2 + n 3 + . . . n k
Case Study
As the manager of a hotel, you want to establish a u chart to monitor the room-preparation
activities. After a guest checks out, the room is cleaned and prepared for the next guest.
An audit check sheet lists several items that relate to this preparation. The following nonconformities list will be used:
1. Ground fault failed test
2. Too few towels, face cloths in bathroom
3. Covers on bed not smooth or even
4. Dust on furniture
5. Floors dirty
6. No toiletries in bathroom
7. No house cleaning I.D./welcome
card left
8. Bathroom not adequately cleaned
Each day for three weeks, 100 percent of the rooms have been audited for the eight
items listed. Develop a u chart based on the data provided. In this example, only 21 samples will be used to characterize the process.
H1317_CH14.qxd
122
10/18/07
11:46 AM
Page 122
Sample
number
Date
Sample size n
Number
of defects c
Defects/unit
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
11/1
11/2
11/3
11/4
11/5
11/6
11/7
11/8
11/9
11/10
11/11
11/12
11/13
11/14
11/15
11/16
11/17
11/18
11/19
11/20
11/21
75
103
53
68
88
115
90
48
65
100
85
75
55
90
60
95
70
105
87
92
70
6
8
2
6
5
11
5
2
5
3
6
4
2
5
2
7
6
4
4
6
3
0.080
0.075
0.038
0.088
0.057
0.096
0.056
0.042
0.077
0.030
0.071
0.053
0.036
0.056
0.033
0.074
0.086
0.038
0.046
0.065
0.043
c1 + c 2 + c 3 + . . . c k
6+8+2+6+. . . 3
=
= 0.060
n1 + n 2 + n 3 + . . . n k
75 + 103 + 53 + 68 + . . . 70
H1317_CH14.qxd
10/18/07
11:46 AM
Page 123
123
0.060
= 0.060 + 0.082 = 0.142
80.4
LCL = 0.060 3
0.060
= 0.060 0.082 = 0.022 = 0.000
80.4
Note: The LCLs will not be calculated. A default LCL value of zero will be used, since all
LCLs using a process average defects/unit of 0.060 will yield a negative control limit when
the sample size is less than 146. As a matter of fact, the relationship of average defects per
unit and sample size give the following relationship.
For a given average defects/unit u, the minimum sample size to give a zero LCL is
defined by the relationship
n=
9u
.
(u )2
For example, a process with an average number of defects/unit of 0.15 would require
a sample size of
n=
( 9 )( 0.15 )
9u
=
= 60 .
2
(u )
( 0.15 ) 2
The actual sample size should be a little larger than n = 60 such that the LCL will be
slightly larger than zero. It is suggested that the sample size be increased approximately
10 percent to, for example, n = 66.
UCL = 0.060 + 3
0.060
= 0.060 + 0.072 = 0.132
103
UCL = 0.060 + 3
0.060
= 0.060 + 0.101 = 0.161
53
UCL = 0.060 + 3
0.060
= 0.060 + 0.069 = 0.129
115
Sample #3, n = 53
The remaining samples may be calculated in a similar manner. A summary of the historical data and UCLs follows. (All LCLs defaulted to 0.000.)
H1317_CH14.qxd
124
10/18/07
11:46 AM
Page 124
Sample
number
Date
Sample size n
Number
of defects c
Defects/unit
UCL
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
11/1
11/2
11/3
11/4
11/5
11/6
11/7
11/8
11/9
11/10
11/11
11/12
11/13
11/14
11/15
11/16
11/17
11/18
11/19
11/20
11/21
75
103
53
68
88
115
90
48
65
100
85
75
55
90
60
95
70
105
87
92
70
6
8
2
6
5
11
5
2
5
3
6
4
2
5
2
7
6
4
4
6
3
0.080
0.075
0.038
0.088
0.057
0.096
0.056
0.042
0.077
0.030
0.071
0.053
0.036
0.056
0.033
0.074
0.086
0.038
0.046
0.065
0.043
0.142
0.132
0.161
0.142
0.142
0.129
0.142
0.167
0.142
0.142
0.142
0.142
0.159
0.142
0.142
0.142
0.142
0.132
0.142
0.142
0.142
Notice that as the sample size increases, the control limits become closer to the average line. This increases the probability of detecting a process change as indicated by a
point outside the control limit.
Step 4. Construct the chart and plot historical data.
The control chart with the initial 21 data points plotted and the control limits with the
process average defects per unit follows. Note the variable control limits for those data
points resulting when the sample size exceeds the interval of = 0.75n to 1.25n. The process
average defects line is drawn as a solid line and extends into the future beyond the last historical sample. This is done because it is this historical average to which we want to compare all future sample results.
History
.16
.14
.12
.10
.08
.06
.04
.02
0.0 1
10
15
20
25
H1317_CH14.qxd
10/18/07
11:46 AM
Page 125
125
Date
Sample size n
Number
of defects c
Defect/unit u
22
23
24
25
26
27
28
29
11/22
11/23
11/24
11/25
11/26
11/27
11/28
11/29
105
90
58
88
95
75
55
93
11
12
2
4
1
2
2
4
0.105
0.130
0.034
0.045
0.011
0.027
0.036
0.043
Upon completion of the control chart, it appears that the process average has decreased
relative to the historical value of .060 defects/unit. This conclusion is supported by the violation of rule #2: seven consecutive points below the historical average.
Completed u chart:
History
.16
.14
.12
.10
.08
.06
.04
.02
0.0 1
10
15
20
25
30
Optional problem:
The Stockton Plastics Corporation has been collecting data for approximately one week on
the frequency and nature of defects that occur in the injection molding department. Based
on this data, do you feel that the process is in control, and where would you focus your
attention to improve it?
H1317_CH14.qxd
10/18/07
126
11:46 AM
Page 126
Defect list:
Characteristic
Defect code
Cracked parts
Splay
Undersize
Chips
Flash
Color
Underweight
Plagimitized
Cost to repair, $
23
44
100
58
168
135
75
80
8.00
1.25
11.50
0.80
3.25
0.75
2.00
3.60
Lot size n
300
400
400
580
350
400
500
280
500
350
400
500
Total defects
5
12
8
15
14
22
10
14
25
7
24
18
23
44
100
58
168
135
75
80
1
3
0
2
0
4
1
2
5
3
0
1
0
2
2
4
0
5
4
2
4
3
0
4
2
3
3
3
2
5
3
2
3
0
4
4
1
0
5
0
9
7
2
6
5
0
9
6
0
1
0
4
1
0
0
1
3
0
5
3
1
0
0
0
2
1
0
1
2
0
0
2
0
0
0
1
0
0
0
0
2
0
2
1
0
0
0
1
0
0
0
0
1
1
4
0
What is the most frequently occurring defect, and where would you focus your attention to improve quality? Why?
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Wheeler, D. J., and D. S. Chambers. 1992. Understanding Statistical Process Control. 2nd
edition. Knoxville, TN: SPC Press.
H1317_CH15.qxd
10/18/07
11:48 AM
Page 127
Control charts for attributes are broadly classified into the following four types:
1.
2.
3.
4.
For both the c chart and the u chart there is a presumption that all defects are equal, or
at least the severity of the nature of the defect will not influence behavior of the control
chart. As a point of practical concern, we know that in some cases, defects are not always
equal in their impact on the quality of the product or performance of the service to which
they are associated. For example, in the audit of a sample of hotel rooms, if one finds that
no hot water is available versus having two towels where three are required, the resulting
level of customer satisfaction is significantly different. Another example would be the
results of the final audit of the readiness of a New York to Seattle commercial flight where
three defects were found: (1) inadequate pillows relative to the specification for the flight,
(2) a shortage of 25 meals for the flight, and (3) in-flight movie equipment not working.
With some imagination, defects in most operations and services can be divided into two or
more classification levels with respect to the severity of the defects. By categorizing
defects into classifications and giving a relative weighting scale to specific defects, we can
develop a control chart that will not be overly sensitive to minor defects or under sensitive
to critical defects. Such a treatment of defects will lead to the use of what is commonly
referred to as a demerit/unit control chart. The following example illustrates the construction of the demerit/unit, or Du, control chart.
Essential operations for the Du control chart include:
1.
2.
3.
4.
5.
6.
H1317_CH15.qxd
128
10/18/07
11:48 AM
Page 128
Major
No towels
No telephone directory
No shampoo
Drapery drawstring
broken
Internet connection
not working
Minor
Missing mint on bed in evening
Dust on furniture
Drawer moderately difficult
to open
No motel stationery
Motel phone number missing
from room phone
H1317_CH15.qxd
10/18/07
11:48 AM
Page 129
129
Sample
number
Size n
Date
Criticals
Majors
Minors
Total
demerits
Demerits/
unit
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
40
40
40
40
40
40
40
40
40
40
40
40
40
40
40
7/7/94
7/8/94
7/9/94
7/10/94
7/11/94
7/12/94
7/13/94
7/14/94
7/15/94
7/16/94
7/17/94
7/18/94
7/19/94
7/20/94
7/21/94
1
0
1
1
2
0
0
0
1
0
0
1
1
0
0
2
3
3
2
0
4
3
1
2
1
3
4
0
3
1
6
2
9
4
3
6
5
6
5
4
4
6
8
7
5
180
85
220
170
215
130
100
55
175
45
95
230
140
110
50
4.50
2.13
5.50
4.25
5.75
3.25
2.50
1.38
4.38
1.13
2.38
5.75
3.50
2.75
1.25
Totals:
600
32
80
2000
3.33
the binomial is given by npq, where n = sample size, p = fractional defective, and q = 1 p.
Rearranging this equation based on a proportional defective rather than a number
defective, the standard deviation is equal to p (1 p ) .
n
The average fractional defective term p varies from one type of defect to another. Of
the 600 samples inspected (rooms audited), there are 8 defects that are critical, 32 that are
major, and 80 that are minor. In order to combine the effects of these defects and allow a
relative weighting in accordance with the demerit assignment, we must weigh the calculation of the standard deviation. This is accomplished by calculating a weighted average
standard deviation, Du:
Du = Wc
p c (1 p c )
+ Wma
n
p ma (1 p ma )
+ Wmi
n
p mi (1 p mi ) ,
n
H1317_CH15.qxd
130
10/18/07
11:48 AM
Page 130
Rearranging, we have
Du =
Wc 2 p c (1 p c ) + Wma 2 p ma (1 p ma ) + Wmi 2 p mi (1 p mi ) .
n
pc = the fractional contribution of the critical defects and is equal to the total number
of critical defects found, divided by the total number of units inspected:
pc =
8
= 0.013 .
600
In a similar manner, we calculate the fractional contribution of the major and minor
defects:
p ma =
32
= 0.053
600
and
p mi =
80
= 0.133 .
600
Total demerits
2000
=
= 3.33.
Total units inspected 600
H1317_CH15.qxd
10/18/07
11:48 AM
Page 131
131
Record all information in the data area, including date; sample size; total demerits;
demerits/unit; and number of critical, major, and minor defects found in the sample.
A wavy vertical line is drawn between sample #15 and sample #16 to separate the
future values from the historical data that were used to develop the control chart average
and control limits. Label the left portion of the control chart as History.
History
8
7
Du
6
5
4
3
2
1
0
1
10
15
20
25
Step 6. Continue to monitor the process, looking for changes in the future.
The process appears to be statistically stable in that approximately half of the points are
distributed above the average and half below, and runs of length seven are not present.
Data collected in the future can be used to detect a shift in the process average. The following are evidence of a process change:
1. There is a run of seven points in a row on one side of the average
2. A single point is present outside the control limits
3. Seven points are steadily increasing or decreasing
These three indicators or rules are statistically rare with respect to their probability of
occurrence for a stable process. They are sufficiently rare enough that when these patterns
do occur, we assume that a process change has taken place. Using these rules and the following data, determine if there has been a process change. Do the data indicate that the
performance has gotten better or worse when compared to the historical average of
3.33 demerits per unit?
Note that the sample size has been increased to n = 120. A recalculation of the control
limits is required anytime the sample size changes. The historical values for Du, pc, pma,
H1317_CH15.qxd
10/18/07
132
11:48 AM
Page 132
and pmi are used in the recalculation. Only the sample size n is changed in the computation. The new value for three standard deviations is calculated by
3 for Du = 3
162.12
= 3.49.
120
The new upper and lower control limits using n = 120 are:
UCL = 3.33 + 3.49 = 6.82
LCL = 3.33 3.49 = 0.16 = 0.00.
Future data from the process are as follows:
Sample
number
Size n
Date
Criticals
Majors
Minors
120
120
120
120
120
120
120
120
120
120
120
7/22/94
7/23/94
7/24/94
7/25/94
7/26/94
7/27/94
7/28/94
7/29/94
7/30/94
8/1/94
8/2/94
2
1
3
3
2
5
7
5
3
6
4
4
12
10
2
4
3
10
9
2
13
0
12
16
21
9
2
13
13
3
1
3
14
16
17
18
19
20
21
22
23
24
25
26
Total
demerits
Demerits/
unit
The total demerits and demerits/unit columns have been left blank intentionally for the
reader to complete. The finished control chart can be seen in the following figure. It
appears that the process average demerits/unit has shifted positively.
UCL = 9.36
9
History
UCL = 6.82
Du
6
5
4
3
2
1
LCL = 0
0
1
10
15
20
25
H1317_CH15.qxd
10/18/07
11:48 AM
Page 133
133
Optional problem:
A complex electronic device goes through a comprehensive final audit. Defects are categorized into four classifications and are assigned demerit weighting factors accordingly:
1.
2.
3.
4.
Critical C = 100
Severe S = 50
Major M = 10
Incidental I = 1
Sample size n
100
100
100
100
100
400
400
400
250
250
250
1
0
0
1
0
3
2
1
5
2
5
0
1
2
3
1
2
5
8
2
4
2
4
5
1
2
5
6
14
10
11
7
4
9
3
5
8
6
24
30
18
12
16
11
D = demerits/unit
Wc = weighting factor for criticals
Wma = weighting factor for majors
Wmi = weighting factor for minors
and
Uc = counts for criticals per unit
Uma = counts for majors per unit
Umi = counts for minors per unit.
If the weighting factors for the critical, major, and minors are 12, 9, and 4, respectively, the demerits per unit would be calculated
D = 12Uc + 9Uma + 4Umi.
H1317_CH15.qxd
134
10/18/07
11:48 AM
Page 134
(12 ) 2 U c + ( 9 ) 2 U mi + ( 4 ) 2 U mi .
n
Example:
The sample size is n = 40, and the average defects/unit for the critical, major, and minor defects
are 0.07, 0.38, and 1.59, respectively. What is the center line, UCL, and LCL for this chart?
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Hayes, G. E., and H. G. Romig. 1982. Modern Quality Control. 3rd edition. Encino, CA:
Glencoe.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
H1317_CH16.qxd
10/18/07
11:49 AM
Page 135
Descriptive Statistics
Processes can be thought of as defined activities that lead to a change. Processes are dynamic
and always add value to services or manufactured devices. As quality practitioners, we are
all interested in processes. Processes can be measured in a numerical sense, and all observations regarding processes can be ultimately converted into numerical data. Examples of
measurements include service-related and manufacturing-related measurements.
Service-related
Length of time required to answer customer questions
Proportion of customers who are not satisfied with a service
Number of errors found on expense reports
Units of production per hour
Error on shipping papers
Nonconformities on a service audit
Manufacturing-related
Total defects on 45 printed circuit boards
Surface finish
Diameter of a part
pH of a solution
In most cases, we have an overabundance of both the opportunity and the available
data for measurements of a characteristic.
All of the opportunities for a process to be measured can be considered as a universe
from which we can develop a characterization. This entire opportunity can be considered
as the universe or population of a process. In magnitude, this opportunity is unlimited.
For our discussion, we will equate process = population = universe. The abbreviation
for the population is N. N consists of all the observations that can be made on a process.
In most cases, N is exceedingly large.
Process
Population N
Universe
135
H1317_CH16.qxd
136
10/18/07
11:49 AM
Page 136
Since this population is so large, we cannot from a practical point measure all of it.
The process might be one of answering calls at a phone help line where hundreds of calls
are received each hour. We are, nevertheless, interested in the length of time required to
answer each call. The size of the opportunity (population) does not diminish our interest
in the process. If we cannot measure all of the calls, we can always take a sample of the
population. Samples are subsets of the larger population or universe.
We can act upon data from samples to calculate such information as averages, proportions, and standard deviations. The following are examples of this population/sample
relationship as used in quality:
1. You are interested in determining the percentage of customers who use your service
and would consider using it again. You have sampled 500 and found that six would
not use your service again. This proportion is 6/500, or 0.012.
2. A shipment of 1200 bolts has been received. The average weight of a random sample of 25 is determined to be 12.68 grams.
3. Approximately 45 patients a day are admitted to a local hospital. A daily sample of
20 are chosen each day to determine the standard deviation for the time required to
complete the admission process.
Process
Population N
Universe
Time = $$$$
Sample n
Information calculated from samples is called statistics. Examples of statistics frequently used in quality are:
H1317_CH16.qxd
10/18/07
11:49 AM
Page 137
Descriptive Statistics
137
Average, X
Standard deviation, S
Traditional statistics such as the average are abbreviated using English letters or symbols
made up of English letters. The average proportion is written with a letter P and a bar over
it. Any statistic with a bar over it refers to the average of that statistic. For example, R represents the average range. The abbreviations may or may not be capitalized.
Statistics by their very nature are subject to error. Statistics are estimates of the true
description of the process. If we could and did use all of the data in the population or universe to calculate a defining characteristic, then there would be no error. Such numbers
derived from all of the data are not called statistics but rather parameters. The true proportion is still called the proportion; however, it is abbreviated with a lowercase Greek letter for the English letter Ppi or . The true standard deviation is abbreviated using the
Greek lowercase letter for the English letter Ssigma or . The case for the average is
unique in that the name of the true average is the mean, and the abbreviation for the mean
is the lowercase Greek letter for the English letter Mmu or .
This parameter is estimated by the corresponding statistic, and it is the parameter that
actually describes the process without error.
Process
Population N
Universe
Time = $$$$
Average X
Proportion P
Standard deviation S
Calculate
Sample n
H1317_CH16.qxd
138
10/18/07
11:49 AM
Page 138
Parameter
Describes
Process
Population N
Universe
Time = $$$$
Estimate
Statistic
Average X
Proportion P
Standard deviation S
Calculate
Sample n
H1317_CH16.qxd
10/18/07
11:49 AM
Page 139
Descriptive Statistics
139
Descriptive statistics
Location:
Variation:
Average X
Median X
Mode M
Average proportion P
Range R
Standard deviation S
numbers used in conjunction with each other give a better overall description of data sets
and processes.
The nature of the actual data collected can be defined as variables or attributes. A variable is any data collected initially as a direct number and can take on any value within the
limits in which it can exist. Examples of variables data would be temperature, pressure,
time, weight, and volume.
Attribute data are data that evolve from a direct observation where the response is yes
or no. The actual attribute data can and must, however, be translated into a number so that
mathematical operations can be made. An example of attribute data would be the characteristic of meeting a specification requirement. The specification is either met (yes response)
or it is not (no response).
Variables data can be converted to attribute data by comparing them to a specification
requirement.
Example:
The weight of a product is 12.36 poundsa variable. The specification for the weight is
12.00 0.50 pounds. Observation: the product meets the specification requirement, so the
response is yesan attribute.
H1317_CH16.qxd
140
10/18/07
11:49 AM
Page 140
Bibliography
Gryna, F. M., R.C.H. Chua, and J. A. De Feo. 2007. Jurans Quality Planning and Analysis
for Enterprise Quality. 5th edition. New York: McGraw-Hill.
Hayes, E., and H. G. Romig. 1982. Modern Quality Control. 3rd edition. Encino, CA:
Glencoe.
Petruccelli, J. D., B. Nandram, and M. Chen. 1999. Applied Statistics for Engineers and
Scientists. Upper Saddle River, NJ: Prentice Hall.
Sternstein, M. 1996. Statistics. Hauppauge, NY: Barrons Educational Series.
H1317_CH17.qxd
10/18/07
11:51 AM
Page 141
Designed Experiments
Designed experiments provide a statistical tool to allow for the efficient examination of
several input factors and to determine their effect on one or more response variables.
Input factors are those characteristics or operating conditions that we may or may not
have direct control over but that can and do affect a process. Response variables or factors
are the observations or measurements that result from a process.
Examples of this input factor/response variable relationship follow.
Process: Wood-Gluing Operation
Input factors
Amount of glue
Type of glue
Drying temperature
Drying time
Moisture content of wood
Type of wood
Relative humidity of environment
Response variables
Tensile strength, pounds/inch2
For all of these input factors, there is an optimum level of setting that will maximize
the bond strength. Some of these factors may be more important than others.
Designed experiments (frequently referred to as design of experiments, or DOE) can
assist in determining which factors play a role in affecting the level of response.
There are essentially five steps in a DOE:
1. Brainstorming to identify potential input factors (or factors) and output responses (or
responses) and establishing levels for the factors and the measure for the response(s)
2. Designing the experimental design or matrix
3. Performing the experiment
4. Analyzing the results
5. Performing a validation run to test the results
Example A:
A manufacturer of plastic/paper laminate wants to investigate the lamination process to see
if improvements can be made.
141
H1317_CH17.qxd
142
10/18/07
11:51 AM
Page 142
Step 1. Brainstorm.
During this initial stage, individuals gather to discuss and define input factors and output
responses. The results of this meeting yield the following:
Input factors
Top roll tension setting, lb.
Rewind tension setting, lb.
Bottom roll tension setting, lb.
Take-up speed, ft./min.
Type of paper
Thickness of plastic film, mils.
Type of plastic film
Response variables
Amount of curl
Number of wrinkles per 100 ft.
Peel strength, lb./in.
For this initial experiment, three factors and one response variable are selected. The
input factors are identified as A, B, and C. The response variable is the amount of curl, and
the three input factors are the following:
Input factor
A
B
C
Description
Top roll tension
Bottom roll tension
Rewind tension
Response variable
Curl, inches
Description
A 36-inch strip of paper/plastic film is hung against a
vertical plane. The distance from the end of the laminate
to the vertical plane is measured in millimeters.
Laminate
Curl
H1317_CH17.qxd
10/18/07
11:51 AM
Page 143
Designed Experiments
143
The objective of this experiment is to better understand the effects of the three chosen
factors on the amount of curl in the laminate. The current method calls for these factors to
be set as follows:
Factor
A
B
C
Setting
22
22
9
For this experiment, we will set each of the factors to a + level and a level. The
notation for this three-factor, two-level experiment is
23,
where: 2 = number of levels
3 = number of factors.
The total number of experimental conditions is given by one 23 = 8.
The general case for any two-level experiment, where all possible experimental conditions are run, is given by
2n,
where n = the number of factors.
Each of the factors will be set at a + and level around the traditional value at which
they are run.
Note: Some text on DOE uses 1 for one level and 2 for the second level.
Factor
A
B
C
Setting
16
16
6
+ Setting
28
28
12
The new conditions are set at such a level as to give a change in the response variable
if the factor is a contributor to the response variable. The change in the factor setting
should be large enough to change the response variable but not enough to critically affect
the process.
Step 2. Design the experiment.
This experiment is a full 23 and will have eight experimental setups. Start by having eight
runs and three columns for the factors A, B, and C.
H1317_CH17.qxd
10/18/07
144
11:51 AM
Page 144
Factor
Run
1
2
3
4
5
6
7
8
Start by placing a for the first run of factor A and then alternating + and for the
first column.
Factor
Run
1
2
3
4
5
6
7
8
The second column is treated in a similar manner, except there are pairs of s and +s.
The third column is completed with groups of four s and four +s. This completed table
represents the design or matrix for the two-level, three-factor experiment, or the 23 design.
Factor
Run
1
2
3
4
5
6
7
8
+
+
+
+
+
+
+
+
H1317_CH17.qxd
10/18/07
11:51 AM
Page 145
Designed Experiments
145
The order of one through eight listed in numerical sequence is called the standard
order. The actual sequence for the eight experiments should be a random order. The and
+ signs determine the setting for the particular run.
For example, run #3 would have:
Factor A (Top roll tension) set at 16
Factor B (Bottom roll tension) set at 28
Factor C (Rewind tension) set at 6
Step 3. Perform the experiment.
Each of the eight experimental runs is made in a random order. The results of each run are
recorded, and a single measurement observation will be made for each run.
Step 4. Analyze the data.
Main Effects
For each of the main factors A, B, and C, an effect will be determined. The effect will be
the average effect obtained when the factor under consideration is changed from a setting to a + setting. The signs of the column to which a factor has been assigned will be
used to determine the effect.
Step 5. Validation.
The original objective was to minimize the curl response. From the graphical analysis of
the effects, we have concluded that factors A and C are major factors with respect to their
effect on curl. Since we are minimizing the response of curl, we must set the factors A and
C to the + and settings, respectively. The top roll tension should be set to 28, the rewind
tension set to 6, and the remaining factor B to the most economical setting, as it has no significant effect on tension.
The process should now be run for a more extended period of time to collect data on
the curl response. These data from the improved process can now be compared with the
historical data using traditional methods of hypothesis testing. Essentially we want to
know if there has been a measured, statistically significant reduction in the amount of curl.
Factor
Run
Response, curl
1
2
3
4
5
6
7
8
+
+
+
+
+
+
+
+
87
76
90
83
101
92
100
92
H1317_CH17.qxd
146
10/18/07
11:51 AM
Page 146
Effects are determined by taking the difference between the average response when
the factor is set + and the average response when the factor is set .
Main Effect A:
76 + 83 + 92 + 92 87 + 90 + 101 + 100 = 85.75 94.50 = 8.75
4
4
The average effect in going from a + setting for factor A to a setting for factor A is
8.75 units of curl.
Main Effect B:
90 + 83 + 100 + 92 87 + 76 + 101 + 92 = 91.25 89.00 = +2.25
4
4
Main Effect C:
101 + 92 + 100 + 92 87 + 76 + 90 + 83 = 96.25 84.00 = +12.25
4
4
The main effects can be visualized by drawing a cube plot of the responses.
100
92
101
+
C
87
92
90
83
+
76
It can be seen that in order to minimize the curl response, the following settings for
the factors should be made:
Factor
Factor A (Top roll tension)
Factor B (Bottom roll tension)
Factor C (Rewind tension)
Setting
A + = 28
B = 16
C=6
In addition to the main effects, there can be interaction effects. Interactions result
when the magnitude of one main effect depends on or is related to another effect. For
example, the rate at which a chemical reaction occurs can be influenced by a catalyst. The
selection of the catalyst can be a determining factor. Catalyst A might perform well but
only at a higher temperature, whereas catalyst B would perform as well as catalyst A but
H1317_CH17.qxd
10/18/07
11:51 AM
Page 147
Designed Experiments
147
only at a lower temperature. The catalyst and temperature factors interact with each other;
in other words, there is a catalyst-temperature interaction.
In order to evaluate interactions of factors A, B, and C, the interactions will be determined. There are three interactions that involve two factors:
AB = the interaction of top roll tension and bottom roll tension
BC = the interaction of bottom roll tension and rewind tension
AC = the interaction of top roll tension and rewind tension
These interactions may also be written as A B, B C, and A C. In addition, there
is a single three-factor interaction expressed as ABC.
The computation of these interactions is performed using the signs of the columns as
was the case with the main effects. The signs for the interaction columns are determined
by multiplying the signs of the main effects used in the interaction.
Factor
Run
AB
AC
BC
ABC
1
2
3
4
5
6
7
8
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
Response, curl
AB effect:
87 + 83 + 101 + 92 76 + 90 + 92 + 100 = 90.75 89.50 = +1.25
4
4
AC effect:
87 + 90 + 92 + 92 76 + 83 + 101 + 100 = 90.25 90.00 = +0.25
4
4
BC effect:
87 + 76 + 100 + 92 90 + 83 + 101 + 92 = 88.75 91.50 = 2.75
4
4
ABC effect:
76 + 90 + 101 + 92 87 + 83 + 92 + 100 = 89.75 90.50 = 0.75
4
4
87
76
90
83
101
92
100
92
H1317_CH17.qxd
148
10/18/07
11:51 AM
Page 148
Are these effects significant, or are the differences due to random, experimental error?
There are two ways to determine if these effects are statistically significant. One way is to
repeat the experiment and determine the actual experimental error, and the other is utilized
when only one observation is made for each run (as in this case). When only one observation per run is available, the use of a normal probability plot can be helpful.
Effect value
8.75
2.75
0.75
0.25
1.25
2.25
12.25
Effect
A
BC
ABC
AC
AB
B
C
Note that there is one less effect than there are experimental runs.
Step 2. Calculate the percent median rank values.
%MR =
i .3
100
n + .4
H1317_CH17.qxd
10/18/07
11:51 AM
Page 149
Designed Experiments
Order i
Effect value
Effect
1
2
3
4
5
6
7
8.75
2.75
0.75
0.25
1.25
2.25
12.25
A
BC
ABC
AC
AB
B
C
9.5
23.0
36.5
50.0
63.5
77.0
90.5
149
Step 3. Plot the effects as a function of the percent median rank on the normal probability
paper.
Draw a straight line through the points, concentrating the emphasis around the points nearest the zero effect. Effects falling off this straight line are judged to be significant. To maximize a response variable, set the main effect to the same sign as the effect. To minimize
the response variable, set the factor to the reverse of the effect. In this example, factors A
and C are statistically significant.
99.99
99.9
99.8
99
98
95
xC
70
60
50
40
30
AB
AC
x
ABC
x
BC
20
10
5
2
1
.5
.2
.1
.05
.01
10 8 6 4 2
2 4
Value
10 12
Median rank
80
xB
x
90
H1317_CH17.qxd
150
10/18/07
11:51 AM
Page 150
80
70
60
A+
Since factor A was chosen for the horizontal axis, factor B will be plotted as a pair of
linear linesone for B and one for B+.
The B line will be plotted first. There are two experimental runs where A is set to A
and B is set to B. These two runs are run #1 and run #5. The average response for these
two is
87 + 101
= 94.0.
2
When factor A is set to A and factor B is set to B, the average response is 94.0.
The other end of the B line is determined by calculating the average response when
A is set at A+ and B is set at B. A+ and B are found in two experimental runsrun #2
and run #6. The average response for these two is
76 + 92
= 84.0.
2
These two values define the B line on the graph.
H1317_CH17.qxd
10/18/07
11:51 AM
Page 151
Designed Experiments
151
94.0
90
84.0
80
70
60
A+
The remaining line is the B+ line. There are two runs where B is set + and A is set .
These are run #3 and run #7. The average of these two responses is
90 + 100
= 95.0.
2
The other end of the B+ line is at the A+ position. There are two runs where B is set +
and A is set +. These are run #4 and run #8. The average of these two responses is
83.0 + 92.0
= 87.5.
2
100
95.0
94.0
B+ Line
90
87.5
B Line
84.0
80
70
60
AB interaction
A+
H1317_CH17.qxd
152
10/18/07
11:51 AM
Page 152
C+
87.0
86.5
C
80
81.5
70
60
B+
87.0 + 76.0
= 81.5
2
C B + point =
90.0 + 83.0
= 86.5
2
C + B point =
93.0 + 81.0
= 87.0
2
C + B + point =
95.0 + 90.0
= 92.5
2
The C+ line:
In the original experiment, only one observation was made for each experiment run.
The only technique to determine if the effects were statistically significant was to perform
a normal probability plot and make a subjective judgment if any of the points fell off the
straight line (with emphasis on the zero point). This method is subject to errors of judgment, but it is the best technique available when only one observation is available.
In order to determine experimental error, the entire experiment must be replicated at
least one more time in order to calculate a measure of variationthe standard deviation.
The entire set of eight runs is replicated. The new data from the first and second replications yield an average and standard deviation for each of the eight runs:
H1317_CH17.qxd
10/18/07
11:51 AM
Page 153
Designed Experiments
153
Factor
Run
AB
AC
1
2
3
4
5
6
7
8
+
+
+
+
+
+
+
+
+
+
BC
+
+
+
+
ABC
1st
2nd
Avg.
Std. dev.
+
+
87
76
90
83
101
92
100
92
88
78
92
80
96
91
104
91
87.5
77.0
91.0
81.5
98.5
91.5
102.0
91.5
0.707
1.414
1.414
2.121
3.535
0.707
2.828
0.707
The new calculated effects, in decreasing order, based on the average of two observations/
run are shown in the following table:
Factor
Effect
A
ABC
BC
AB
AC
B
C
9.38
1.13
1.13
0.63
0.63
2.88
11.63
While a normal probability plot would yield the same conclusion that factor A and factor B are probably statistically significant, again, the conclusion is somewhat subjective.
By using the t-distribution and determining the confidence interval for each effect, a
better judgment can be made.
The following method can test if the effect is within the normal variation of the experimental error or if the effect is statistically removed from the limits of normal variation.
That is, the effect is real and not just a random value.
A confidence interval will be determined for each effect. If zero is found to be within
the confidence interval limits, the conclusion is that the observed effect is simply random
variation within the expectation for the calculated experimental error. If, however, the confidence interval does not contain zero, the conclusion is that the effect is real and is statistically significant.
Calculate the pooled standard deviation for all the observations.
Sp =
V1 S12 + V2 S 22 + . . . Vn S n2
V1 + V2 + . . . Vn
H1317_CH17.qxd
154
10/18/07
11:51 AM
Page 154
Sp =
Sp = 1.952
Note: If the number of replicates is the same for all runs, then the pooled standard deviation is simply the square root of the average variances.
Each of the calculated effects is actually a point estimate. The confidence interval for
each of these estimated effects is determined by
Effect Error.
The error is calculated as
t / 2 ,( r 1 ) 2 k f Sp
2
,
pr
2
E = t / 2,(r 1)2k f Sp
pr
2
E = 2.306 1.952
(4)(2)
E = 2.25.
Any effect that is contained within the limits of 0 2.25 is considered not to be statistically significant.
H1317_CH17.qxd
10/18/07
11:51 AM
Page 155
Designed Experiments
155
Note: Any effect whose absolute value is greater than the error is statistically significant.
The main effect B is deemed to be statistically significant at the 95 percent level of
confidence. This was not evident using the normal probability plot method.
Factor
Run
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
H1317_CH17.qxd
10/18/07
156
11:51 AM
Page 156
Three-level interactions are very rare. If the assumption is made that any three-level
interaction is actually due to experimental error, then by letting the ABC interaction of this
24 experiment equal the D main effect, the full 16 runs for a full 24 can be cut in half. In
other words, let D = ABC. D = ABC is called a generator, and the identity is I = ABCD.
By mixing the effect of main effect D with the three-level interaction ABC, we have
confounded the two effects. D and ABC are called aliases. Since D and ABC are aliases,
we cannot differentiate between D and ABC.
Factor
Run
ABC
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
Let D = ABC.
Factor
Run
D
ABC
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
H1317_CH17.qxd
10/18/07
11:51 AM
Page 157
Designed Experiments
157
Notice that run #1 through run #8 are now duplicated by run #9 through run #16. Run
#9 through run #16 can be eliminated. Four factors can be run using eight experimental
runs. This experimental design is called a 241 design.
Fractional factorial designs can be designated by the general form of
2kf,
where: k = number of factors 2kf
f = degree of fractionation
f = 1 or 12 fractional design
f = 2 or 14 fractional design
f = 3 or 18 fractional design.
The number of experimental runs is determined by the value of 2kf. In a 273, there
are 16 runs24 = 16.
The resolution of a design determines the degree to which confounding is present.
Resolutions are expressed as roman numerals III, IV, or V and are named by the number
of letters in the shortest word of the defining relationship. In the example of a 241, the
defining relationship was D = ABC. The identity element is I = ABCD. There are four letters in the word; therefore, the resolution is IV.
The proper designation for this design is
2 IV 4 1.
Possible experimental design resolutions are as follows:
RIII: Main effects and two-factor interactions are confounded, so watch out! RIII are
good for screening designs.
RIV: Main effects are clear from any two-factor interactions, and two-factor
interactions are confounded with other two-factor interactions.
RV: Main effects are clear, and two-factor interactions are clear.
Example:
In a 252, the defining relationships are D = AB and E = AC.
This is a 1/4 fractional design. The number of factors is five, the number of runs is eight,
and the resolution is III (I = ABD and I = ACE); three letters make up the identity element.
This is a 2III52 experiment.
Fractional designs are created by confounding main effects with interactions. Table 1
lists other designs available using such confounding relationships, or generators.
H1317_CH17.qxd
158
10/18/07
11:51 AM
Page 158
Number of factors k
Designation
Number of runs
2III31
A = BC
2IV41
D = ABC
51
16
2III52
2VI61
32
F = ABCDE
2IV62
16
E = ABC
F = BCD
2VII71
64
G = ABCDEF
2IV72
32
F = ABCD
G = ABDE
2IV73
16
E = ABC
F = BCD
G = ACD
D = AB
E = AC
F = BC
G = ABC
2V
2III74
Design generator
E = ABCD
D = AB
E = AC
Case Study
The AstroSol company manufactures solar panels used to generate electricity. The manufacturing process consists of deposition of alternating layers of tin (Sn) and cadmium telluride (CdTe) and then baking these deposited layers in an oven. The power output is then
measured and reported in units of watts/ft2.
The R&D department wants to maximize the output of the cells. After considerable
review and discussion, the design team decides to look at five factors in a 1/2 fractional
factorial design, 2V51, as this design will isolate a main and two factorial interactions.
The five factors to be considered are:
Thickness of Sn in microns
Thickness of CdTe in microns
Bake temperature, F
Bake time, minutes
Source of CdTe
All other factors in the process will remain fixed at their current levels.
A relative and + for each of the factors around the current operating conditions is
determined for each factor.
H1317_CH17.qxd
10/18/07
11:51 AM
Page 159
Designed Experiments
Factor
Description
Thickness of Sn,
Thickness of CdTe,
Bake temperature, F
Bake time, minutes
Source of CdTe
A
B
C
D
E
159
20
30
350
40
United States
40
50
400
50
China
For a 2V51 fractional design, the base design will be that of a 24 full factorial where
the fifth factor will be generated by E = ABCD and I = ABCDE.
Interactions:
AB = CDE
AC = BDE
AD = BCE
AE = BCD
BC = ADE
BD = ACE
BE = ACD
CD = ABE
CE = ABD
DE = ABC
Since this is a resolution V, all main effects are clear and all two-factor interactions
are clear.
The 16 runs were replicated three times so that experimental error and statistical significance could be determined. The individual responses, response average, and standard
deviation are recorded in Table 2.
Calculation of Main Effects
A effect:
9.23 + 9.76 + 8.27 + 9.83 + 8.23 + 10.17 + 8.30 + 9.50
8
9.30 + 10.73 + 9.30 + 9.73 + 10.10 + 10.97 + 8.30 + 10.77
= 0.74
B effect:
C effect:
9.30 + 8.27 + 9.73 + 9.83 + 8.50 + 8.30 + 10.77 + 9.50
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
CDE
AB
+
+
+
+
+
+
+
ABCD
E
+
+
+
+
+
+
+
BDE
AC
+
BCE
AD
+
+
+
+
+
+
+
+
BCD
AE
+
+
+
+
+
+
+
+
ADE
BC
+
+
+
+
+
+
+
+
ACE
BD
+
+
ACD
BE
+
+
+
+
+
+
+
+
ABE
CD
+
+
+
+
+
+
ABD
CE
+
+
+
+
ABC
DE
9.3
9.1
10.8
9.8
9.5
8.0
9.9
9.8
10.4
8.5
10.8
10.2
8.6
7.8
11.0
9.2
Y1
9.4
9.4
10.8
9.6
9.0
8.4
9.6
9.9
10.5
8.0
11.1
9.7
8.2
8.4
10.6
9.9
Y2
9.2
9.2
10.6
9.9
9.4
8.4
9.7
9.8
9.4
8.2
11.0
10.6
8.7
8.7
10.7
9.4
Y3
9.30
9.23
10.73
9.76
9.30
8.27
9.73
9.83
10.10
8.23
10.97
10.17
8.50
8.30
10.77
9.50
Avg.
0.10
0.15
0.12
0.15
0.26
0.23
0.15
0.06
0.61
0.25
0.15
0.45
0.26
0.46
0.21
0.36
11:51 AM
Run
Response
10/18/07
H1317_CH17.qxd
Page 160
160
H1317_CH17.qxd
10/18/07
11:51 AM
Page 161
Designed Experiments
D effect:
10.10 + 8.23 + 10.97 + 10.17 + 8.50 + 8.30 + 10.77 + 9.50
8
9.30 + 9.23 + 10.73 + 9.76 + 9.30 + 8.27 + 9.73 + 9.83
= 0.05
E effect:
Two-Factor Interactions
AB interaction:
9.30 + 9.76 + 9.30 + 9.83 + 10.10 + 10.17 + 8.50 + 9.50
AC interaction:
9.30 + 10.73 + 8.27 + 9.83 + 10.10 + 10.97 + 8.30 + 9.50
AD interaction:
9.30 + 10.73 + 9.30 + 9.73 + 8.23 + 10.17 + 8.30 + 9.50
AE interaction:
10.73 + 9.76 + 9.30 + 8.27 + 10.10 + 8.23 + 10.77 + 9.50
161
H1317_CH17.qxd
162
10/18/07
11:51 AM
Page 162
BC interaction:
9.30 + 9.23 + 9.73 + 9.83 + 10.10 + 8.23 + 10.77 + 9.50
BD interaction:
9.30 + 9.23 + 9.30 + 8.27 + 10.97 + 10.17 + 10.77 + 9.50
BE interaction:
9.23 + 9.76 + 9.30 + 9.73 + 10.10 + 10.97 + 8.30 + 9.50
CD interaction:
9.30 + 9.23 + 10.73 + 9.76 + 8.50 + 8.30 + 10.77 + 9.50
CE interaction:
9.23 + 10.73 + 8.27 + 9.73 + 10.10 + 10.17 + 8.50 + 9.50
DE interaction:
9.23 + 10.73 + 9.30 + 9.83 + 8.23 + 10.97 + 8.50 + 9.50
H1317_CH17.qxd
10/18/07
11:51 AM
Page 163
Designed Experiments
163
Main and two-factor interaction effects are ranked in ascending order, and percent
median rank values are calculated as follows:
Order i
Effect value
Effect
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
0.74
0.54
0.52
0.27
0.06
0.03
0.01
0.05
0.07
0.08
0.11
0.14
0.16
0.29
1.28
A
C
E
AD
CD
CE
DE
D
AB
AE
BC
BE
AC
BD
B
4.5
11.0
17.5
24.0
30.5
37.0
43.5
50.0
56.5
63.0
69.5
76.0
82.5
89.0
95.5
The effects are plotted on normal probability paper to estimate which effects and interactions are significant.
A normal probability plot of effects is shown in the following figure:
99.99
99.9
99.8
99
98
95
90
80
70
60
50
40
30
AD
AB
20
C
E
C
10
5
2
1
.5
.2
.1
.05
.01
2.0
1.5
1.0
0.5
0.00
Value
0.5
1.0
1.5
Median rank
BD
BC
H1317_CH17.qxd
164
10/18/07
11:51 AM
Page 164
Based on the analysis of the normal probability plot, it appears that main effects A, C,
E, and B are statistically significant. In addition, consideration should be given to the BD
and AD interactions.
In order to maximize the power output factors, A, C, and E should be set at the level
and factor B at the + level.
Statistical Significance Based on t-test
Step 1. Calculate the pooled standard deviation.
Sp =
Sp =
Sp =
3.9855
Sp = 0.288
48
Error, E = t / 2 ,( r 1 ) 2 k f Sp
= 0.10
f =1
2
pr
p=8
r=3
k=5
2
E = 1.698 0.288
E = 0.14
( 8 )( 3 )
Any effect 0.14 that contains zero is judged not statistically significant at the 90 percent confidence level.
H1317_CH17.qxd
10/18/07
11:51 AM
Page 165
Designed Experiments
Order i
Effect value
Effect
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
0.74
0.54
0.52
0.27
0.06
0.03
0.01
0.05
0.07
0.08
0.11
0.14
0.16
0.29
1.28
A
C
E
AD
CD
CE
DE
D
AB
AE
BC
BE
AC
BD
B
90% confidence
interval
0.88
0.68
0.66
0.41
0.20
0.17
0.15
0.09
0.07
0.06
0.03
0.00
0.02
0.15
1.14
to
to
to
to
to
to
to
to
to
to
to
to
to
to
to
0.60
0.40
0.38
0.13
0.08
0.11
0.13
0.19
0.21
0.22
0.25
0.28
0.30
0.43
1.42
165
Statistically
significant?
Yes
Yes
Yes
Yes
No
No
No
No
No
No
No
No
Yes
Yes
Yes
Conclusion
Factors A, B, C, and E appear to be statistically significant at the 90 percent confidence
level. The following factor setting will maximize the output.
Set A, C, and E at the level, and set factor B at the + level.
Factor
A
B
C
E
D
Description
Optimum setting
Thickness of Sn
Thickness of CdTe
Bake temperature
Source of CdTe
Bake time
20 microns
50 microns
350
United States
Does not significantly affect power output; therefore,
set a lower time of 40 minutes to reduce cycle time.
Variation Reduction
Variation reduction is accomplished when we reduce the standard deviation; however,
standard deviations are not normally distributed and cannot be used directly as a response
to minimize. One approach is to take multiple observations during an experimental run,
calculate the standard deviation, and take the log, or ln. The resulting transformed response
can then be treated as a normally distributed response and the effects calculated in the traditional manner. The objective of the experiment is to reduce the log S, or ln S. Dr. George
Box (1978) of the University of WisconsinMadison and his colleagues published a series
of papers suggesting the use of log10(s) as a response variable when the objective is to
reduce variation.
H1317_CH17.qxd
10/18/07
166
11:51 AM
Page 166
Run
AB
AC
BC
ABC
Average
S2
1
2
3
4
5
6
7
8
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
1900
1250
1875
1160
2100
1740
1940
1350
3.9
5.4
15.1
14.8
10.3
8.0
13.6
11.1
15.21
29.16
228.01
219.04
106.09
64.0
184.96
123.21
(S )2 =
F=
( S ) 2 larger
( S ) 2 smaller
F=
133.57
= 1.23
108.85
For effect B:
(S )2 + =
(S )2 + =
F=
( S ) 2 larger
( S ) 2 smaller
F=
188.81
= 3.52
53.62
H1317_CH17.qxd
10/18/07
11:51 AM
Page 167
Designed Experiments
Effect
F value
A
B
C
AB
AC
BC
ABC
1.23
3.52
1.03
1.09
1.25
1.75
1.01
167
A critical F value is determined. This value is found in an F table using one-half the
degrees of freedom as calculated previously for the t value used in testing for statistical
significance.
( r 1) 2 k f
2
df = 8
df =
Factor
Effect
A
B
C
AB
AC
BC
ABC
+579
373
+236
74
+104
109
41
H1317_CH17.qxd
168
10/18/07
11:51 AM
Page 168
Experimental error, E = t / 2 ,( r 1 ) 2 k f Sp
2
pr
Si 2
= 10.38.
n
t / 2 ,( r 1 ) 2 k f Sp
2
2
= t 0.05 ,16 10.38
= (1.746 )(10.38 )( 0.41) = 7.4
pr
( 4 )( 3 )
At the 90 percent level of confidence, all of the factors in their interactions are statistically significant. This is because the magnitude of the effects is relatively large compared
to the experimental error of 7.4.
In this case, to maximize the compressive strength, we would set factors A and C to
the + level and factor B to the level. Setting B to the level also minimizes the variation.
Empirical Predictions
In a designed experiment, the number of + and conditions or settings is always the
same. The effects for the main factor and the interactions are calculated by taking the difference between the average of the + settings and the average of the settings for the
experimental data. In a balanced experiment, half the runs are set and the other half are
set +. The grand average of all the experimental data represents the expected average
response if all the calculated effects were equal to zero. We will add or subtract one-half
the effects to the grand average. Only those effects that are statistically significant will be
included in our empirical equation. There is one exception: If an interaction effect is statistically significant, it must be included in our model along with the main effects used in
the interaction effect. The main effects are included even if they are individually not statistically significant. This is due to the hierarchy rule for defining a model.
Consider the following example.
A full factorial experiment for three factors was conducted. The error for the effects is
7.5, and the average of all the responses is 125. The effects are:
Factor
A
B
C
AB
AC
BC
ABC
Effect
+60
6.5
+30
12
+8
5
7
H1317_CH17.qxd
10/18/07
11:51 AM
Page 169
Designed Experiments
169
Main effects A and C and two-factor interaction AB are statistically significant. Since
the AB interaction is to be included in the model, so must main effects A and B even though
B is not statistically significant. We will add one-half the effects to the grand average to
obtain the predicted expected response.
The values in parentheses are orthogonal values and can only take on the values
between 1 and +1.
Y = 125 + 30( A) 15( B) 6( A)( B) + 4( A)(C )
The factors A, B, and C in this example are defined by:
Factor
Description
()
(+)
A
B
C
Time, sec.
Pressure, psi
Surface finish, rms
15
20
70
45
50
120
In this example the response is pounds/bond strength between two samples of wood
that have been bonded together.
What would be the expected average bond strength if A = +1, B = 1, and C = +0.35?
We simply substitute the above values in our empirical equation and solve for the predicted response.
Y = 125 + 30( A) 15( B) 6( A)( B) + 4( A)(C )
Y = 125 + 30(+1) 15(1) 6(+1)(1) + 4(+1)(+0.35)
Y = 125 + 30 + 15 + 6 + 1.4
Y = 177.4
The values in this equation are orthogonal settings (OSs) and can vary from 1 to +1.
In order to evaluate the conditions, we must convert the OS values to real process or
machine settings (MSs). To convert from a specified OS (or OStgt) to the corresponding
MS value, we use the relationship
MS = 0.5[(OStgt )( MS)] + MS,
where: OStgt = the specified orthogonal value
MS = the average of the + and settings for the experiment
MS = the + setting minus the setting for the experiment.
Recall that for our experiment, the following and + settings were used:
H1317_CH17.qxd
170
10/18/07
11:51 AM
Page 170
Factor
Description
()
(+)
A
B
C
Time, sec.
Pressure, psi
Surface finish, rms
15
20
70
45
50
120
15 + 45
= 30.
2
MS or MS A = MS+ MS = 45 15 = 30
Factor
Description
()
(+)
MS
MS
A
B
C
Time, sec.
Pressure, psi
Surface finish, rms
15
20
70
45
50
120
30
20
50
30
35
95
For our example we had the OS values for the empirical model, and we set the factors
to +1, 1, and +0.35, respectively, for factors A, B, and C.
The +1 for A and 1 for B are no problem. These are simply the +1 setting for A, which
is 45 seconds, and the 1 setting for B, which is 20 psi. The +0.35 OS value for factor C
is a bit more complicated. We will use the conversion equation.
MS = 0.5[(OStgt )( MS)] + MS where the value is +0
0.35
MS = 0.5[(+0.35)(50)] + 95
MS = 0.5[17.5] + 95
MS = 8.75 + 95
= 103.75
In this example, we specified a set of orthogonal settings, predicted the expected
response, and converted the orthogonal settings to a set of corresponding machine settings.
Background
A valve manufacturer has evaluated the effect of two factors (seat chamfer angle and
hardness of o-ring) on the life of a valve. The response is the mean time between failure
(MTBF) in hours.
Factor
A
B
Description
Chamfer angle, degrees
O-ring hardness, shore A
()
(+)
46
62
32
78
H1317_CH17.qxd
10/18/07
11:51 AM
Page 171
Designed Experiments
171
Effects
Factor
Effect
A
B
AB
+350
450
120
We start by developing the empirical equation. All effects are statistically significant.
Adding one-half of each effect to the average of all responses, we have
Y = 1800 + 175( A) 225( B) 60( A)( B).
What is the expected average MTBF if we choose a chamfer angle of 38 and an o-ring
hardness, shore A, of 75?
These values are given in MSs and cannot be directly substituted into our empirical equation. We must convert each MS to an orthogonal setting (OS). To accomplish this, we can utilize the following relationship:
MStgt MS
,
OS = 2
MS
MSA =
Note that the MS may be negative. This can occur when the + setting is less than the
setting.
MStgt MS
OS = 2
MS
OS = 2
1
14
OS = 2
38 39
14
OS = 2(0.07) = 0.14
H1317_CH17.qxd
172
10/18/07
11:51 AM
Page 172
Performing a similar calculation for factor B set to an MS of shore A = 75, we find the
corresponding OS to be
MStgt = 75
MS = 70
MStgt MS
OS = 2
,
MS
OS = 2
MS = 16
75 70
,
16
OS = 2
5
= +0.63.
16
We now have OS values for factors A and B, which are +0.14 and +0.63, respectively.
These values will be used in our empirical equation:
Y = 1800 + 175( A) 225( B) 60( A)( B),
Y = 1800 + 175(+0.14) 225(+0.63) 60(+0.14)(+0.63),
Y = 1800 + 24.5 141.75 5.29,
Y = 1677.46.
We have demonstrated how to convert from MSs to OSs and vice versa when converting specific values of OS to MS. The following is an example of setting the OS values
to achieve an objective average response.
Testing Curvature or Nonlinearity of a Model (and Complete Case Study)
In traditional designs of experiments (DOEs) we determine effects to establish a linear or
first-degree relationship to model the expected average response as a function of the statistically significant factors. The resulting model will always work when the factors are set
to their () and (+) OSs. But will the model work at settings between the () and (+) OSs?
We will test for curvature or nonlinearity. The grand average of all our base-run experimental data is actually the average where half the settings were () and half were (+) or
equivalent to running all the experiments at an OS of (0). These (0) points, or what we call
center points, should yield the same average as that of the average of all the base-run
experimental data. Any difference in the average of the center points and the average of all
the base-run data is called curvature or the nonlinearity effect. If the difference is zero, we
have a perfect linear model. Center points can be determined only if the factor is quantitative and not qualitative. Examples of qualitative factors would be the type of polymer
used in an o-ring, the brand of cutting fluid, the production shift, and so on.
Consider the following case.
Objective: To achieve complete dissolution of a material in 40 seconds, with minimum
variation.
Factor
A
B
Description
Particle size, microns
pH
()
(+)
100
6.2
200
7.9
H1317_CH17.qxd
10/18/07
11:51 AM
Page 173
Designed Experiments
173
Experimental data:
Run
AB
Response
Average
Standard
deviation S
Variance
S2
1
2
3
4
+
+
32, 31, 32
58, 55, 61
101, 111, 93
18, 20, 21
31.7
58.0
101.7
19.7
0.58
3.00
9.02
1.53
0.34
9.00
81.36
2.34
Effect B =
In addition to the data obtained from the base-run experiment, we will collect data
from five center points. The setting for the center point for factor A will be the midpoint
or average of the two settings for the base experiment,
100 + 200
= 150.
2
The setting for the center point for factor B will be the midpoint or average of the two settings for the base experiment,
6.2 + 7.9
= 7.05.
2
Randomly during the base experiment, we will run five experiments where the particle size is 150 and the pH is 7.05. The response for our five center points is 53.1, 54.6,
48.5, 47.8, and 46.2.
We will calculate the curvature or nonlinearity effect NL by taking the difference
between the grand average of the base run Y B and the average of the center points Y C .
N L = Y B Y C = 52.78 50.04 = +2.74
H1317_CH17.qxd
174
10/18/07
11:51 AM
Page 174
, df B , dfC
SPC
4
NB
where:
a = 1 C and C = the level of confidence C = 0.95 for this example
dfb = (r 1)2KF = (3 1)22 = 8
dfc = # center points 1 = 4
t = t-score from table
t.025, 8+4 = t.025, 12 = 1.782
SPc = standard deviation with center points
SPC =
4
4
=
= 0.58
Nb
12
N B = total number of observations in the base experimeent = 12
Error of effects = (1.782)(4.46)(0.58) = 4.61
Any effects that are outside the limits of 4.61 are statistically significant at a level of
confidence of 95 percent.
We dont know the true value for effect A, but we are 95 percent confident it is
between 32.46 (27.85 4.61) and 23.24 (27.85 + 4.61). The fact that the interval does
not contain zero means that the probability that the true effect is zero is 5 percent or less.
A similar treatment can be made for all effects. Applying this criterion to all the
effects, we conclude that they are all statistically significant at a level of confidence of 95
percent. All effects will be included in our prediction model.
The empirical model is:
Effect A
Effect B
Effect AB
( A)( B),
( A) +
( B) +
2
2
2
54.15
27.85
+15.85
Y = 52.78 +
( B) +
( A)( B),
( A) +
2
2
2
Y = 52.78 + 13.93( A) + 7.93( B) 27.08( A)( B).
Y = Grand average +
H1317_CH17.qxd
10/18/07
11:51 AM
Page 175
Designed Experiments
175
We now have our linear empirical model to predict average responses as a function of
particle size and pH.
It should always be our objective to reduce variation wherever possible. Do we have
a factor that is responsible for variation reduction? To answer this question, we will perform an F-test. We calculate an F ratio for each of the main effects and compare this calculated F ratio (Fcalc) to a critical F ratio (Fcrit). If Fcalc > Fcrit, we reject the hypothesis that
the variance for the + setting is equal to the variance of the setting and conclude that the
factor in question does affect variation.
The Fcalc ratio is calculated by averaging the variances of the runs where the factor was
set positive, averaging the variances where the factor was set negative, and then dividing
the larger result by the smaller result.
Fcalc factor A:
(3.00)2 + (1.53)2 9 + 2.34
=
= 5.67,
2
2
(0.58)2 + (9.02)2 0.34 + 81.36
S12 =
=
= 40.85,
2
2
40.85
= 7.20.
Fcalc =
5.67
S+2 =
Fcalc factor B:
(9.02)2 + (1.53)2 81.36 + 2.34
=
= 41.85,
2
2
(0.58)2 + (3.00)2 0.34 + 9
=
= 4.67,
S2 =
2
2
41.85
Fcalc =
= 8.96.
4.67
S+2 =
Fcrit is determined by looking up the F value at a/2 and two equal degrees of freedom
where a = risk. For 95 percent confidence, the risk is 5 percent and a/2= 2.5 percent. The
two degrees of freedom are determined by
(# Replicates 1)(# Runs) (2)(4)
=
= 4,
2
2
= F2.5,4,4 = 9.40.
df1 = df2 =
Fcrit
Neither of the factors is significant at a 5 percent risk; however, if we increase our risk
(of being wrong) to 10 percent, the Fcrit becomes F10,4,4 = 6.39 at 90 percent confidence,
and both factors are significant in their effect on variation reduction. We will restrict factor B such that variation is minimal.
Recall that the calculated F ratio for factor B was calculated as
Fcalc =
S+2 41.85
=
.
S2
4.67
H1317_CH17.qxd
10/18/07
176
11:51 AM
Page 176
The greater the variance, the greater the variation. We have more variation when B is set
+; therefore, to minimize variation, we will set factor B to its OS. Letting B = , we set
B = 6.2.
Y = 52.78 13.93( A) + 7.93(11) 27.08( A)(1)
Y = 52.778 13.93( A) + 7.93 + 27.08( A)
Y = 44.85 + 13.15( A)
We now have a model that provides a minimum variation. But is a linear model
suitable in this case? Recall our calculation of the nonlinearity effect, NL = +2.74. If the
nonlinearity effect is sufficiently large, we should consider using another model such
as a second-degree relationship. Our linear model will always work at the extreme settings (OSs of +1 or 1) but not for settings between 1 and +1.
Consider the following case.
A 23 experiment has been conducted to evaluate three factors and how they affect
RMS surface finish.
Factor
A
B
C
Description
Cutting speed, rpm
Feed rate, inches/minute
Cutting angle, degrees
()
1000
1
5
(+)
1800
1.5
18
Predicted response
0
+
+
0
0
+
64
100
146
The model we have established assumes a linear relation is appropriate. The average
of our actual center points from experimental data yields 160, significantly different from
our predicted value of 100. The reason for this discrepancy is nonlinearity. See the following figure:
H1317_CH17.qxd
10/18/07
11:51 AM
Page 177
Response
Designed Experiments
177
160
146
100
64
A
B+
C
A
B
C
A+
B
C+
Even under the best circumstances we will always have some deviation from the
straight line linear model. The question is how much deviation or nonlinearity can we have
and still assume that a linear model is acceptable?
Recall that in our 22 case, the nonlinearity was
N L = Y B Y C = 52.78 50.04 = +2.74.
Is an NL of +2.74 acceptable? To answer this question, we will calculate the error of
nonlinearity. If the error of nonlinearity is greater than the nonlinearity effect, we conclude
that nonlinearity is not a problem and accept our linear model.
t a
2
, df B , dfC
SPC
1
1
+
N B CP
Note that the first and second terms are exactly the same as used in the calculation of
the error of effects. The third term changes slightly to
1
1
+
.
N B CP
NB is the number of total observations in the base experiment (12), and CP is the number
of center points (5 for our example).
t a
2
, df B , dfC
SPC
1
1
1 1
+
= (1.782)(4.46)
+ = (1.782)(4.46)(0.532) = 4.22
N B CP
12 5
H1317_CH17.qxd
178
10/18/07
11:51 AM
Page 178
The error of NL is greater than the NL; therefore, nonlinearity is not a problem at 95
percent confidence.
Y = 44.85 + 13.15( A)
Our objective is to achieve a dissolution time of 40 seconds. We have addressed the objective of minimizing variation by setting factor B to a 1 OS. We set to 40 and solve for the OS.
For factor A:
40 = 44.85 + 13.15(A).
Solving for A, we find the OS to be 0.37 in order to achieve a dissolution time of 40
seconds.
The OS of 0.37 must be converted to an equivalent MS.
OS
0.37
+1
MS
90
97.5
105
H1317_CH17.qxd
10/18/07
11:51 AM
Page 179
Designed Experiments
179
(S+ )
Example:
Consider the 22 experiment where the objective is to minimize the variation in the percent
shrinkage of pressure-treated wood. Note that we are not minimizing the shrinkage but
rather the variation in the shrinkage. Two factors will be evaluated. Factor A is pressure,
where + = 150 psi and = 100 psi. Factor B is time, where + = 4 hours and = 2 hours.
Four experimental runs are planned. Three replicates are run, and the percent shrinkage is
determined for each run.
Run
AB
1
2
3
4
+
+
Responses
2.31,
2.46,
4.86,
3.75,
2.42,
2.51,
4.77,
3.87,
2.46
2.53
4.28
4.61
Average
Standard deviation
2.397
2.500
4.637
4.077
0.0777
0.0361
0.3121
0.4658
2
2
B main effect:
0.3121 + 0.4658 0.0777 + 0.361 = 0.3890 0.0569 = 0.3321
2
2
AB interaction:
0.0777 + 0.4658 0.0361 + 0.3121 = 0.2718 0.1741 = 0.0977
2
2
Test for statistical significance as follows:
Effect A:
2
2
S+
0.0361 + 0.4658
=
= 0.0630
2
S
0.0777 + 0.3121
=
= 0.0380
ln
(S+ )
(S )
= ln
0.0630
= 0.51 0.51 >/ 1.96; therefore, facto
or A is not significant.
0.0380
H1317_CH17.qxd
10/18/07
180
11:51 AM
Page 180
Effect B:
2
S+2
0.3121 + 0.4658
=
= 0.1513
S2
0.0777 + 0.0361
=
= 0.0032
ln
(S+ )2
2
(S )
= ln
0.1513
= 3.86 3.86 > 1.96; therefore, factor B is significant.
0.0361
Effect AB interaction:
2
0.0777 + 0.4658
S+2 =
= 0.0738
2
2
0.0361 + 0.3121
S2 =
= 0.0303
2
ln
(S+ )2
2
(S )
= ln
0.0738
= 0.89 0.89 >/ 1.96; therefore, the AB interaction is not significant.
0.0303
Conclusion
To minimize the variation in the actual percent shrinkage, the time of treatment should be
two hours. The pressure is not a statistically significant factor with respect to minimizing
variation. Note that we reverse the signs of the effect to determine the setting for the factor in order to minimize the response. Since the sign for the main effect for B is +, we chose
to set factor B to the setting, or two hours.
No calculated effects were made using the shrinkage response; however, it is obvious
that setting B would minimize the actual percent shrinkage as well as the variation in the
shrinkage.
++++
n = 12
++++++
n = 16
++++++++
n = 20
++++++++++
n = 24
++++++++++++
H1317_CH17.qxd
10/18/07
11:51 AM
Page 181
Designed Experiments
181
To construct the design matrix, the appropriate generating column is selected. The first
column is generated by placing the signs in a vertical column. The last sign of the first column becomes the first sign in the second column, and the remainder of the second column
follows the sequence designated by the generating column. After n 1 columns have been
developed, the final row is made by using all s.
An example is shown for a seven-factor, eight-run Plackett-Burman design.
Run
1
2
3
4
5
6
7
8
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
In the event that fewer than eight factors are to be screened, simply drop the required
number of columns (factors).
Case Study
You have been asked to examine a process with respect to three factors that are suspected
to influence a response. The objective is for the process to yield a response of 55.0 with a
minimum of variation. The factors and their settings for this experiment are as follows:
Setting
Factor
Description
A
B
C
40
30
22
60
70
30
Run
AB
AC
BC
ABC
II
Average
Standard
deviation
1
2
3
4
5
6
7
8
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
40.48
30.01
50.92
42.05
69.69
47.95
69.54
42.46
45.52
27.99
67.08
31.95
72.31
50.05
64.46
47.54
43.0
29.0
59.0
37.0
71.0
49.0
67.0
45.0
3.56
1.43
5.71
7.14
1.85
1.48
3.59
9.98
H1317_CH17.qxd
182
10/18/07
11:51 AM
Page 182
4
4
Main effect C:
71 + 49 + 67 + 45 43 + 29 + 59 + 37 = +16
4
4
Interaction effect AC:
43 + 59 + 49 + 45 29 + 37 + 71 + 67 = 2
4
4
Interaction ABC:
29 + 59 + 71 + 45 43 + 37 + 49 + 67 = 2
4
4
Main effect B:
59 + 37 + 67 + 45 43 + 29 + 71 + 49 = +4
4
4
Interaction effect AB:
43 + 37 + 71 + 45 29 + 59 + 49 + 67 = 2
4
4
Interaction BC:
43 + 29 + 67 + 45 59 + 37 + 71 + 49 = 8
4
4
Determine the statistical significance for effects based on average responses. The level
of confidence is chosen to be 95 percent. The error for the calculated effects is given by
Error, E = t / 2 ,( r 1 ) 2 k f Sp
where: = 0.05
r=2
f=0
p = 4.
2 ,
pr
H1317_CH17.qxd
10/18/07
11:51 AM
Page 183
Designed Experiments
Sp =
183
2
= 0.50
(4)(2)
t 0.025 ,8 = 2.303
Error, E = ( 2.303 )( 5.20 )( 0.50 ) = 5.99
Any effects greater than the experimental error 5.99 are deemed statistically significant at the 95 percent level. The effects that are significant are main effects A, C, and the
two-factor interaction effect BC.
The model for predicting the response using only those effects that are significant is
Yexpected = grand average 12 (effect A)(A) + 12 (effect C)(C) 12 (effect BC)(B)(C),
Yexpected = 50 5(A) + 8(C) 4(B)(C).
Effects based on standard deviation are as follows:
Main effect A:
1.43 + 7.14 + 1.48 + 9.98 3.56 + 5.71 + 1.85 + 3.59 = 0.62
4
4
Main effect B:
5.71 + 7.14 + 3.59 + 9.98 3.56 + 1.43 + 1.85 + 1.48 = 2.31
4
4
Main effect C:
1.85 + 1.48 + 3.59 + 9.98 3.56 + 1.43 + 5.71 + 7.14 = 0.11
4
4
Interaction AB effect:
3.56 + 7.14 + 1.85 + 9.98 1.43 + 5.71 + 1.48 + 3.59 = 1.23
4
4
Interaction AC effect:
3.56 + 5.71 + 1.48 + 9.98 1.43 + 7.14 + 1.85 + 3.59 = 0.78
4
4
Interaction BC effect:
3.56 + 1.43 + 3.59 + 9.98 5.71 + 7.14 + 1.85 + 1.48 = 0.27
4
4
H1317_CH17.qxd
10/18/07
184
11:51 AM
Page 184
4
4
Determine the statistical significance for effects based on standard deviation as a
response. Each effect based on standard deviation will be tested independently for statistical significance. The level of significance chosen is 90 percent.
The test statistic will be the natural log of the ratio of the square of the average standard deviation of the s to the +s.
Test statistic: ln
( S+ ) 2
( S ) 2
If the test is greater than Z/2, we will reject the hypothesis that the effect is not statistically significant at a level of confidence of 1 . In this case, the test statistic must be
greater than 1.645.
Calculate the test statistic for the effects as follows:
Main effect A:
2
4
2
4
ln
(S+ )2
5.01
25.10
= ln
= ln 1.85 = 0.62..
2 = ln
(S )
3.68
13.54
Since the test statistic is less than the critical value of 1.645, we cannot reject the null
hypothesis that the variation from a positive setting of A is any different than the variation
from a negative setting of A. We conclude that factor A is not a significant factor affecting
variation.
Similar calculations for all of the factors and interactions give the following test statistics:
Factor
A
B
C
AB
AC
BC
ABC
Test statistic
0.62
2.31 (Statistically significant at a confidence level of 90%)
0.11
1.23
0.78
0.27
0.37
H1317_CH17.qxd
10/18/07
11:51 AM
Page 185
Designed Experiments
185
Bibliography
Barrentine, L. B. 1999. An Introduction to Design of Experiments. Milwaukee, WI: ASQ
Quality Press.
Box, G. E. P., W. G. Hunter, and J. S. Hunter. 1978. Statistics for Experimenters. New
York: John Wiley & Sons.
Montgomery, D. C. 1997. Design and Analysis of Experiments. 4th edition. New York:
John Wiley & Sons.
Schmidt, S. R., and R. G. Launsby. 1994. Understanding Industrial Designed Experiments. 4th edition. Colorado Springs, CO: Air Academy Press.
H1317_CH17.qxd
10/18/07
11:51 AM
Page 186
H1317_CH18.qxd
10/15/07
4:26 PM
Page 187
Discrete Distributions
When the characteristic being measured can only take on integer values such as 0, 1,
2, . . . , the probability distribution is a discrete distribution. The number of nonconforming units in a sample would be a discrete distribution.
Hypergeometric Distribution
The hypergeometric distribution is the only discrete distribution that has the lot or population size as an element. The hypergeometric distribution is appropriately used when randomly sampling n items without replacement from a lot of N items of which D items are
defective. The probability of finding exactly x defective items is given by
D N D
x n x
,
P( x ) =
N
n
where: D = Nonconforming units in the population or lot
N = Size of the population or lot
n = Sample size
x = Number of nonconforming units in the sample.
a
a!
Note: The expression
.
b b !(a b )!
The mean and standard deviation of the hypergeometric distribution are
Mean , =
nD
N
and
Standard deviation, =
187
nD
D N n
.
1
N
N
N 1
H1317_CH18.qxd
188
10/15/07
4:26 PM
Page 188
Binomial Distribution
The binomial distribution is independent of the population size and assumes the following:
1. Only two outcomes are possible for each trial. In quality applications, the only outcomes resulting from an inspection are that the items are either conforming or nonconforming to a requirement. Only two possibilities are availablegood or bad.
2. Independence of outcome means that the result of one trial does not influence the
outcome of another trial. For example, tossing a coin and getting heads does not
influence the probability of getting heads on the next toss. The probability of getting heads in a fair toss is always 50 percent.
H1317_CH18.qxd
10/15/07
4:26 PM
Page 189
Discrete Distributions
189
n
P( x ) = P x Q n x
x
68
P( x = 3) = (0.07)3 (0.93)65 = (50116)(0.00034)(0.0089) = 0.1517
3
Example 2:
A process is running approximately 25 percent defective. What is the probability that three
or fewer defective units will be found in a random sample of 20 units?
The probabilities of getting exactly zero, one, two, and three must be determined and
these individual probabilities summed in order to determine the probability of getting three
or fewer.
P(x 3) = P(x = 0) + P(x = 1) + P(x = 2) + P(x = 3)
The probability of getting exactly zero is a unique case for the binomial equation. The
binomial equation for this case reduces to
P(x = 0) = (1 p)n
H1317_CH18.qxd
10/15/07
190
4:26 PM
Page 190
or
P(x) = Qn.
In this example, P(x = 0) = (0.75)20 = 0.0032.
The probability of getting exactly one defective is
20
P ( x = 1) = (0.25)1 (0.75)19 = (20)(0.25)(0.0042) = 0.0211.
1
In a similar manner, we calculate the probability of getting exactly two and three
defective units.
20
P ( x = 2) = (.25)2 (.75)18 = 0.0669,
2
20
P ( x = 3) = (.25)3 (.75)17 = 0.1339.
3
The probability of getting three or fewer defective units is the sum of the probabilities
of getting zero, one, two, and three defective units.
P(x 3) = 0.0032 + 0.0211 + 0.0669 + 0.1339 = 0.2281
Poisson Distribution
The Poisson distribution can be thought of as a limiting form of the binomial distribution.
If the factor p in the binomial approaches zero and the sample n approaches infinity such
that the term np approaches a constant , the binomial becomes the Poisson distribution.
The Poisson distribution has only one parameter , which is the product of the process
proportion defective p and the sample size n chosen from the process or population.
The Poisson distribution is given by
P( x ) =
e x
,
x!
where = np.
The Poisson is unique in that both the variance and the mean are equal to .
Mean, = = np
Standard deviation, = np
A typical application of the Poisson distribution is that of the u and c control charts,
where defects per unit or defects are monitored with respect to time. Any randomly
H1317_CH18.qxd
10/15/07
4:26 PM
Page 191
Discrete Distributions
191
occurring event that takes place on a per unit basis (such as defects per hour, defects
per meter, or defects per square yard) represents a suitable application of the Poisson
distribution.
Example 1:
A textbook has 450 pages on which nine typographical errors are found. These errors are
randomly distributed throughout the text. What is the probability of not finding an error on
40 randomly selected pages?
p = proportion nonconforming =
9
= 0.02
450
np = = ( 40 )( 0.02 ) = 0.89
x=0
P( x ) =
e x
e 0.8 0 ( 0.4493 )(1)
P( x = 0 ) =
=
= 0.4493
x!
0!
1
Example 2:
During a period of one week, 126,000 garments have been manufactured in which
4536 defects were found. If the process continues to perform at this level, what is the probability of finding no more than two defects in a sample of 60 garments?
p = proportion defective =
4536
= 0.036
126 , 000
n = sample size = 60
np = = ( 60 )( 0.063 ) = 2.16
The probability of finding no more than two defects is the probability of finding two
or fewer defectives. This probability is determined by summing the probabilities of finding exactly zero, one, and two defects.
P ( x 2 ) = P ( x = 0 ) + P ( x = 1) + P ( x = 2 )
e x e 2.16 0 ( 0.1153 )(1)
=
=
= 0.1153
x!
0!
1
e 2.16 1 ( 0.1153 )( 2.16 ) 1
P ( x = 1) =
=
= 0.2490
1!
1!
e 2.16 2 ( 0.1153 )( 2.16 ) 2
P( x = 2 ) =
=
= 0.2690
2!
2!
P( x = 0 ) =
The probability of getting two or fewer defectives is the sum of the probabilities of
getting zero, one, and two defectives.
P(x 2) = 0.1153 + 0.2490 + 0.2690 + 0.6333
H1317_CH18.qxd
192
10/15/07
4:26 PM
Page 192
1
1
[ ( a np ) 2 / np ( 1 p ) ]
e2
2 np (1 p )
2np(1 p)
2np(1 p)
where = the cumulative normal distribution function
The normal distribution is appropriate and acceptable for p of approximately 0.50 and
n > 10. For larger values of p, larger sample sizes are needed. The normal distribution is
not appropriate for p < 1/(n + 1) or p > n/(n + 1).
Example 1:
A sample of n = 80 is chosen from a population characterized by p = 0.15. What is the
probability of getting exactly two defectives using the normal distribution?
np > 10 and p < .50
n = 80
p = 0.15
np = 12.0
2 + 0.5 12
2 0.5 12
P( x = 2) =
2(3.14)(12)(1 0.15)
2(3.14)(12)(1 0.15)
P( x = 2) = ( 1.19) ( 1.31)
P( x = 2) = 0.11702 0.09510
P( x = 2) = 0.02192
Example 2:
A sample of n = 60 items is chosen from a process where p = 0.15. What is the probability of getting exactly two defectives using the normal, binomial, and Poisson distribution
as a model? Which distribution would be the best for this calculation?
H1317_CH18.qxd
10/15/07
4:26 PM
Page 193
Discrete Distributions
2 np (1 p )
2 np (1 p )
P( x = 2 ) =
2 + 0.5 9
2 ( 3.14 )( 9 )(1 .15 )
2 0.5 9
P ( x = 2 ) = ( 0.94 ) (1.08 )
P ( x = 2 ) = 0.17361 0.14007
P ( x = 2 ) = 0.03354
Using the binomial distribution:
n = 60
p = 0.15
q = 0.85
x=2
n
P( x = 2 ) = p x q n x
x
60
P ( x = 2 ) = ( 0.15 ) 2 ( 0.85 ) 60 2
2
P ( x = 2 ) = (1770 )( 0.0225 )( 0.000081) = 0.0032
Using the Poisson distribution:
e x
x!
e 9.0 ( 9.0 ) 2
P( x = 2 ) =
2!
( 0.000123 )( 81)
P( x = 2 ) =
2
P ( x = 2 ) = 0.0050
P( x ) =
193
H1317_CH18.qxd
194
10/15/07
4:26 PM
Page 194
Summary:
Distribution
Normal
Binomial
Poisson
P(x = 2)
0.0335
0.0032
0.0050
Pascal Distribution
The Pascal distribution is similar to the binomial distribution in that it is based on independent trials with the probability of success for a given trial not being influenced by the
result of other trials. If we have a series of independent trials with a probability of success of p, then x will be the trial on which the rth success will occur. This value of x is a
Pascal random variable. The probability distribution function defining this function is
given by
x 1 r
P( x ) =
p (1 p ) x r x = r , r + 1, r + 2 , . . .
r 1
A special case of the Pascal distribution is when r = 1. In this case, we are interested
in determining the probability of the number of trials required until the first success.
This particular case where we fix the number of successes to one and determine the
probability of the number of trials is the geometric distribution. Other values of r > 0 (and
r is not necessarily an integer) define the negative binomial distribution.
Example 1:
Given a process performing at p = 0.08, what is the probability of getting two defectives
on the fifteenth sample?
p = 0.08
x = 15
r=2
x 1 r
P( x ) =
p (1 p ) x r
r 1
14
P ( x = 15 ) = ( 0.08 ) 2 ( 0.92 ) 13
1
P ( x = 15 ) = (14 )( 0.0064 )( 0.3383 )
P ( x = 15 ) = 0.0303
H1317_CH18.qxd
10/15/07
4:26 PM
Page 195
Discrete Distributions
195
4.1
4.2
4.3
4.4
4.5
4.6
4.7
4.8
4.9
5.0
0
1
2
3
4
5
6
7
0.017
0.085
0.224
0.414
0.609
0.769
0.879
0.943
0.015
0.078
0.210
0.395
0.590
0.753
0.868
0.936
0.014
0.072
0.197
0.377
0.570
0.737
0.856
0.929
0.012
0.066
0.185
0.360
0.551
0.720
0.844
0.921
0.011
0.061
0.174
0.342
0.532
0.703
0.831
0.913
0.010
0.056
0.163
0.326
0.513
0.686
0.818
0.905
0.009
0.052
0.152
0.310
0.495
0.668
0.805
0.896
0.008
0.048
0.143
0.294
0.476
0.651
0.791
0.887
0.007
0.044
0.133
0.279
0.458
0.634
0.777
0.877
0.007
0.040
0.125
0.265
0.441
0.616
0.762
0.867
H1317_CH18.qxd
10/15/07
196
4:26 PM
Page 196
The following figure shows the selection of distributions for the approximation of the
hypergeometric:
Hypergeometric
P(x) =
D
x
ND
nx
N
n
n/N 0.1
p 0.1
p-Binomial
P(x) =
n x
p qnx
x
f-Binomial
P(x) =
np < 5
np 5
Poisson
Normal
p > 0.1
Np
x
n
N
None
n
N
Npx
P(x) =
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Gryna, F. M., R.C.H. Chua, and J. A. De Feo. 2007. Jurans Quality Planning and Analysis
for Enterprise Quality. 5th edition. New York: McGraw-Hill.
Hayes, G. E., and H. G. Romig. 1982. Modern Quality Control. 3rd edition. Encino, CA:
Glencoe.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Petruccelli, J. D., B. Nandram, and M. Chen. 1999. Applied Statistics for Engineers and
Scientists. Upper Saddle River, NJ: Prentice Hall.
Sternstein, M. 1996. Statistics. Hauppauge, NY: Barrons Educational Series.
Walpole, R. E., and R. H. Myers. 1993. Probability and Statistics for Engineers and Scientists. 5th edition. Englewood Cliffs, NJ: Prentice Hall.
H1317_CH19.qxd
10/17/07
2:24 PM
Page 197
Evolutionary Operation
% Solids, Factor A
60.0
58.0
56.0
58.9
x
59.3
x
x
58.8
x
59.0
54.0
52.0
50.0
0.4
0.6
0.8
1.0
1.2
Coating thickness, mils
Factor B
197
1.4
H1317_CH19.qxd
10/17/07
198
2:24 PM
Page 198
The initial trial change will be to set the percentage of total solids to 52.0 percent and
56.0 percent (2 percent) around the normal operating condition and to set the coating
thickness to 0.6 and 1.0 mils.
The operating conditions and the responses are treated as a traditional 22, two-level
experiment.
Factor
Description
A
B
% Solids
Coating thickness
52.0
0.6
56.0
1.0
Phase 1, cycle 1:
Run
Response
1
2
3
4
+
+
58.8
58.9
59.0
59.3
The effects will not be calculated, because with only a single observation, we cannot
determine the experimental error. An additional set of runs (cycle 2) will be made using
the same set of operating conditions (phase 1).
Phase 1, cycle 2:
Run
Cycle 1
Cycle 2
Average
Standard deviation, S
Variance, S 2
1
2
3
4
+
+
58.8
58.9
59.0
59.3
59.1
58.8
58.4
60.0
58.95
58.85
58.70
59.65
0.212
0.071
0.424
0.495
0.0449
0.0050
0.1798
0.2450
Main effects:
A:
2
2
B:
2
2
H1317_CH19.qxd
10/17/07
2:24 PM
Page 199
Evolutionary Operation
199
Error of effects:
t / 2 ,( r 1 ) 2 k f Sp
2
pr
where: = risk
k = number of factors
f = number of times the experiment is fractionated
r = number of replicates or cycles
Sp = pooled standard deviation
p = number of +s in a column
For this example, we will use a 95 percent confidence level ( = 0.05).
= 0.05
k =2
f =0
r =2
p =2
Sp = 0.345
E = (2.776)(0.345)(0.707)
E = 0.68
Since the effects are smaller than the error, we cannot conclude that the effects are statistically significant at the 95 percent level of confidence. We will replicate the experiment
again by running cycle 3.
Phase 1, cycle 3:
Run
Cycle 1
Cycle 2
Cycle 3
1
2
3
4
+
+
58.8
58.9
59.0
59.3
59.1
58.8
58.4
60.0
58.7
59.5
59.6
61.0
Average
58.87
59.07
59.00
60.10
Standard
deviation, S
Variance, S 2
0.208
0.379
0.600
0.854
A:
2
2
B:
2
2
0.0433
0.1436
0.3600
0.7293
H1317_CH19.qxd
200
10/17/07
2:24 PM
Page 200
Error of effects:
t / 2 ,( r 1 ) 2 k f Sp
2
pr
= 0.05
k =2
f =0
r =3
p =2
Sp = 0.565
E = (2.306)(0.565)(0.58) = 0.756
Since the error is still larger than the effects, cycle 4 will be run.
Phase 1, cycle 4:
Run
Cycle
1
Cycle
2
Cycle
3
Cycle
4
Average
Standard
deviation, S
Variance,
S2
1
2
3
4
+
+
58.8
58.9
59.0
59.3
59.1
58.8
58.4
60.0
58.7
59.5
59.6
61.0
58.9
59.1
59.0
59.9
58.88
59.08
59.00
60.05
0.170
0.310
0.490
0.705
0.0289
0.0961
0.2401
0.4970
A:
2
2
B:
2
2
Error of effects:
t / 2 ,( r 1 ) 2 k f Sp
= 0.05
k=2
f=0
r=4
p=2
Sp = 0.46
2
pr
H1317_CH19.qxd
10/17/07
2:24 PM
Page 201
Evolutionary Operation
201
E = (2.179)(0.46)(0.50) = 0.50
The error is now less than the effect, so we conclude that the effects are statistically
significant. Factor A and factor B should both be set positive (+) in order to maximize the
peel strength.
% Solids, Factor A
60.0
Phase 1
58.0
56.0
59.08
x
60.05
x
x
58.88
x
59.00
54.0
52.0
50.0
0.4
0.6
0.8
1.0
1.2
Coating thickness, mils
Factor B
1.4
A new set of conditions will be established (phase 2) around the new settings, where
A is + (percentage of solids = 56.0 percent) and B is + (coating thickness = 1.0).
Factor
Description
A
B
% Solids
Coating thickness
54.0
0.8
58.0
1.2
Phase 2, cycle 1:
Run
Response
1
2
3
4
+
+
58.9
59.3
59.5
60.6
H1317_CH19.qxd
10/17/07
202
2:24 PM
Page 202
Phase 2, cycle 2:
Run
Cycle 1
Cycle 2
Average
Standard deviation, S
Variance, S 2
1
2
3
4
+
+
58.9
59.3
59.5
60.6
59.1
59.5
59.4
60.9
59.00
59.40
59.45
60.75
0.141
0.141
0.071
0.212
0.0199
0.0199
0.0050
0.0449
Main effects:
59.40 + 60.75 59.0 + 59.45
A:
2
2
59.45 + 60.75 59.00 + 59.40
B:
2
2
Error of effects:
t / 2 ,( r 1 ) 2 k f Sp
2
pr
= 0.05
k=2
f=0
r=2
p=2
Sp = 0.150
E = (2.776)(0.150)(0.707) = 0.817
The error for both effects is statistically significant.
H1317_CH19.qxd
10/17/07
2:24 PM
Page 203
Evolutionary Operation
203
% Solids, Factor A
60.0
58.0
59.4
x
60.76
x
x
59.0
x
59.45
56.0
54.0
52.0
50.0
0.4
0.6
0.8
1.0
1.2
Coating thickness, mils
Factor B
1.4
After phase 2, we would devise another set of process conditions (phase 3) to continue
our process optimization. The suggested conditions for phase 3 are as follows:
Phase 3, cycle 1:
Factor
Description
A
B
% Solids
Coating thickness
56.0
1.0
60.0
1.4
The EVOP is continued until no additional gain is achievable at the designated level
of confidence.
Bibliography
Box, G. E. P., and N. R. Draper. 1969. Evolutionary Operation. New York: John Wiley &
Sons.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Montgomery, D. C. 1997. Design and Analysis of Experiments. 4th edition. New York:
John Wiley & Sons.
H1317_CH19.qxd
10/17/07
2:24 PM
Page 204
H1317_CH20.qxd
10/18/07
12:56 PM
Page 205
Exponentially Weighted
Moving Average
The individual/moving range control chart has the following two modest disadvantages:
1. It is relatively weak in detecting a small process change in a timely manner
2. It is potentially affected by nonnormal distributions
An alternative to the individual/moving range control chart (and the average/range
control chart) is the exponentially weighted moving average (EWMA) control chart.
An EWMA is a moving average of past data where each of the data points is assigned
a weight. These weights decrease in an exponential decay manner from the present point
value to subsequent points. The EWMA tends to reflect the current condition biased with,
to varying degrees, the values of the immediate historical data points. The amount of
decrease of the weights is an exponential function of the weighting factor , which can
assume any value between zero and one.
Wt1 = (1 )1
where: Wt 1 = weight associated with observation Xt 1, with Xt being the most recent
observation
= weighting factor 0 to 1
When is very small, the moving average at any time t carries with it a great deal of
inertia from the previous data; therefore, it is relatively insensitive to short-lived process
changes.
In selecting the value for , use the following relationship between the weighting fac
tor and the sample size n for a Shewhart X /R control chart:
=
2 .
n +1
If = 1 is selected for the EWMA, all of the weight is given to the current sample, and
in effect it is an individual Shewhart control chart.
The weights (1 )1 decrease geometrically with the age of the sample. If = 0.2,
then the weight assigned to the current sample (used in calculating EWMA) is 0.20, and
the weights given to the preceding samples are 0.16, 0.128, 0.1024, and so on. Because
205
H1317_CH20.qxd
206
10/18/07
12:56 PM
Page 206
2
If the X i are independent random variables with variance
, then the variance of
n
Zt is
2 zt =
2
[1 (1 )2 t ].
n 2
2z =
2
.
n 2
The upper and lower control limits for the EWMA are
UCL/LCL = X 3
.
2
X3
2t
[1 (1 ) ].
n 2
H1317_CH20.qxd
10/18/07
12:56 PM
Page 207
Sample
t
Xt
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
60
91
132
100
80
120
70
95
14
80
110
100
82
105
65
207
X = 95.3
S = 23.5
Note: The sample standard deviation S is used as an
estimate of . Any appropriate estimate based on
SPC relationships may be used.
Zt = EWMA at time t
= weighting factor ( = 0.2 for this example)
Z t 1 = EWMA for the immediate preceding point.
H1317_CH20.qxd
208
10/18/07
12:56 PM
Page 208
The remaining EWMAs are calculated and tallied in the following table, which summarizes individual observations Xt and EWMAs Zt:
Sample
t
Xt
Zt
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
60
91
132
100
80
120
70
95
140
80
110
100
82
105
65
88.2
88.8
97.4
97.9
94.3
99.4
93.5
93.8
103.0
98.4
100.7
100.6
96.9
98.5
91.8
H1317_CH20.qxd
10/18/07
12:56 PM
Page 209
209
Sample
Xt
Zt
LCL
UCL
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
60
91
132
100
80
120
70
95
140
80
110
100
82
105
65
88.2
88.8
97.4
97.9
94.3
99.4
93.5
93.8
103.0
98.4
100.7
100.6
96.9
98.5
91.8
81.2
77.2
75.1
73.9
73.1
72.6
72.3
72.1
72.0
.
.
.
.
.
.
109.4
113.4
115.5
116.7
117.5
118.0
118.3
118.5
118.6
.
.
.
.
.
.
H1317_CH20.qxd
210
10/18/07
12:56 PM
Page 210
120
110
= 0.2
Cl = 95.3
100
90
80
LCL = 72.0
70
1
10
15
Step 5. Continue to collect and plot the data, looking for signs of a change.
Ten new data values are obtained. Calculate the EWMA values and plot the results.
Sample
#
Xi
(15
16
17
18
19
20
21
22
23
24
25
65
110
115
111
118
126
117
122
131
124
129
Zt
91.9)
95.4
99.4
101.7
104.9
109.2
110.7
112.3
116.6
118.1
120.2
H1317_CH20.qxd
10/18/07
12:56 PM
Page 211
211
120
History
110
= 0.2
Cl = 95.3
100
90
80
LCL = 72.0
70
1
10
15
20
25
Bibliography
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Sweet, A. L. 1986. Control Charts Using Coupled Exponentially Weighted Moving Averages. Transactions of IIE 18, no. 1: 2633.
Wheeler, D. J. 1995. Advanced Topics in Statistical Process Control. Knoxville, TN: SPC
Press.
H1317_CH20.qxd
10/18/07
12:56 PM
Page 212
H1317_CH21.qxd
10/17/07
2:33 PM
Page 213
F-test
One of the main objectives in the pursuit of quality in manufacturing (and service for that
matter) is to reduce variability. This goal is consistent with the Taguchi philosophy of performing at the target with minimum variation. One of the best methods to assess whether
variation reduction has been accomplished is to compare the variances (or square of the
standard deviation).
Consider a process variable 1 that has a standard deviation of 1 and a process variable
2 that has a standard deviation 2. For example, 1 might represent the standard deviation
of the tablet weight of a pharmaceutical product before attempts were made to reduce variability, and 2 might represent the standard deviation after efforts were made to reduce the
variation. A sample of seven tablets is taken from the production line prior to the improvement, and five tablets are taken from the production line after implementation of the modification. By comparing the variances of the two samples, we may test the hypothesis that
one of the variances is greater than the other. This test relies on the F statistic, named after
Ronald A. Fisher (18901962), who developed the probability distribution in the 1920s.
The weights of tablets prior
to process improvement:
12.3
12.7
12.5
12.6
12.4
12.9
12.2
S1 = 0.24
S2 = 0.08
S12 = 0.0576
S22 = 0.0064
The F statistic or ratio is determined by dividing the larger variance by the smaller variance.
F=
S12 0.0576
=
= 9.00
S22 0.0064
213
H1317_CH21.qxd
214
10/17/07
2:33 PM
Page 214
With this F-test, we are testing the hypothesis that 12 is greater than 22 (the alternative
hypothesis), as opposed to 12 equal to 22 (the null hypothesis).
Ho: 12 = 22
Ha: 12 > 22
We will reject the null hypothesis and accept the alternative hypothesis if our F calculated
value is greater than the critical F value. A level of significance must be chosen. For this
example, a confidence level of 95 percent is selected, meaning that a risk, = 5%, is
accepted.
The critical F value is found in Table A.13 in the appendix. The degrees of freedom
are determined by the sample size minus one. The horizontal rows represent the degrees
of freedom for the sample giving the smaller variance, v2. The vertical columns represent
the degrees of freedom for the sample giving the larger variance, v1. Each row of v2 has
four levels of corresponding to 10, 5, 2.5, and 1 percent.
For this example v1 = 7 1 = 6 (the sample size of the larger variance) and v2 = 5 1
= 4 for the smaller variance.
10
5
2.5
1
v1
v2 ............... 6
4
4.01
6.16
9.2
15.2
The critical F value for v1 = 6, v2 = 4, and a risk of = 0.05 is 6.16. Since our calculated
F-ratio is 9.0 and exceeds 6.16 we will reject the null hypothesis (12 22) and accept
the alternative hypothesis (12 > 22), concluding that a reduction in variation has been
demonstrated.
H1317_CH21.qxd
10/17/07
2:33 PM
Page 215
F-test
215
Example 1:
Two different methods of training have been evaluated. One test group consists of 16 participants, and the other consists of 11. The respective standard deviations are 14.0 and 7.
Assume a risk, = 0.05.
The F-ratio is
F=
196
= 4.00 / 2 = 0.025.
49
The critical F-ratio, F15,10,0.025 = 3.52. Since F calculated is greater than F critical, we reject
the null hypothesis, Ha: 12 = 22, and accept the alternative hypothesis, Ha: 12 22.
Example 2:
A method of testing (X) has been used extensively to measure the water content in a latex
co-polymer. Based on a very large number of test results, this method has a standard deviation of SX = 0.071 units. A new method (Y) has been developed in an effort to reduce the
variation of test results. Eight independent tests gave a standard deviation of Sy = 0.038. It
would be an improvement if the test variation were to be reduced as a result of using the
new method. Do the data support the hypothesis that a reduction in variation is evident?
Ho: X 2 = Y 2
Ha: X 2 Y 2 (for a two-tailed test)
F=
SX 2 (0.071)2
=
= 3.49
SY 2 (0.038)2
H1317_CH21.qxd
10/17/07
2:33 PM
Page 216
H1317_CH22.qxd
10/15/07
1:54 PM
Page 217
Histograms
39.9
42.1
38.5
38.7
35.1
42.2
31.4
35.6
40.8
47.3
41.2
38.3
42.2
34.6
41.2
37.0
36.4
37.9
36.4
40.9
37.8
36.5
35.7
38.0
40.5
41.4
43.7
42.0
34.3
37.5
42.9
33.9
37.7
45.1
29.5
40.2
217
44.2
46.0
31.0
39.6
47.3
41.3
41.2
39.5
40.5
37.3
42.4
42.1
40.5
38.5
36.1
40.6
45.9
44.8
41.9
46.6
43.6
39.2
38.5
42.6
H1317_CH22.qxd
218
10/15/07
1:54 PM
Page 218
46.7
40.1
35.5
34.8
43.0
43.1
41.5
41.8
42.0
38.0
37.9
47.7
33.4
36.8
33.0
43.5
40.7
42.5
39.4
43.6
43.0
35.8
40.1
39.2
40.6
37.4
38.1
38.9
40.8
44.6
39.2
40.0
36.3
38.9
47.3
36.1
40.7
40.4
47.2
41.5
41.7
32.8
37.7
36.4
42.0
38.9
36.9
40.6
41.5
39.2
41.0
44.8
45.1
41.3
Step 2. Determine the number of class intervals or cells K into which the data will be
sorted.
The number of class intervals K depends on the total number of data values or observations N. Jurans Quality Handbook (1999, 44.1) gives the following table:
Number of observations N
2050
51100
101200
201500
5011000
> 1000
Juran suggests that the following equation be used to estimate the number of class
intervals. The resulting value should be rounded to the nearest integer. This relationship
will give values approximately equal to those given in Jurans handbook.
K = 1.5 ln N + 0.5
For our example, the number of observations N = 126 and the number of class intervals K is given by
K = 1.5(4.84) + 0.5,
K = 7.76 = 8.
Step 3. Determine the class interval width.
The class interval width is determined by dividing the difference in the largest and smallest observations by the number of class intervals.
Interval width =
47.7 29.5
= 2.2
8
Step 4. Calculate all class intervals and categorize the data into appropriate class intervals.
H1317_CH22.qxd
10/15/07
1:54 PM
Page 219
Histograms
219
There will be eight intervals, each with a class interval width of 2.00 units.
The beginning of the first interval will be at the smallest observation minus one-half
the interval width (1.00). The end of the first interval will be at the beginning point plus
the full interval width (2.00).
First interval bounds: 29.5 1.1 to 29.5 + 2.2 or 28.4 to 31.7.
Each of the following intervals is calculated by taking the ending point of the previous interval and adding the interval width, 2.2.
1st
2nd
3rd
4th
5th
6th
7th
8th
9th
These intervals will hold observations from the beginning point up to but not including the ending point. For example, the fifth interval will contain all of the observations
from 39.4 up to but not including 41.4. If we encounter an observation of 41.4, it will be
assigned to the sixth interval.
A table of the intervals and a tally count of all the observations is made, assigning each
observation to its appropriate interval:
Interval
28.4 to 31.7
31.7 to 33.9
33.9 to 36.1
36.1 to 38.3
38.3 to 40.5
40.5 to 42.7
42.7 to 44.9
44.9 to 47.1
47.1 to 49.3
H1317_CH22.qxd
10/15/07
220
1:54 PM
Page 220
40
35
Frequency
30
25
20
15
10
5
0
28.4
to
31.7
31.7
to
33.9
33.9
to
36.1
36.1
to
38.3
38.3
to
40.5
40.5
to
42.7
42.7
to
44.9
44.9
to
47.1
47.1
to
49.3
Observation
An alternative to the traditional frequency histogram is a technique called the leafstem plot. This method is a quick and relatively simple way to present data in a graphical
format. Start by determining the largest and smallest observations and record these to one
decimal place less than that of the data.
In the previous example, the smallest and largest values were 29.5 and 47.7. These values will be recorded as 29 and 47.
A vertical table is made beginning with 29 and ending with 47. A vertical line is drawn
to the right of the values.
A
29
30
31
32
33
34
35
36
37
38
39
B
29
30
31
32
33
34
35
36
37
38
39 1
C
29
30
31
32
33
34
35
36
37
38
39 1
H1317_CH22.qxd
10/15/07
1:54 PM
Page 221
Histograms
40
41
42 4
43
44
45
46
47
40
41
42 4
43
44
45
46
47
221
40
41
42 4
43 3
44
45
46
47
1. The first observation from the 126 values is 42.4. Find 42 in the table, then place a
4 to the right of the vertical line. This will represent the location of the observation 42.4.
2. The second observation from the 126 values is 39.1. Find 39 in the table, then place
a 1 to the right of the vertical line. This will represent the location of the observation 39.1.
3. The third observation from the 126 values is 43.3. Find 43 in the table, then place
a 3 to the right of the vertical line. This will represent the location of the observation 43.3.
Continue until all 126 observations have been posted. The resulting leaf-stem plot will
aid in visualizing the distribution of the observations.
Completed leaf-stem plot:
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
5
4
8
4
2
5
8
4
4
1
6
3
4
3
6
1
1
3
0
0
8
1
4
9
5
4
1
9
0
6
2
9
7
7
9
6
6
4
0
7
9
1
5
1
0
8
1
0
3
3
7
5
9
3
4
8
8
2
1
8
8
3
8
0
2
7
2
5
5
6
3 2
1
4
0
2
9
2
2
6
4
5
1
6
1
4
0
0
1
7
9
5
6
3
9
7
9
3
9
2
5
2
4
6
7
5
2
2
5
1
5 9
8 0 5 7 4 5 6 6
7 9 5 0 3
0 6
H1317_CH22.qxd
222
10/15/07
1:54 PM
Page 222
Optional problem:
Construct a leaf-stem plot for the following 76 weights of employees. Is this distribution
normally distributed? What could contribute to the shape of this distribution?
136
198
106
190
94
150
142
116
206
144
136
156
124
108
195
231
172
151
92
184
181
138
112
163
211
183
81
162
207
205
194
168
225
107
111
196
139
216
240
204
152
148
111
147
222
196
146
144
122
194
236
205
153
128
132
133
232
207
146
192
193
135
213
206
193
133
212
204
223
180
120
117
Bibliography
Betteley, G., N. Mettrick, E. Sweeney, and D. Wilson. 1994. Using Statistics in Industry.
New York: Prentice Hall.
Ishikawa, K. 1993. Guide to Quality Control. 2nd edition. White Plains, NY: Quality
Resources.
Juran, J. M., and A. B. Godfrey. 1999. Jurans Quality Handbook. 5th edition. New York:
McGraw-Hill.
Tague, N. R. 2004. The Quality Toolbox. 2nd edition. Milwaukee, WI: ASQ Quality Press.
H1317_CH23.qxd
10/15/07
4:28 PM
Page 223
Hypothesis Testing
Hypothesis tests are used to derive statistically sound judgments regarding evaluating sample data compared to population parameters or other external criteria. A hypothesis is a
systematic approach to assessing beliefs about reality; it is confronting a belief with evidence and then deciding, in light of this evidence, whether the initial belief (or hypothesis) can be maintained as acceptable or must be rejected as untenable.
The hypothesis for which we want to develop supporting evidence and that we want to prove is called
the alternative hypothesis, or Ha.
Formulate the appropriate null hypothesis and alternative hypothesis for testing that
the starting salary for graduates with a B.S. degree in electrical engineering is greater than
$38,000 per year.
223
H1317_CH23.qxd
224
10/15/07
4:28 PM
Page 224
Null hypothesis is
accepted
Null hypothesis is
rejected
The level of significance must be chosen for the hypothesis. This significance, or risk,
is the probability that we will reject the null hypothesis when, in reality, the null hypothesis is, in fact, true. Sometimes the significance is expressed as a level of confidence of the
hypothesis. Confidence is 1 . Typical levels of confidence are 0.99, 0.95, and 0.90.
H1317_CH23.qxd
10/15/07
4:28 PM
Page 225
Hypothesis Testing
225
The level of significance, one of the easily controlled requirements, is the percentage
of time over an extended period that the statistical test will make a type I error.
For this example, we will set the significance, or , risk at = 0.05.
The first step is to sample the population in order to determine a sample average X.
In this step of the hypothesis test, we select a test statistic, which is the type of statistic to be computed from a random sample taken from the population of interest. This test
statistic will be used to establish the probability of truth or falsity of the null hypothesis.
Hypothesis tests are used to test statistical inferences regarding the following:
X
,
X
Assume from the sample data that we find that X = $39,750 and S = $2816.
H1317_CH23.qxd
10/15/07
226
4:28 PM
Page 226
The general formula for computing a test statistic for making an inference is
test statistic =
where the observed sample statistic is the statistic of interest from the sample, tested value
is the hypothesized population parameter, and standard error is the standard deviation of
the sampling population divided by the square root of the sample size.
Hypothesis tests can be divided into several types depending on the following:
1. Large vs. small sample sizes (large 30 and small < 30)
2. Hypothesis test about a population mean
Hypothesis test about a population proportion
3. Hypothesis test about the difference between two population means, 1 2
Hypothesis test about the difference between two population proportions, 1 2
H1317_CH23.qxd
10/15/07
4:28 PM
Page 227
Hypothesis Testing
227
In the following example, we illustrate the hypothesis for a small-sample test about a
population mean.
Ho: = 0
Ha: 0
Test statistic:
Test statistic:
t=
X 0
s
n
t=
X 0
s
n
Reject Ho if
Reject Ho if
the average is X = 14.2 ounces and the sample standard deviation is S = 1.3. What is your
conclusion?
Step 1. Formulate two opposing hypotheses.
Ho: = 15
Ha: < 15
Step 2. Select a test statistic.
t=
X 0
s
n
H1317_CH23.qxd
228
10/15/07
4:28 PM
Page 228
n = 25
X = 14.2
S = 1.3
t=
14.2 15.0
= 3.08
1.3
25
Since the test statistic is less than 1.318, we will reject the null hypothesis and conclude that the true average weight is less than 15.0 ounces.
Note: It is assumed that the variances, while unknown, are equal. If the variances of
the two populations are assumed not to be equal, we modify the degrees of freedom when
looking up the critical t-score. The degrees of freedom for these cases are determined by
df =
S12 S22
n + n
1
2
2
S12
S22
n
n
1
+ 2
n1
n2
X 0
s
n
H1317_CH23.qxd
10/15/07
4:28 PM
Page 229
Hypothesis Testing
229
From the 15 data values, we find the average X = 25.8 and the standard deviation S = 4.70.
The resulting test statistic is
t=
X 0
25.8 22.0
, t=
, t = 3.14.
s
4.70
n
15
Since the test statistic is greater than the rejection region of t > 3.14, we will reject the
null hypothesis and conclude that the true average build time is not equal to 22.0 minutes.
Hypothesis Test about a Population Mean:
Large-Sample Case
a. One-tail test
b. Two-tail test
Ho: = 0
Ha: > 0
(or Ha: < 0)
Ho: = 0
Ha: 0
Test statistic:
Test statistic:
Z=
X 0
s
n
Z=
X 0
s
n
Reject Ho if
Reject Ho if
Average, X = 1.93
H1317_CH23.qxd
230
10/15/07
4:28 PM
Page 230
X X0
S
n
Z=
1.93 2.00
0.28
120
Z = 2.73
X u0
S
n
Z = 3.14
H1317_CH23.qxd
10/15/07
4:28 PM
Page 231
Hypothesis Testing
231
Ho: (1 2) = D0
Ho: (1 2) D0
Test statistic:
( X1 X 2 ) D0
( X1 X2 )
ZCalc =
( X1 X 2 ) D0
( X1 X2 )
where:
( X1 X2 ) =
12 22
+
n1 n2
Reject Ho if:
Reject Ho if:
CG = Carolina Giant
BR = Big Red
X 2 = 418
S1 = 23.5
S2 = 31.0
D0 = 45
H1317_CH23.qxd
232
10/15/07
4:28 PM
Page 232
( X1 X 2 ) D0 (470 418) 45
=
( X1 X2 )
( X1 X2 )
ZCalc =
7
(470 418) 45
=
= 2.15
3.25
3.25
12 22
+
n1 n2
( X1 X2 ) =
( X1 X2 ) =
(23.5)2 (31.0)2
+
= 3.25
1201
160
Ho: (1 2) = D0
Ha: (1 2) D0
Test statistic:
Test statistic:
t=
( X 1 X 2 ) D0
1
1
S p2 +
n1 n 2
t=
( X 1 X 2 ) D0
1
1
S p2 +
n1 n 2
Reject Ho if:
Reject Ho if:
t > t or (t < t)
where:
S p2 =
( n1 1) S12 + ( n 2 1) S 22
n1 + n 2 2
H1317_CH23.qxd
10/15/07
4:28 PM
Page 233
Hypothesis Testing
233
( X1 X 2) D0
S p2
1 1
n + n
1
2
We will sample the wear rate for 15 samples of brand A and 18 samples of brand B
and will calculate the average, standard deviation for each and determine the pooled variance Sp2.
n1 = 15
X 1 = 18.4
S1 = 3.4
n2 = 18
X 2 = 20.6
S2 = 3.8
S p2 =
S p2 = 13.14.
Using this pooled variance of 13.14, we calculate the test statistic.
t=
(X1 X2) D0
1
1
Sp2 +
n1 n2
t=
(18.4 20.6) 0
1
1
13.14 +
15 18
t = 1.74
H1317_CH23.qxd
234
10/15/07
4:28 PM
Page 234
Ho: = 0
Ha: 0
Test statistic:
Test statistic:
Z=
0
0 (1 0 )
n
Z=
0
0 (1 0 )
n
Reject Ho if:
Reject Ho if:
H1317_CH23.qxd
10/15/07
4:28 PM
Page 235
Hypothesis Testing
235
Z=
0
0 (1 0 )
n
.785 .750
.750 (1 .750 )
125
Z = 0.90
H1317_CH23.qxd
236
10/15/07
4:28 PM
Page 236
Ho: ( 2) = 0
Ha: ( 2) 0
P1 P2
( P1 P2 )
Reject Ho if:
Reject Ho if:
Z < Z or
Z < Z /2
Z > Z for
Z > Z /2
Ha: ( 2) > 0
*This test can be applied to differences other than zero.
Example 1:
Over the past decade it has been suspected that the percentage of adults living in the
southeastern region of the United States who eat grits (a gourmet delicacy indigenous to
this area) has declined. A sample survey by the Grain Research Institute of Technology
(GRIT) in 1986 found that from a random sample of 1500 adults, 525 ate grits on a regular basis. The same survey taken in 1996 with a sample of 1850 found that 685 adults
ate grits. Based on this data, has the consumption of grits increased or has it remained
the same?
Step 1. Formulate the opposing hypotheses.
Ho: ( 2) = 0
Ha: ( 2) > 0
1 = true proportion of population that ate grits in 1996
2 = true proportion of population that ate grits in 1986
H1317_CH23.qxd
10/15/07
4:28 PM
Page 237
Hypothesis Testing
237
P1 P2
( P1 P2 )
P1 =
(np)1
n1
528
= 0.352
1500
1 1
( P1 P2 ) = P(1 P) +
n1 n2
or
( P1 P2 ) =
p1q1 p2 q2
+
n1
n1
(np)1 + (np)2
685 + 528
= 0.362
P = combined proportion =
=
1850
+
+ 1500
n
n
1
2
1 1
( P1 P2 ) = P(1 P) +
n1 n2
1
1
( P1 P2 ) = (0.362)(0.638)
+
1850 1500
( P1 P2 ) = (0.362)(0.638)(0.0012)
( P1 P2 ) = 0.017
Test statistic:
Z=
P1 P2
( P1 P2 )
Z=
0.370 0.352
= 1.06
0.017
H1317_CH23.qxd
238
10/15/07
4:28 PM
Page 238
p1 p2
p1q1 p2 q2
+
n1
n2
Z=
0.075 0.050
= 1.30
(0.075 )(0.9925 ) (0.050 )(0.95 )
+
350
280
H1317_CH23.qxd
10/15/07
4:28 PM
Page 239
Hypothesis Testing
239
s 2 ( n 1)
2
The sample standard deviation S from the 30 sample observations is 3.2 minutes.
2 =
( 3.2 ) 2 ( 29 )
= 6.06
( 7) 2
We reject the null hypothesis that the true standard deviation is equal to 7 minutes and
conclude that the new system does, in fact, have a lower standard deviation.
H1317_CH23.qxd
240
10/15/07
4:28 PM
Page 240
Ho: 12 = 22
Ha: 12 < 22
Ha: 12 22
Test statistic:
S22
S12
(or FCalc =
S12
when Ha : 12 > 22 )
S22
FCalc =
FCalc =
S12
2
2
2 when S1 > S2
S2
or
S22
when S22 > S12
S12
Rejection criteria:
Reject Ho in favor of Ha if
Rejection criteria:
Reject Ho in favor of Ha if
1 = U of S
S12
S22
FCalc =
(18.5)2
= 2.09
(12.8)2
2 = MSC
H1317_CH23.qxd
10/15/07
4:28 PM
Page 241
Hypothesis Testing
241
n11 = 30
n21 = 19
F.10,30,19 = 1.76
Bibliography
Dixon, W. L., and F. J. Massey, Jr. 1969. Introduction to Statistical Analysis. 3rd edition.
New York: McGraw-Hill.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Petruccelli, J. D., B. Nandram, and M. Chen. 1999. Applied Statistics for Engineers and
Scientists. Upper Saddle River, NJ: Prentice Hall.
Sternstein, M. 1996. Statistics. Hauppauge, NY: Barrons Educational Series.
Walpole, R. E., and R. H. Myers. 1993. Probability and Statistics for Engineers and Scientists. 5th edition. Englewood Cliffs, NJ: Prentice Hall.
H1317_CH23.qxd
10/15/07
4:28 PM
Page 242
H1317_CH24.qxd
10/17/07
2:34 PM
Page 243
Individual-Median/Range
Control Chart
The selection of control charts for variables characteristics are traditionally the individual/
moving range, average/range, average/standard deviation, and the median/range. In all of
these cases, there is an assumption that the characteristic is originating from a single
population. In some cases, there is a macro contributing to a process that is influencing a
set of smaller microprocesses. An example of such a case would be a multi-cavity mold
operation. Each of the individual cavities would represent several microprocesses, and the
injection pressure common to all of the cavities would be the macroprocess. Another
example would be a filling line where several filling heads would represent the individual
microprocesses, and the master pump supplying all of the individual filling heads would
represent the macroprocess. If in the latter example a 12-head filling line were to be monitored using an average/range control chart where five samples were randomly chosen
every hour and the average weight was charted, there would be a 67 percent probability of
not including a specific filling head in the sample. Even with hourly sampling, production
could continue for a full day without sampling one of the filling heads.
If a faulty filling head were to develop a state of short filling that led to an out-ofcontrol condition, the usual response would be to adjust the primary pump. This action
would lead to an increased overall average causing all of the other positions of the filling
heads to be increased, when our original intention was to increase only the filling head in
question.
An optional method to monitor the overall process and address the performance of the
individual microprocesses is found in the application of the individual-median/range control chart. The larger macroprocess will be monitored by the movement of the individual
medians around the average median. The microprocesses will be monitored by the movement of the individuals around the average. These two charts will be constructed on a common chart. The process variation will be monitored using a traditional range control chart.
The steps for the construction of this chart, as with all variable control charts, involve:
1. Collection of historical data to characterize the current performance of the process
2. Calculation of a location and variation statistic
3. Calculation of statistical control limits for the location and variation statistics based
on an average 3 standard deviations
4. Construction of the control chart, including average, control limits, and plotting
historical data
5. Continue plotting future data, looking for signs of a process change
243
H1317_CH24.qxd
10/17/07
244
2:34 PM
Page 244
8
4
11
7
8
11
7
8
7
8 = median
Sample #1
3/6/95
AM
Sample #2
3/6/95
PM
Sample #3
3/7/95
AM
Sample #4
3/7/95
PM
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
42.0
44.7
46.1
43.8
43.6
43.8
4.1
46.0
45.4
40.1
44.5
45.6
45.4
5.9
45.8
44.9
43.7
41.5
42.4
43.7
4.3
43.9
46.2
45.2
43.2
40.9
43.9
5.3
H1317_CH24.qxd
10/17/07
2:34 PM
Page 245
245
Sample #5
3/8/95
AM
Sample #6
3/8/95
PM
Sample #7
3/9/95
AM
Sample #8
3/9/95
PM
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
40.0
42.7
44.5
43.3
46.2
43.3
6.2
45.7
43.0
45.4
45.5
42.2
45.4
3.5
Sample #9
3/10/95
AM
Sample #10
3/10/95
PM
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
45.2
45.7
46.0
43.0
42.5
45.2
3.5
43.0
42.8
45.0
42.0
42.9
42.9
3.0
46.9
44.8
43.4
43.2
41.5
43.4
5.4
43.1
46.5
44.4
43.8
45.6
44.4
3.4
= 44.14.
X
The variation statistic will be the average range R and is determined by averaging the
10 ranges:
4.1 + 5.9 + 4.3 + 5.3 + 6.2 + 3.5 + 3.0 + 5.4 + 3.5 + 3.4
,
10
R = 4.46.
R=
Step 3. Determine the statistical control limits for the location and variation statistic.
H1317_CH24.qxd
246
10/17/07
2:34 PM
Page 246
2R
UCLx /LCLx = X
= average median
where: X
2 = factor dependent on the subgroup sample size n
R = average range
The value for the factor 2 for this example, where n = 5 is 0.691, is found in
Table 1.
E2R,
UCLXi/LCLXi = X
= average median
where: X
E2 = factor dependent on the subgroup sample size
R = average range.
= 44.14, R = 4.46, and E2 = 1.290, the upper and lower conFor our example, where X
trol limits for individuals are
E2R
UCLXi /LCLXi = X
UCLR = D4R
H1317_CH24.qxd
10/17/07
2:34 PM
Page 247
247
Subgroup n
2 (medians)
E2 (individuals)
D3 (LCL range)
D4 (UCL range)
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
1.880
1.187
0.796
0.691
0.548
0.508
0.433
0.412
0.362
0.350
0.316
0.308
0.282
0.276
0.255
0.251
0.234
0.231
0.218
0.215
0.203
0.201
0.191
0.189
2.660
1.772
1.457
1.290
1.184
1.109
1.054
1.010
0.975
0.945
0.921
0.899
0.811
0.864
0.849
0.836
0.824
0.813
0.803
0.794
0.786
0.778
0.770
0.763
0.076
0.136
0.184
0.223
0.256
0.283
0.307
0.328
0.347
0.363
0.378
0.391
0.403
0.415
0.425
0.434
0.443
0.451
0.459
3.267
2.574
2.282
2.114
2.004
1.924
1.864
1.816
1.777
1.744
1.717
1.693
1.672
1.653
1.637
1.622
1.608
1.597
1.585
1.575
1.566
1.557
1.548
1.541
= 44.14
Average median, X
H1317_CH24.qxd
10/17/07
248
2:34 PM
Page 248
Step 4. Construct the control chart, sketch in control limits and averages, and plot historical data.
Control limits are traditionally drawn with a broken line (----------------------), and averages
are drawn with a solid line (). For subgroup sample sizes less than
seven, the average line for the range is not drawn.
Since we are plotting two location statistics (median and individuals) on one chart, we
will use broken lines for the control limits for the individuals and a short-long dash for the
median control limits.
Notes on Plotting Data
All individual points should be identified as a single dot (.), and the particular individual
that is identified as the median will be represented by an x. The median values will be
connected with a line. Any median values falling outside the median limits represent a rule
violation, and any individual values (including the median) that fall outside the individual
control limits also represent a rule violation. If an individual value falls outside its control
limit, it should be identified relative to its origin (such as head #1, head #2, and so on) by
placing a number next to the dot.
The completed chart showing the first 10 sets of data follows:
History
55
UCL XM = 47.22
50
UCL Xi = 49.89
45
x x
x x
40
LCL Xi = 38.39
35
LCL XM = 41.06
30
1
10
x = median
Note that all of the medians and individuals are within their respective control limits.
This process is statistically stable and well behaved.
Step 5. Continue to monitor the process, looking for evidence of a process change.
Signs of a process change will be indicated when one of the SPC rules is violated. These
rules are numerous, and most have the following two common characteristics:
H1317_CH24.qxd
10/17/07
2:34 PM
Page 249
249
1. They are patterns that are statistically rare to occur with a normal distribution
2. They indicate a direction in which the process has changed, either increased or
decreased
The three more frequently used SPC detection rules are:
1. A single observation outside the upper or lower control limit
2. Seven consecutive points below the average or above the average
3. A trend of seven consecutive points all steadily increasing or decreasing in value
These rules will be applied both to the movement of the individuals with respect to
their control limits and to the median with respect to its control limit. The only rule applicable to monitoring the range will be rule #1, when the sample size is less than seven.
Data for the next seven samples are provided below. Plot the data and determine if
there is any evidence of a process change.
Sample #11
Sample #12
Sample #13
Sample #14
3/11/95
AM
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
3/11/95
PM
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
3/12/95
AM
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
3/12/95
PM
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
42.1
44.8
44.7
43.3
42.5
43.3
2.7
43.4
40.4
43.3
45.1
40.7
43.3
4.7
46.5
44.0
42.4
43.5
41.2
43.5
5.3
Sample #15
Sample #16
Sample #17
3/13/95
AM
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
3/13/95
PM
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
3/14/95
AM
Head 1
Head 2
Head 3
Head 4
Head 5
Median
Range
46.0
40.6
45.0
41.8
44.4
44.4
5.4
43.0
44.4
40.6
43.6
42.4
43.0
3.8
45.1
40.0
38.0
42.5
43.3
42.5
7.1
42.0
43.0
37.6
45.0
40.0
42.0
7.4
The completed control chart follows. Note that there are two incidences where head
#3 had individual values that were below the LCL for individuals (38.39). This indicates
that this individual fill head is delivering less than would be expected based on the historical characterization.
Control limits for the range are calculated in the traditional manner for Shewhart
charts. The UCL for the range is UCL = D4 R , and the LCL for the range is LCL = D3 R .
For this example, the LCL is undefined as n = 5, and the UCL is calculated to be
UCL = 9.43.
H1317_CH24.qxd
10/17/07
250
2:34 PM
Page 250
History
55
UCL XM = 47.22
50
UCL Xi = 49.89
45
x x
x x x
x x
40
(3)
LCL Xi = 38.39
(3)
35
LCL XM = 41.06
30
1
10
Individual-median chart
15
20
x = median
= individual
10
UCL = 9.43
0.0
1
10
15
20
Range chart
Sample #14 and sample #17 have individuals from filling head #3 falling below the
LCL for individuals, while the medians for these samples remain in control. This supports
the conclusion that filling head #3 is not performing as was established for the historical
data. Perhaps the filling nozzle is blocked slightly, causing a restriction in the filling rate.
In any case, the output has been reduced enough to cause a process change in this position.
H1317_CH25.qxd
10/15/07
2:12 PM
Page 251
Individual/Moving Range
Control Chart
Control charts are used to detect changes in processes. When processes are monitored
using variables data and the opportunity for obtaining data is relatively infrequent, we may
elect to use an individual/moving range control chart. An example of such a case would be
measurements of a characteristic of a batch process where individual batches are made
once every four hours.
The ability of a control chart to detect a change in the process is related to:
1. The subgroup sample size
2. The amount of change that has occurred
The individual/moving range control chart has a subgroup sample size of n = 1 and is
the weakest of all Shewhart control charts with respect to its ability to detect a process
change.
The general principles and use of control charts are universal and are based on the following elements:
1.
2.
3.
4.
5.
H1317_CH25.qxd
10/15/07
252
2:12 PM
Page 252
The best single value to express the location of the performance of this process at a
specific time is the data point taken at that time. The best indication of the variation of the
process at a given point in time is the range between two successive values around that
point in time. The range between successive points is defined as the moving range and is
calculated by taking the difference between the current value and the previous value.
Moving ranges are always positive in value. The first data point collected will not have
a moving range, because it does not have a previous data point associated with it. There
will always be one less moving range than there are data points.
At the time we collect our initial data, we will also calculate and record the moving
ranges. The first moving range is 260 196 = 64, the second moving range is 260 187
= 73, and the last moving range is 230 210 = 20.
Date
Time
Admission time
Moving range
11/2/94
11/2/94
11/2/94
11/3/94
11/3/94
11/3/94
11/4/94
11/4/94
11/4/94
11/5/94
11/5/94
11/5/94
11/6/94
11/6/94
11/6/94
8:45A
1:12P
3:48P
9:04A
1:00P
3:55P
9:00A
12:55P
4:10P
8:53A
1:07P
3:58P
9:11A
12:55P
3:57P
196
260
187
170
296
171
320
139
196
228
176
125
226
210
230
64
73
17
126
125
149
181
57
32
52
51
101
16
20
H1317_CH25.qxd
10/15/07
2:12 PM
Page 253
253
The variation statistic we will use is the average moving range. It is calculated by averaging the moving ranges.
Average moving range MR =
64 + 73 + 17 + 126 + 20 1064
=
= 76.0
14
14
Step 3. Determine control limits for the location and variation statistics.
The average of the individuals represents the best overall location statistic for the process.
There is variation of the individual observations relative to the average. Control limits will
be determined for the individual values that make up the average. These limits will define
the upper and lower values we expect to encounter approximately 99.7 percent of the time
and will be based on the average 3 standard deviations.
Our original variation statistic was the moving range, which was determined for each
successive pair of data points. Using these moving ranges, we calculated an average mov
ing range MR. An estimate of three standard deviations can be made from the relationship
3S = 2.66MR ,
3S = 2.66 76.0 = 202.2.
The upper and lower control limits for the individual values are determined by adding
and subtracting three standard deviations to the average.
label appropriately X = 208.6, LCL = 6.4, and UCL = 410.8. Draw the average line using
a solid line, and draw the control limits using a broken line. Repeat this for the UCL for
H1317_CH25.qxd
254
10/15/07
2:12 PM
Page 254
the moving range. It is not necessary to show the average of the moving range. Label the
UCL for the moving range as UCL = 248.3.
Plot the individual and moving range values, and connect the points as they are
located. Draw a vertical wavy line between sample #15 and sample #16. This line will
serve to separate the historical data used to characterize the process and establish control
limits from values plotted from the future. In the area above the UCL for the individuals
in the vicinity of sample #6 through sample #9, record the word History to identify those
data points used for historical characterization.
An individual/moving range chart (historical data only) follows:
300
UCL = 248.3
200
100
0
1
10
15
Moving range
UCL = 410.8
400
History
300
Avg. = 208.6
200
100
LCL = 6.4
0
1
10
15
Individuals
H1317_CH25.qxd
10/15/07
2:12 PM
Page 255
255
Rule Two:
Rule Three: A lack of control is indicated when seven consecutive points are
steadily increasing or decreasing in value (trend detection).
Rule Four:
Two out of three consecutive points are greater than the two
standard deviation limits with the third falling anywhere. No points
fall outside the three standard deviation limits.
Rule Five:
Four out of five points are greater than one standard deviation, and
the fifth point falls anywhere. No points fall outside the three
standard deviation limits.
There are other patterns representing nonstable situations, some of which may not
indicate a direction of change. The five patterns described by the aforementioned rules
are the most frequent used to detect process changes. Only rule one through rule three utilize the 3 standard deviation limits and are the rules traditionally used in an operational
environment.
Only rule one should be applied to the moving range portion of the control chart for
the individual/moving range. This is because the distribution of the moving range is not
normal and is not symmetrical with respect to the average moving range. For a more
detailed discussion of control chart patterns, see Statistical Quality Control Handbook
(Western Electric Company 1958, 149183).
H1317_CH25.qxd
256
10/15/07
2:12 PM
Page 256
Based on rule one, rule two, and rule three, plot the following additional data points,
and determine if there is evidence of a process change. Has the process average changed?
Has the process variation changed?
Date
Time
Admission time
11/7/94
11/7/94
11/7/94
11/8/94
11/8/94
11/8/94
11/9/94
11/9/94
11/9/94
11/10/94
11/10/94
11/10/94
11/11/94
11/11/94
11/11/94
11/12/94
11/12/94
11/12/94
8:56A
12:40P
3:50P
9:00A
1:10P
4:05P
9:05P
12:58P
3:50P
8:30A
1:10P
4:00P
8:50A
1:11A
3:51P
9:11A
1:00A
4:00P
115
175
250
215
330
220
280
215
305
275
125
150
250
285
190
105
330
250
After completing your chart, compare with the following finished chart.
UCL = 410.8
400
History
300
Avg. = 208.6
200
100
LCL = 6.4
0
1
10
15
20
Individuals
25
30
H1317_CH25.qxd
10/15/07
2:12 PM
Page 257
257
300
UCL = 248.3
200
100
0
1
10
15
20
Moving range
25
30
The chart indicates the possibility of an increase in the average with no change in the
variation. Eight consecutive points are seen, starting with sample #18 (11/7/94 3:50 PM).
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Hayes, G. E., and H. G. Romig. 1982. Modern Quality Control. 3rd edition. Encino, CA:
Glencoe.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Western Electric Company. 1958. Statistical Quality Control Handbook. 2nd edition. New
York: Western Electric Company.
Wheeler, D. J., and D. S. Chambers. 1992. Understanding Statistical Process Control. 2nd
edition. Knoxville, TN: SPC Press.
H1317_CH25.qxd
10/15/07
2:12 PM
Page 258
H1317_CH26.qxd
10/15/07
4:33 PM
Page 259
Measurement Error
Assessment
All of the conclusions reached when using data obtained from observations depend on the
accuracy and validity of the data. We assume that the information fairly represents the
truth. Measurement data can be derived from either pure attribute inputs such as visual
standards or from direct measured variables such as gages. The latter type of measurement
error will be discussed first.
259
H1317_CH26.qxd
260
10/15/07
4:33 PM
Page 260
0.50011
Mary
0.49995
Jim
0.50008
Mark
0.50000
We will assume the lack of agreement is associated not with the measuring device (this
has already been acknowledged via the repeatability, or EV ) but rather with the different
techniques in using the measuring device. The lack of agreement is due to the reproducibility error or appraiser variation (AV ).
The amount of variation for each of these contributing factors can be determined by
simply calculating the standard deviation of each set of data for the repeatability and
reproducibility:
Standard deviation for Bobs five measurements: 0.0000482
Standard deviation for four measurements from all operators: 0.0000733
In most formal studies to assess measurement, since there is a relatively small number
of operators who typically measure a given part, the range of measurements is used to estimate the standard deviation. From statistical process control (SPC), the relationship
=
R
,
d 2*
may be used to estimate the standard deviation of measurement error. In this relationship,
R is the average range of repeated measures, and d2 is a constant that is dependent on the
number of observations to calculate the range. In pure SPC calculations, the actual number of samples from which the average range is determined is assumed to be very large
(n > 25) for most control chart applications. For this reason, the d2 factors must be modified to compensate for the small number of samples. These modified factors are abbreviated d2*. These modified values depend on both the size of the sample from which the
range is determined and on the number of samples used to determine the average.
The values for d2* can be found in Table 1.
H1317_CH26.qxd
10/15/07
4:33 PM
Page 261
261
Size of sample m
Number of samples g
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
>15
10
11
1.41
1.28
1.23
1.21
1.19
1.18
1.17
1.17
1.16
1.16
1.16
1.15
1.15
1.15
1.15
1.128
1.91
1.81
1.77
1.75
1.74
1.73
1.73
1.72
1.72
1.72
1.71
1.71
1.71
1.71
1.71
1.693
2.24
2.15
2.12
2.11
2.10
2.09
2.09
2.08
2.08
2.08
2.08
2.07
2.07
2.07
2.07
2.059
2.48
2.40
2.38
2.37
2.36
2.35
2.35
2.35
2.34
2.34
2.34
2.34
2.34
2.34
2.34
2.326
2.67
2.60
2.58
2.57
2.56
2.56
2.55
2.55
2.55
2.55
2.55
2.55
2.55
2.54
2.54
2.534
2.83
2.77
2.75
2.74
2.73
2.73
2.72
2.72
2.72
2.72
2.72
2.72
2.71
2.71
2.71
2.704
2.96
2.91
2.89
2.88
2.87
2.87
2.87
2.87
2.86
2.86
2.86
2.85
2.85
2.85
2.85
2.847
3.08
3.02
3.01
3.00
2.99
2.99
2.99
2.98
2.98
2.98
2.98
2.98
2.98
2.98
2.98
2.970
3.18
3.13
3.11
3.10
3.10
3.10
3.10
3.09
3.09
3.09
3.09
3.09
3.09
3.08
3.08
3.078
3.27
3.22
3.21
3.20
3.19
3.19
3.19
3.19
3.18
3.18
3.18
3.18
3.18
3.18
3.18
3.173
12
13
3.35 3.42
3.30 3.38
3.29 3.37
3.28 3.36
3.28 3.35
3.27 3.35
3.27 3.35
3.27 3.35
3.27 3.35
3.27 3.34
3.27 3.34
3.27 3.34
3.27 3.34
3.27 3.34
3.26 3.34
3.258 3.336
14
15
3.49
3.45
3.43
3.43
3.42
3.42
3.42
3.42
3.42
3.42
3.41
3.41
3.41
3.41
3.41
3.407
3.55
3.51
3.50
3.49
3.49
3.49
3.48
3.48
3.48
3.48
3.48
3.48
3.48
3.48
3.48
3.472
Operator B
Part #
Trial 1
Trial 2
Range
Trial 1
Trial 2
Range
1
2
3
4
5
0.889
0.855
0.868
0.888
0.867
0.886
0.859
0.870
0.890
0.870
0.003
0.004
0.002
0.002
0.003
0.888
0.852
0.866
0.885
0.862
0.886
0.851
0.866
0.884
0.863
0.002
0.001
0.000
0.001
0.001
R
,
d 2*
where the average range R is the grand average range of the two operators ranges for the
duplicate measurements for each of the five sample parts.
H1317_CH26.qxd
262
10/15/07
4:33 PM
Page 262
0.0028 + 0.0010
= 0.0019.
2
This average range is the same as if we had averaged all 10 sets of ranges. Each operator measured the samples two times. In terms of SPC, the subgroup sample size is two in
terms of measurement error, m = 2 (m = the number of trials). The total number of samples measured g is determined by multiplying the number of parts (five) by the number of
operators participating in the study (two). In this case, g = 2 5 = 10.
Looking up the appropriate d2* in Table 1, where m = 2 and g = 10, we find d2* = 1.16.
=
R1
d2*
0.0019
= 0.0016
1.16
This estimate is for one standard deviation for the repeatability error and should be
written as EV. Traditionally, the AV or repeatability is reported as 5.15 to reflect a 99
percent level of confidence.
EV = 5.15 0.0016
EV = 0.0082
By dividing the EV by two and rewriting, we may also express the EV as 0.0041.
Any measurement we make with the system under evaluation is subject to a measurement of 0.0041 99 percent of the time due to equipment error.
Example of reproducibility (AV ) error determination:
In the example illustrating the concept of repeatability, each operator measured the same
five parts twice. If there were no measurement error due to the actual measuring device
(EV = 0), then the difference between the grand average for all ten observations (5 parts 2
measurements each = 10 measurements) would be equal. Any differences in the average
of operator A and operator B would be attributed to differences in technique. This error
factor is referred to as reproducibility error or AV.
The AV standard deviation is calculated similar to that of the EV. In this case, we are
using a range of one sample of two observations. The appropriate d2* is obtained from
H1317_CH26.qxd
10/15/07
4:33 PM
Page 263
263
Table 1, and it is dependent on the number of operators (m = 2) and the number of samples (g = 1) since there is only one range calculation.
d2* = 1.41
Each operators average is determined by averaging all 10 samples. The range is determined by subtracting the lowest from the highest.
Average operator A = 0.8742
Average operator B = 0.8704
Difference R2 = 0.0038
Estimate of one standard deviation for appraiser variation:
=
R2
d2*
0.0038
= 0.0027
1.41
AV = 0.0027
Unadjusted AV = 5.15 AV
Unadjusted AV = 0.014
This unadjusted AV is contaminated by the variation due to the measuring device (EV)
and must be adjusted by subtracting a proportion of the EV from it. Since standard deviations cannot be added or subtracted, we convert to the variance by squaring the standard
deviation, performing operations, and then taking the square root to convert back to standard deviations.
Corrected AV compensated for EV:
2
5.15 R2 (5.15 EV )2
AV =
*
nt
d2
where: R2 = range for average values of operator measurements
R2 = 0.0038
n = number of parts used in evaluation
n=5
t = number of times (trials) each operator measures the part
t=2
EV = 0.0016
2
(5.15)(0.0038) [(5.15)(0.0016)]2
AV =
1.41
(5)(2)
AV = 0.000193 0.0000068
AV = 0.0136 or 0.0068
H1317_CH26.qxd
264
10/15/07
4:33 PM
Page 264
R&R
100,
T.T
%GR&R=
0.016
1000,
0.060
%GR&R=26.7%.
As a result of measurement error, 26.7 percent of the total tolerance is consumed. As
a rule of thumb, the following guidelines may be used:
%GR&R
%GR&R
%GR&R
H1317_CH26.qxd
10/15/07
4:33 PM
Page 265
265
0.530, then the conclusion would be to accept the part as meeting the specification requirement. With this information, we can calculate the probability that the part exceeds the
upper specification and that measurement error has led us to falsely accept the part when
we should, in fact, reject the part. This probability can be determined using the normal distribution, the standard deviation of measurement, and a point of interest as the upper specification limit.
Given an upper specification of 0.530, a standard deviation of measurement error of
0.0032, and an observed measurement of 0.529, what is the probability that the true measurement is greater than 0.530?
.470
Lower
specification
.500
Nominal
ZU =
.529
.530
Upper
specification
0.530 0.529
= 0.31
0.0032
A Z-score of 0.31 corresponds to a probability of exceeding 0.530, the upper specification of 0.3783. There is a 37.8 percent chance that the part actually exceeds the upper
specification limit of 0.530; this is approximately a one-in-three chance.
Supplemental problem:
An R&R of 0.355 is reported for a measuring system. The lower specification is 12.0. An
inspected unit is deemed out of specification with an observed measurement of 11.92.
What is the probability that the part actually meets the specification requirement?
Confidence Interval for Repeatability, Reproducibility,
and Combined R&R
Confidence intervals for the AV, EV, and R&R can be calculated similarly to those of the
standard deviation. Recall that the confidence interval for the standard deviation is determined by
(n 1)s 2
(n 1)s 2
<
<
,
X 2 / 2
X2
1
H1317_CH26.qxd
10/15/07
266
4:33 PM
Page 266
( EV )2
2 / 2
v1
UCL =
( EV )2
21 / 2
v1
Size of sample m
Number of samples g
2
1 1.0
2 1.9
3 2.8
4 3.7
5 4.6
6 5.5
7 6.4
8 7.2
9 8.1
10 9.0
11 9.9
12 10.8
13 11.6
14 12.5
15 13.4
>15
c.d. 0.88
2.0
3.8
5.7
7.5
9.3
11.4
12.9
14.8
16.6
18.4
20.2
22.0
23.9
25.7
27.5
2.9
5.7
8.4
11.2
13.9
16.6
19.4
22.1
24.8
27.6
30.3
33.0
35.7
38.5
41.2
3.8
7.5
11.1
14.7
18.4
22.0
25.6
29.3
32.9
36.5
40.1
43.7
47.4
51.0
54.6
6
4.7
9.2
13.6
18.1
22.6
27.1
31.5
36.0
40.5
44.9
49.4
53.9
58.4
62.8
67.3
7
5.5
10.8
16.0
21.3
26.6
31.8
37.1
42.4
47.7
52.9
58.2
63.5
68.8
74.0
79.3
10
11
6.3
7.0
7.7
8.4
12.3 13.8 15.1 16.5
18.3 20.5 22.6 24.6
24.4 27.3 30.1 32.7
30.4 34.0 37.5 40.9
36.4 40.8 45.0 49.0
42.5 47.5 52.4 57.1
48.5 54.3 59.9 65.2
54.5 61.0 67.3 73.3
60.6 67.8 74.8 81.5
66.6 74.6 82.3 89.6
72.7 81.3 89.7 97.7
78.7 88.1 97.2 105.8
84.7 94.8 104.6 113.9
90.8 101.6 112.1 122.1
6.76
7.45
8.12
12
9.0
17.8
26.5
35.3
44.1
52.8
61.6
70.3
79.1
87.9
96.6
105.4
114.1
122.9
131.7
8.76
13
9.6
19.0
28.4
37.8
47.1
56.5
65.9
75.3
84.6
94.0
103.4
112.8
122.2
131.5
140.9
9.38
14
10.2
20.2
30.2
40.1
50.1
60.1
70.0
80.0
90.0
99.9
109.9
119.9
129.9
139.8
149.8
15
10.8
21.3
31.9
42.4
52.9
63.5
74.0
84.6
95.1
105.6
116.2
126.7
137.3
147.8
158.3
9.97 10.54
H1317_CH26.qxd
10/15/07
4:33 PM
Page 267
267
Example calculation:
Calculate the 90 percent confidence interval for EV given the following (from the previous example):
Part I. Lower confidence interval limit:
EV = 0.0082
Number of samples = 5 parts
Number of operators = 2
Total measurements taken, g = 10 (Parts Operators = 5 2 = 10)
Number of trials, m = 2
Level of confidence = 90%
We get v1 from Table 2. For the EV calculation, m = 2 and g = 10.
v1 = 9.0
2/2 for 9.0 degrees of freedom and /2 = 0.05
2/2 = 16.92
LCL =
( EV )2
2 / 2
v1
LCL =
(0.0082)2
16.92
9
LCL = 0.0059
H1317_CH26.qxd
268
10/15/07
4:33 PM
Page 268
UCL =
(EV )2
21 / 2
v1
UCL =
(0.0082)2
3.33
9
UCL = 0.0135
Summary
The 90 percent confidence interval for the EV is 0.0059 to 0.0135.
An appropriate statement would be: We do not know the true repeatability error (EV),
but we are 90 percent confident it is between 0.0059 and 0.0135.
Confidence Interval for Reproducibility, AV
Lower confidence limit
LCL =
( AV )
2 / 2
v2
UCL =
( AV )2
21 / 2
v2
Calculate the 90 percent confidence interval for EV given the following (from the previous example):
The calculations for the AV are done exactly as for the EV except that the degrees of
freedom are changed. With this example, the number of trials is two since we are dealing
with a range of two averages, operator A and operator B, and we have only one such sample range. For this case, m = 2 and g = 1.
Part I. Lower confidence interval limit:
AV = 0.0136
Number of samples, m = 2 (each operator has an average, and there are two
operators)
Total measurements taken, g = 1
Level of confidence = 90%
We get v2 from Table 2. For the AV calculation, m = 2 and g = 1.
v2 = 1.00
2/2 for 1.00 degrees of freedom and /2 = 0.05
2/2 = 3.84
H1317_CH26.qxd
10/15/07
4:33 PM
Page 269
LCL=
(AV )2
2 / 2
v2
LCL=
(0.0136)2
3.84
1
269
LCL=0.0069
UCL =
UCL =
(0.0136)2
0.0039
1
UCL = 0.2178
Summary
The 90 percent confidence interval for the AV is 0.0069 to 0.2178.
An appropriate statement would be: We do not know the true repeatability error (EV),
but we are 90 percent confident it is between 0.0069 and 0.2178.
Confidence Interval for Combined Repeatability
and Reproducibility, R&R
Lower confidence limit
LCL=
(R&R)2
2 / 2
v1 + v2
(R&R)2
21 / 2
v1 + v2
(R&R)2
2 / 2
v1 + v2
LCL =
(0.016)2
18.3
10.0
LCL = 0.0118
H1317_CH26.qxd
270
10/15/07
4:33 PM
Page 270
(R&R)2
21 / 2
v1 + v2
UCL =
(0.016)2
3.94
10
UCL = 0.0254
Summary
The 90 percent confidence interval for the R&R is 0.0118 to 0.0254.
An appropriate statement would be: We do not know the true R&R, but we are 90
percent confident that it is between 0.0118 and 0.0254.
In the event that the measurement R&R is excessive, a short-term fix could be realized
by reporting an average of several observations. The amount of error would decrease proportionally to the square root of the sample size. This is a direct result of the central limit
theorem. Reporting an average of n = 4 will cut the measurement error in half relative to
the error expected by reporting a single observation.
H1317_CH26.qxd
10/15/07
4:33 PM
Page 271
271
of the defect, and the nature of the defect. The units found to be conforming are screened
by the check inspector. The defective units missed by the participating inspector but found
by the check inspector are tallied as B. The items found defective by the participating
inspector are tallied as D. These units are then reinspected by the check inspector. Any
items in the defective group D that are judged acceptable by the check inspector are tallied as K. The inspection efficiency or accuracy is determined by
% inspection accuracy or % of defects correctly identified =
DK
.
DK+B
Before inspection
Test lot (contains
250 good and 40 defective
units)
Inspection efficiency =
35 4
100 = 77.8%
35 4 + 9
D = 35
Defects found to be OK
by check inspector
K=4
True
defects = 31
H1317_CH26.qxd
272
10/15/07
4:33 PM
Page 272
Observed characteristic
MAD
In this case, the inspected unit would be acceptable (obvious in this example). As the
characteristic becomes closer and closer to the MAD, the more likely the call will be
incorrect.
In order to evaluate the level of effectiveness for a sensory inspection, it is suggested
that a test group of approximately 40 samples be made, 50 percent of which pass the
requirement. In our case, spots that are less than the MAD. Of the 50 percent that are
less than the MAD, half of these are marginally less than the MAD and half are definitely less than the MAD. The terms marginally and definitely are subjective. A committee will have the responsibility of determining which samples are good, which ones
are definitely good, and which ones are good but to a lesser degree than those that are
definitely good.
Of the remaining samples that are nonconforming, half of them are marginally bad and
half of them are definitely bad. It is important that 50 percent of the test group be good and
50 percent be bad (according to the committee).
The 40 samples are randomly numbered 140. A record is kept regarding the truth
of the status of the numbered samples.
H1317_CH26.qxd
10/15/07
4:33 PM
Page 273
MAD =
(0.310)
273
Sample
1
(0.314)
(0.280)
(0.320)
40
(0.275)
Sample
Date
Trial 1
Trial 2
Trial 3
Actual
GG =
GB =
3
4
5
6
7
8
9
10
11
12
BG =
BB =
H1317_CH26.qxd
274
10/15/07
4:33 PM
Page 274
The samples are presented to the participant one at a time in a random order, and his
or her decision is recorded on the worksheet. This process continues until all 40 samples
have been inspected. For this example case, only 12 samples are used. Inspection
responses that are classified as bad by the participant are recorded as a and those that
are good are recorded as a +. The samples are then presented to the inspector in a random order for a first, second, and third trial inspection. The column Actual shows the
true condition of the sample as determined by the committee.
Data for the attribute study:
Inspector
MC
Date 11/2/12
Sample
Trial 1
Trial 2
Trial 3
Actual
GG =
15
GB =
BG =
4
5
BB =
14
10
11
12
H1317_CH26.qxd
10/15/07
4:33 PM
Page 275
275
There are five values or indices related to the overall effectiveness of the inspector:
1.
2.
3.
4.
5.
Correct = C
Consumers risk = Cr
Manufacturers risk = Mr
Repeatability = R
Bias = B
1. Correct, C
The proportion of the time the inspector got it right.
GG + BB
Total Inspected
15 + 14
C=
= 0.81
36
C=
Inspector
MC
Date 11/2/12
Sample
Trial 1
Trial 2
Trial 3
Actual
GG =
15
GB =
BG =
4
5
BB =
14
10
11
12
H1317_CH26.qxd
276
10/15/07
4:33 PM
Page 276
2. Consumers Risk, Cr
As consumers, we are most interested in the probability of accepting a bad product. Cr is
the probability of accepting a part we believe to be good that in reality is bad.
4
BG
=
= 0.22
BB + BG 14 + 4
Cr =
Inspector
MC
Date 11/2/12
Sample
Trial 1
Trial 2
Trial 3
Actual
GG =
15
GB =
BG =
4
5
BB =
14
10
11
12
3. % Manufacturers Risk, Mr
The last thing a manufacturer wants to do is to call a part good when in reality it is bad.
The manufacturers risk (false alarm) is determined by
Mr =
3
GB
=
= 0.17
GG + GB 15 + 3
H1317_CH26.qxd
10/15/07
4:33 PM
Page 277
Inspector
MC
277
Date 11/2/12
Sample
Trial 1
Trial 2
Trial 3
Actual
GG =
15
GB =
BG =
4
5
BB =
14
10
11
12
4. Repeatability, R
Repeatability is a measure of how well the participant and the inspector get the same
results when inspecting the same part. It is a measure of consistency. We are not concerned
with getting the correct answer, only how constant we are in our within sample inspection. Each sample has been inspected three times; therefore, there are three opportunities
per sample to have agreement. These opportunities are agreement between trial #1 and #2,
trial #2 and #3, and trial #1 and #3. For each sample, there are three opportunities for
agreement, and there are 12 samples. The total opportunities are 36, and the number of
agreements is 26.
R=
26
= 0.72
36
H1317_CH26.qxd
10/15/07
278
4:33 PM
Page 278
Inspector
MC
Date 11/2/12
Sample
Trial 1
Trial 2
Trial 3
Actual
GG =
15
GB =
BG =
4
5
BB =
14
10
11
12
5. Bias, B
Bias is the ratio of the Mr and Cr. The Mr and Cr must be converted using a bias factor
(BF). The bias factors are listed in Table A.9 in the appendix.
Mr = 0.17
Bf Mr = 0.2541
Cr = 0.22
BfCr = 0.2966
B=
Bf Mr 0.2541
=
= 0.86
BfCr 0.2966
Summary of indices:
Index
1.
2.
3.
4.
5.
Correct, C
Consumers risk, Cr
Manufacturers risk, Mr
Repeatability, R
Bias, B
Value
0.81
0.22
0.17
0.72
0.86
H1317_CH26.qxd
10/15/07
4:33 PM
Page 279
279
Acceptable
0.90
0.02
0.05
0.90
0.80 to 1.20
Marginal
0.80
0.02
0.05
0.80
0.50
or
1.20
to
to
to
to
to
0.90
0.05
0.10
0.90
0.80
to 1.50
Unacceptable
0.80
0.05
0.10
0.80
0.50
or
1.50
Interpretation of the results must be done with caution. If all the test samples were
extremely obvious, then all the inspectors would have achieved a l00 percent, 0 percent,
0 percent, and 100 percent on the four scores. The scores depend significantly on the mix
used in the test samples. What we are looking for are performances that differ significantly from the group. If we have seven or more inspectors, a normal probability plot
(NOPP) would be beneficial in the detection of outliers. It would also be prudent to repeat
the test over time to see if the performance is changing, along with having the committee
take the test.
Bibliography
Automotive Industry Action Group (AIAG). 2002. Measurement Systems Analysis.
3rd edition. Southfield, MI: AIAG.
Duncan, A. J. 1974. Quality Control and Industrial Statistics. 4th edition. Homewood, IL:
Richard D. Irwin.
Juran, J. M. 1935. Inspectors Errors in Quality Control. Mechanical Engineering 59,
no. 10 (October): 643644.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Wheeler, D. J., and R. W. Lyday. 1989. Evaluating the Measurement Process. Knoxville,
TN: SPC Press.
H1317_CH26.qxd
10/15/07
4:33 PM
Page 280
H1317_CH27.qxd
10/15/07
5:50 PM
Page 281
Often processes will have variables that are related to each other, such as the percentage
of calcium carbonate in limestone and the pH of a 25 percent slurry of calcium carbonate.
Rather than maintain a control chart for both of these characteristics, a single control chart
can respond to a process change in the concentration of the slurry and/or the pH of the
slurry. A control chart that allows monitoring of two or more related variables is referred
to as a multivariate control chart. This chart can detect a change in the relationship of two
or more variables.
Two methods of constructing multivariate control charts are presented:
1. Hotellings T 2 statistic (T 2 chart)
2. Standardized Euclidean distance (DE chart)
T 2 Chart
The following example illustrates the T 2 control chart.
Two process characteristics are monitored in the production of breaded shrimp:
X1 = weight of the batter pickup in grams
X2 = volume of the shrimp in cm3
These two characteristics are correlated, and both affect the ratio of shrimp meat to the
total breaded weight.
Samples are taken from the breading line every hour. The weight of the breading is
determined by the difference between the before-washing weight and the after-washing
weight for five randomly selected shrimp. The volume of the shrimp is also determined by
water displacement. Data for the first 14 hours of production are taken.
For each sample, the average, standard deviation S, and range are determined.
281
H1317_CH27.qxd
282
10/15/07
5:50 PM
Page 282
Sample #1
Average:
S:
Range:
Sample #4
X2
X1
X2
X1
X2
X1
X2
6.08
5.70
5.54
6.53
6.45
11.97
11.65
11.58
12.53
12.36
6.09
5.92
5.52
5.98
5.10
11.97
11.82
11.48
11.85
10.96
6.05
5.86
6.06
5.21
6.13
11.91
11.74
11.94
11.27
12.17
6.40
6.14
6.19
6.02
6.38
12.30
12.22
12.25
11.89
12.27
6.06
0.44
0.99
12.02
0.42
0.95
5.72
0.41
0.99
11.62
0.41
1.01
5.86
0.38
0.92
11.81
0.34
0.90
6.23
0.16
0.38
12.19
0.17
0.41
Sample #6
Sample #7
Sample #8
X1
X2
X1
X2
X1
X2
X1
X2
5.96
6.68
5.81
5.53
5.83
11.83
12.72
11.72
11.54
11.74
5.92
6.49
5.79
5.59
6.19
11.80
12.37
11.69
11.62
12.23
6.11
6.13
6.09
5.99
5.80
12.15
12.18
12.08
11.87
11.71
6.53
6.53
5.72
6.05
5.80
12.55
12.53
11.65
11.91
11.70
5.96
0.43
1.15
11.91
0.46
1.18
5.99
0.35
0.90
11.94
0.34
0.75
6.02
0.14
0.33
12.00
0.20
0.47
6.13
0.39
0.81
12.07
0.44
0.90
Sample #9
Average:
S:
Range:
Sample #3
X1
Sample #5
Average:
S:
Range:
Sample #2
Sample #10
Sample #11
Sample #12
X1
X2
X1
X2
X1
X2
X1
X2
5.45
6.52
6.14
6.15
6.39
11.42
12.43
12.19
12.22
12.29
6.02
6.01
6.02
5.74
6.12
11.89
11.87
11.88
11.68
12.16
6.05
5.83
5.77
5.73
5.86
11.92
11.73
11.69
11.65
11.75
6.39
6.56
6.15
5.94
6.11
12.29
12.63
12.23
11.83
12.14
6.13
0.41
1.07
12.11
0.40
1.01
5.98
0.14
0.38
11.90
0.17
0.48
5.85
0.12
0.32
11.75
0.10
0.27
6.23
0.24
0.62
12.22
0.29
0.80
H1317_CH27.qxd
10/15/07
5:50 PM
Page 283
Average:
S:
Range:
Sample #13
Sample #14
X1
X2
X1
X2
5.81
5.52
5.26
5.94
6.45
11.71
11.49
11.39
11.83
12.36
5.87
6.39
6.84
5.67
5.59
11.79
12.28
12.74
11.64
11.63
5.80
0.45
1.19
11.76
0.38
0.97
6.07
0.53
1.25
12.02
0.48
1.11
283
Each of the process variables could be monitored via an average/range control chart
that would, in effect, provide two areas of joint process control. The following square represents the control region using an X chart for each parameter. Any point outside the
square indicates that the process has changedsuch as a subgroup outside the limits for
either one or both of the variables.
UCL (2)
Joint
control
region
Average (2)
LCL (2)
LCL (1)
UCL (1)
Average (1)
However, if we consider the correlation, we can identify a better region of the square.
If a particular unit has a low value for X1, it is likely not to yield a high X2, due to the positive correlation. So, given that the process is under control, a unit produced is unlikely to
yield a low value for X1 and a high value for X2. This means that points are unlikely to
occur in the upper-left or lower-right corner of the square.
A better control region would be one that excludes the two unlikely corners and
enlarges the other remaining corners where the likelihood of occurrence increases. This is
accomplished using the T 2 chart.
H1317_CH27.qxd
10/15/07
284
5:50 PM
Page 284
Individual charts cannot address the correlation between parameters. However, this
does not discount the value in maintaining individual control charts for specific characteristics. If the characteristics are not correlated, then the maintenance of separate control
charts is required.
The data used in this example are highly correlated with r = 0.9770. The joint control
region using the T 2 chart is shown in the following figure:
UCL (2)
Average (2)
LCL (2)
UCL (1)
LCL (1)
Average (1)
The T 2 Equation
T 2 Control Chart for Two Parameters
The value for T 2 for each subgroup is calculated as
Ti 2 =
n
S12
S22
S 2(12 )
S 2( X X )2 + S 2( X X )2 2 S ( X X )( X X ) ,
1
2i
2
(12 )
1i
1
2i
2
1
2 1i
S( 12 ) i =
X11 i X 21 i
( X11 i )( X 21 i )
n
.
n 1
H1317_CH27.qxd
10/15/07
5:50 PM
Page 285
285
Example:
For subgroup sample #1:
Sample #1
X1
Average:
S:
Range:
Sum, S:
X2
X1X2
6.08
5.70
5.54
6.53
6.45
11.97
11.65
11.58
12.53
12.36
72.78
66.41
64.15
81.82
79.72
6.06
0.44
0.99
30.30
12.02
0.42
0.95
60.09
364.88
S( 12 ) i =
S( 12 ) i =
X11 i X 21 i
364.88
( X11 i )( X 21 i )
n
n 1
( 30.30 )( 60.09 )
5
= 0.1837
4
The remaining covariances are calculated in a similar manner for subgroups 214. A
summary of the S(12), subgroup averages, and variances are found in Table 1.
Subgroup (I)
X 1i
X 2i
S 12i
S 22i
S (12) i
1
2
3
4
5
6
7
8
9
10
11
12
13
14
6.06
5.72
5.86
6.23
5.96
5.99
6.02
6.13
6.13
5.98
5.85
6.23
5.80
6.07
12.02
11.62
11.81
12.19
11.91
11.94
12.00
12.07
12.11
11.90
11.75
12.22
11.76
12.02
0.1936
0.1681
0.1444
0.0256
0.1849
0.1225
0.0196
0.1521
0.1681
0.0196
0.0144
0.0576
0.2025
0.2809
0.1764
0.1681
0.1156
0.0289
0.2116
0.1156
0.0400
0.1936
0.1600
0.0289
0.0100
0.0841
0.1444
0.2304
0.1837
0.1666
0.1233
0.0221
0.1982
0.1152
0.0267
0.1708
0.1610
0.0218
0.0128
0.0671
0.1687
0.2557
X1 = 6.00
X 2 = 11.95
S12 = 0.1252
S22 = 0.1220
S ( 12 ) i = 0.1210
Averages:
H1317_CH27.qxd
10/15/07
286
5:50 PM
Page 286
n
S12
S 22
S 2( 12 )
[S ( X
2
2
1i
S 22
S 2( 12 )
X1 ) 2 + S12 ( X 2 i X 2 ) 2 2 S( 12 ) ( X1 i X1 )( X 2 i X 2 )
(S
( 12 ) i
= ( 0.1210 ) 2
n=5
5
= 7893.9
( 0.1252 )( 0.1220 ) ( 0.1210 ) 2
H1317_CH27.qxd
10/15/07
5:50 PM
Page 287
287
T2
1
2
3
4
5
6
7
8
9
10
11
12
13
14
0.2866
6.6167
0.8036
2.4226
0.0657
0.0041
0.9457
0.7063
1.8418
0.9457
3.8192
4.3628
1.6081
0.2011
p ( n 1)
F , p ,n p ,
n p
= probability of a point above the control limit when there has been
no process change
p = number of variables (2 for this example)
n = subgroup sample size (5 for this example)
F,p,n p = value from F distribution table for level of significance with p
and n p degrees of freedom.
2(5 1)
p(n 1)
(9.55) = 25.47.
F , p , n p =
3
n p
The individual T 2 values are plotted, and the upper control limit of 25.47 is drawn
using a broken line.
H1317_CH27.qxd
288
10/15/07
5:50 PM
Page 288
UCL = 25.47
25.0
20.0
History
15.0
10.0
5.0
0
5.0
The process appears to be very well behaved, indicating that the correlation of the data
is stable and well defined.
The control chart of averages for each of the two variables, plotted independently,
follows:
Average control chart, n = 5
Variable X1
7.0
UCL = 6.46
6.0
LCL = 5.54
5.0
13.0
12.0
11.0
LCL = 11.49
H1317_CH27.qxd
10/15/07
5:50 PM
Page 289
Average:
S:
Range:
289
Sample #15
Sample #16
Sample #17
Sample #18
X1
X2
X1
X2
X1
X2
X1
X2
5.60
5.91
5.44
5.80
6.20
11.91
12.05
11.75
11.53
11.61
6.22
6.35
5.84
6.01
6.08
12.17
12.05
11.73
11.55
12.10
6.25
6.55
5.95
6.12
6.38
11.30
12.15
12.05
12.55
12.55
5.94
5.53
5.61
6.05
5.52
12.16
12.01
11.62
11.51
11.95
5.79
0.29
0.76
11.77
0.21
0.52
6.10
0.20
0.51
11.92
0.27
0.62
6.25
0.23
0.60
12.12
0.49
1.25
5.73
0.25
0.53
11.85
0.27
0.65
Average:
S:
Range:
Sample #19
Sample #20
X1
X2
X1
X2
6.44
6.35
6.18
6.17
5.91
12.21
12.46
12.10
12.27
12.36
6.22
5.87
6.05
6.28
6.08
11.75
12.10
11.88
12.00
11.92
6.21
0.20
0.53
12.28
0.14
0.36
6.10
0.16
0.41
11.93
0.13
0.35
The new future data values are used to calculate the T 2 points. Use the process averages, variances, and covariance from the historical data. That is,
Ti 2 = 7893.9.
0.1220( X1i 6.00)2 + 0.1252( X 2i 11.995)2 (2)(0.121)( X1i 6.00)( X 2i 11.95)
A summary of the T 2 for subgroups 1520 follows:
Sample, i
T2
15
16
17
18
19
20
2.282
16.250
10.790
28.511
17.713
13.847
H1317_CH27.qxd
10/15/07
290
5:50 PM
Page 290
The completed control chart with both the historical data and the new future data
follows.
Process change
UCL = 25.47
25.0
20.0
History
15.0
10.0
5.0
0
1
10
15
20
5.0
History
UCL = 6.46
6.0
LCL = 5.54
5.0
1
10
15
20
15
20
13.0
UCL = 12.41
History
12.0
LCL = 11.49
11.0
1
10
Note that the average control chart for both variables X1 and X2 remain in control, failing to detect the change in process correlation.
H1317_CH27.qxd
10/15/07
5:50 PM
Page 291
291
Correlation plot
6.50
17
6.00
X1
11
2
3
13 15
20 16 8
14 1
5
10 6 7
18
412
19
Out-of-control point
5.50
5.00
11.00
11.50
12.00
X2
12.50
13.00
In the preceding plot, the numbered values represent the subgroup sample average.
The bold values show the future data, and the normal text represents the historical data
from which the T 2 chart parameters were derived. Notice that the subgroup sample #18
falls significantly off the best linear line, indicating a departure from the expected value
per the historical data. This departure from the expected correlation is detected using the
T 2 but not using the traditional X /R chart.
X i Ti ,
Sigma ( Ti )
H1317_CH27.qxd
292
10/15/07
5:50 PM
Page 292
where:
Xi Xi
,
Si
Xi Xi
,
SX
A R
S X = standard deviation of the averages can be estimated from 2 .
3
02.995 (p),
where: 20.995 = the 99.5 percentile of the chi-square distribution using p degrees of
freedom (p = number of variables considered).
H1317_CH27.qxd
10/15/07
5:50 PM
Page 293
293
p.
The following table gives the center line values CL and upper control limits
UCL DE = 20.995 ( p) as a function of the number of variables p:
CL
UCL
2
3
4
5
6
7
8
9
10
1.41
1.73
2.00
2.24
2.45
2.65
2.83
3.00
3.16
3.26
3.58
3.85
4.09
4.31
4.50
4.69
4.86
5.02
Batch
X1
Peel, lb/in
X2
Viscosity, cps
1
2
3
4
5
6
7
8
9
10
11
12
17.5
18.3
14.2
16.4
19.0
16.3
19.8
19.8
16.1
17.5
19.4
18.3
1200
1700
825
1025
1600
1320
1725
1600
1150
1300
1825
1550
Batch
X1
Peel, lb/in
X2
Viscosity, cps
13
14
15
16
17
18
19
20
15.5
17.8
15.2
17.5
21.0
17.3
18.2
20.0
1550
1500
950
1300
1850
1375
1400
1925
X 1 = 17.76
X 2 = 1433.5
S 1 = 1.78
S 2 = 305.2
H1317_CH27.qxd
294
10/15/07
5:50 PM
Page 294
For each pair of data points (X1 and X2), a standard Z-score and a combined Z-score
are calculated.
Batch 1:
Standard peel strength, Z1, X 1 =
Standard viscosity, Z1, X 2 =
X 21 X1 18.3 17.76
=
= 0.30
1.78
S1
X 22 X 2 1700 1433.5
=
= 0.87
305.2
S2
(0.30)2 + (0.87)2 = 0.92
Batch 3:
Standard peel strength, Z3, X 1 =
Standard viscosity, Z3, X 2 =
X 31 X1 14.2 17.76
=
= 2.00
1.78
S1
X 32 X 2 825 1433.5
=
= 1.99
S2
305.2
H1317_CH27.qxd
10/15/07
5:50 PM
Page 295
Batch
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
295
X1
Peel, lb/in
X2
Viscosity, cps
Standardized
peel, Z1
Standardized
viscosity, Z2
Combined
DE
17.5
18.3
14.2
16.4
19.0
16.3
19.8
19.8
16.1
17.5
19.4
18.3
15.5
17.8
15.2
17.5
21.0
17.3
18.2
20.0
1200
1700
825
1025
1600
1320
1725
1600
1150
1300
1825
1550
1550
1500
950
1300
1850
1375
1400
1925
0.03
0.30
2.00
0.76
0.70
0.82
1.15
1.15
0.93
0.03
0.92
0.30
1.27
0.02
1.44
0.03
1.82
0.26
0.25
1.26
0.77
0.87
1.99
1.34
0.55
0.37
0.96
0.55
0.93
0.44
1.28
0.38
0.38
0.22
1.58
0.44
1.36
0.19
0.11
1.61
0.77
0.92
2.82
1.54
0.39
0.90
1.50
1.27
1.32
0.44
1.58
0.48
1.32
0.22
2.14
0.44
2.27
0.32
0.27
2.04
X 1 = 17.76
X 2 = 1433.5
S 1 = 1.78
S 2 = 305.2
Using an upper control limit of 3.26 and a center line of 1.41, the DE control chart
using the historical data follows:
UCL = 3.26
3.0
History
2.5
2.0
CL = 1.41
1.5
1.0
0.5
0.0
1
10
Batch number
15
20
H1317_CH27.qxd
10/15/07
296
5:50 PM
Page 296
Batch
Peel, lb/in
Viscosity, cps
Peel, Z1
Viscosity, Z2
DE
21
22
23
24
25
13.5
14.8
15.3
21.3
14.1
1250
1450
1975
2235
1000
2.39
1.66
1.38
1.99
2.06
0.60
0.05
1.77
2.63
1.42
2.46
1.66
2.24
3.30
2.50
UCL = 3.26
3.0
History
2.5
Future
2.0
CL = 1.41
1.5
1.0
0.5
0.0
1
10
Batch number
15
20
25
Control charts for the individual characteristics are presented in the following table.
The control limits are based on X 3S. S is from the sample standard deviation using samples 120.
Characteristic
Peel, X1
Viscosity, X2
Average
LCL
UCL
17.76
1433.50
12.42
517.90
23.10
2349.10
H1317_CH27.qxd
10/15/07
5:50 PM
Page 297
297
UCL = 23.10
History
22.0
Future
20.0
CL = 17.76
18.0
16.0
14.0
LCL = 12.42
12.0
1
10
15
Variable, X 1
20
25
UCL = 2349.1
History
2000
Future
CL = 1433.5
1000
LCL = 517.9
0
1
15
10
Variable, X 2
20
25
Bibliography
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Wheeler, D. J. 1995. Advanced Topics in Statistical Process Control. Knoxville, TN: SPC
Press.
H1317_CH27.qxd
10/15/07
5:50 PM
Page 298
H1317_CH28.qxd
10/18/07
12:01 PM
Page 299
The traditional calculation of Cpk (and other process capability indices) is based on the
assumption of a normal distribution. Some of the distinguishing characteristics of a normal distribution are:
1. The distribution is symmetrical about the average
2. The mean, median, and mode are equal
3. The kurtosis Ku = 0 and the skewness Sk = 0
The calculation of Cpk and Cr can be adjusted to compensate for nonnormalcy by using the
median rather than the average and applying corrections based on Pearson frequency curves.
Knowing the skewness and kurtosis for a given distribution of data allows the determination of the specific values that will result in a probability of 0.00135 and a probability of 0.99865. These two probability values represent the probability of getting a value
less than the average minus three standard deviations and the probability of getting a value
less than the average plus three standard deviations. These specific values are used for the
calculation of Cpk for all distributions, including the normal distribution.
Normal
Sk = 2
Ku = 0
Sk = 0
Ku = 0
Sk = 0
Ku = 1
Sk = +2
Ku = +1
299
H1317_CH28.qxd
300
10/18/07
12:01 PM
Page 300
The Pearson distributions can be used for many distributions, including the normal,
where Sk = 0 and Ku = 0.
As skewness and kurtosis change, so does the shape of the distribution. The following
illustrations demonstrate the effect of these parameter changes on a normal distribution.
Skewness measures how unsymmetrical the distribution is with respect to the average.
The greater the positive skewness, the more the predominant amount of data will be
greater than the average. A negative skewness means that more data will be less than the
average.
Kurtosis is a measure of the peakness of the distribution. A negative kurtosis is indicative of a relative flatness compared to a normal distribution, and a positive kurtosis reflects
a distribution with more peak than a normal distribution.
There are several statistics used to measure skewness and kurtosis. The calculation of
these statistics as applied in this discussion will be defined in terms of various moments
about the arithmetic mean, defined by the formula
( Xi X ) k
,
n
k =
where: X = average
Xi = an individual observation.
The first moment about the mean is always zero, and the second moment about the
mean, adjusted for bias, is the variance.
1 = 0
n
2
= s2
n 1
A measure of skewness often used is
Sk =
= 3 ,
( 2 )3 3
2
4
3.
22
4
,
22
H1317_CH28.qxd
10/18/07
12:01 PM
Page 301
301
Xi
Frequency f
(Xi X )
(Xi X )2
(Xi X )3
(Xi X )4
4
5
6
7
8
9
1
8
6
4
2
1
2.046
1.046
0.046
0.954
1.954
2.954
4.186
1.094
0.002
0.910
3.818
8.726
8.565
1.144
0.000
0.868
7.461
25.777
17.524
1.197
0.000
0.828
14.578
76.145
X = 6.046
= 1.224
S = 1.253
3 = 1.206
n = 22
2 = 1.498
Sk =
Ku =
3
32
4
3
22
1.206
(1.498)2
= 0.66
6.169
3 = 0.25
Ku =
(1.498)2
Average, X = 6.046
Standard deviation, S = 1.253
Upper specification, Uspec = 14.0
Lower specification, Lspec = 2.0
Step 3. Determine the standardized value for getting a probability of less than 0.00135
with specified Ku and Sk, P.00135.
For + Sk, use Table 1.
For Sk, use Table 2.
More exact values are obtained by linear interpolation:
Sk :
0.60
Ku:
0.40
0.25
0.20
0.66
1.496
0.70
1.299
1.486
1.655
1.434
0.0
1.512
1.727
1.966
2.210
2.442
2.653
2.839
3.000
3.140
3.261
3.366
3.458
3.539
3.611
3.674
3.731
3.782
3.828
3.870
3.908
3.943
3.975
4.004
4.031
4.056
4.079
4.101
4.121
4.140
4.157
4.174
4.189
4.204
4.218
4.231
Ku
1.4
1.2
1.0
0.8
0.6
0.4
0.2
0.0
0.2
0.4
0.6
0.8
1.0
1.2
1.4
1.6
1.8
2.0
2.2
2.4
2.6
2.8
3.0
3.2
3.4
3.6
3.8
4.0
4.2
4.4
4.6
4.8
5.0
5.2
5.4
Table 1
1.421
1.619
1.840
2.072
2.298
2.506
2.692
2.856
2.986
3.088
3.164
3.222
3.266
3.300
3.327
3.349
3.367
3.382
3.395
3.405
3.415
3.423
3.430
3.436
3.441
3.446
3.450
3.454
3.458
3.461
3.464
3.466
3.469
3.471
3.473
0.1
1.317
1.496
1.696
1.912
2.129
2.335
2.522
2.689
2.834
2.952
3.045
3.118
3.174
3.218
3.254
3.282
3.306
3.325
3.342
3.356
3.367
3.378
3.387
3.395
3.402
3.408
3.414
3.419
3.423
3.428
3.431
3.435
3.438
3.441
3.444
0.2
1.206
1.364
1.541
1.736
1.941
2.141
2.329
2.500
2.653
2.785
2.896
2.986
3.058
3.115
3.161
3.199
3.229
3.255
3.277
3.295
3.311
3.324
3.326
3.346
3.356
3.364
3.371
3.378
3.384
3.389
3.394
3.399
3.403
3.406
3.410
0.3
1.092
1.230
1.384
1.5551.
1.740
1.930
2.116
2.289
2.447
2.589
2.714
2.821
2.910
2.983
3.043
3.092
3.133
3.167
3.196
3.220
3.241
3.259
3.274
3.288
3.300
3.311
3.321
3.329
3.337
3.344
3.350
3.356
3.362
3.367
3.371
0.4
0.979
1.100
1.232
1.212
1.539
1.711
1.887
2.059
2.220
2.368
2.502
2.622
2.727
2.817
2.893
2.957
3.011
3.055
3.093
3.126
3.153
3.177
3.198
3.216
3.233
3.247
3.259
3.271
3.281
3.290
3.299
3.306
3.313
3.320
3.326
0.5
0.868
0.975
1.089
1.062
1.348
1.496
1.655
1.817
1.976
2.127
2.267
2.396
2.512
2.616
2.708
2.787
2.855
2.914
2.964
3.006
3.043
3.075
3.103
3.127
3.149
3.168
3.184
3.200
3.213
3.225
3.236
3.246
3.256
3.264
3.272
0.6
0.762
0.858
0.957
0.927
1.175
1.299
1.434
1.578
1.726
1.873
2.015
2.148
2.271
2.385
2.488
2.581
2.664
2.736
2.800
2.855
2.904
2.946
2.983
3.015
3.043
3.069
3.091
3.111
3.129
3.145
3.160
3.173
3.186
3.197
3.207
0.7
0.747
0.836
0.804
1.023
1.125
1.235
1.356
1.485
1.619
1.754
1.887
2.013
2.132
2.243
2.345
2.438
2.524
2.600
2.669
2.730
2.784
2.831
2.874
2.911
2.945
2.974
3.001
3.025
3.047
3.066
3.084
3.100
3.114
3.128
0.8
0.692
0.887
0.974
1.065
1.163
1.269
1.382
1.502
1.625
1.748
1.876
1.981
2.089
2.189
2.283
2.369
2.448
2.521
2.586
2.646
2.699
2.747
2.790
2.829
2.864
2.895
2.923
2.949
2.972
2.994
3.013
3.031
0.9
0.766
0.841
0.919
1.000
1.086
1.178
1.277
1.381
1.491
1.602
1.713
1.821
1.925
2.023
2.116
2.202
2.283
2.358
2.427
2.491
2.549
2.602
2.651
2.695
2.735
2.771
2.805
2.835
2.863
2.888
2.911
1.0
0.656
0.723
0.791
0.861
0.933
1.008
1.087
1.172
1.262
1.357
1.456
1.556
1.664
1.755
1.850
1.940
2.026
2.107
2.183
2.254
2.321
2.383
2.440
2.494
2.543
2.588
2.629
2.668
2.703
2.735
2.765
1.1
0.616
0.677
0.739
0.801
0.865
0.931
1.000
1.072
1.149
1.230
1.316
1.404
1.494
1.584
1.673
1.760
1.844
1.924
2.000
2.072
2.140
2.205
2.265
2.321
2.374
2.424
2.470
2.513
2.562
2.589
1.2
0.574
0.630
0.686
0.742
0.799
0.857
0.917
0.979
1.045
1.113
1.185
1.261
1.339
1.420
1.501
1.581
1.661
1.738
1.813
1.684
1.953
2.018
2.080
2.138
2.194
2.246
2.296
2.342
2.386
1.3
0.531
0.583
0.634
0.685
0.736
0.787
0.840
0.894
0.950
1.008
1.068
1.132
1.198
1.267
1.338
1.410
1.483
1.555
1.626
1.695
1.762
1.827
1.889
1.948
2.005
2.059
2.111
2.160
1.4
0.536
0.583
0.629
0.675
0.721
0.768
0.815
0.863
0.913
0.964
1.018
1.073
1.131
1.191
1.253
1.317
1.381
1.446
1.510
1.574
1.636
1.697
1.756
1.813
1.867
1.920
1.5
0.489
0.533
0.575
0.617
0.659
0.701
0.743
0.785
0.828
0.873
0.918
0.965
1.013
1.063
1.115
1.169
1.224
1.281
1.338
1.396
1.453
1.510
1.566
1.621
1.675
1.6
0.484
0.524
0.562
0.600
0.638
0.676
0.714
0.752
0.791
0.830
0.870
0.911
0.953
0.996
1.041
1.088
1.135
1.184
1.234
1.285
1.336
1.387
1.438
1.7
0.475
0.510
0.546
0.580
0.615
0.649
0.683
0.717
0.752
0.787
0.822
0.858
0.895
0.932
0.971
1.011
1.052
1.094
1.137
1.181
1.225
1.8
0.461
0.494
0.526
0.557
0.589
0.620
0.651
0.681
0.712
0.744
0.775
0.807
0.839
0.872
0.905
0.939
0.975
1.010
1.047
1.9
H1317_CH28.qxd
10/18/07
12:01 PM
Page 302
302
0.0
4.243
4.255
4.266
4.276
4.286
4.296
4.305
4.313
4.322
4.330
4.337
4.334
4.351
4.358
4.365
4.371
4.377
4.382
4.388
4.393
4.398
4.403
4.408
Ku
5.6
5.8
6.0
6.2
6.4
6.6
6.8
7.0
7.2
7.4
7.6
7.8
8.0
8.2
8.4
8.6
8.8
9.0
9.2
9.4
9.6
9.8
10.0
10.2
10.4
10.6
10.8
11.0
11.2
11.4
11.6
11.8
12.0
12.2
3.475
3.477
3.478
3.480
3.481
3.483
3.484
3.485
3.486
3.487
3.488
3.489
3.490
3.491
3.492
3.492
3.493
3.494
3.495
3.495
3.496
3.496
3.497
0.1
Table 1 Continued.
3.446
3.448
3.451
3.453
3.454
3.456
3.458
3.459
3.461
3.462
3.464
3.465
3.466
3.467
3.468
3.469
3.470
3.471
3.472
3.473
3.473
3.474
3.475
0.2
3.413
3.416
3.419
3.422
3.424
3.426
3.429
3.431
3.432
3.434
3.436
3.437
3.439
3.440
3.442
3.443
3.444
3.445
3.447
3.448
3.449
3.450
3.451
0.3
3.375
3.379
3.383
3.386
3.389
3.392
3.395
3.398
3.400
3.403
3.405
3.407
3.409
3.411
3.412
3.414
3.416
3.417
3.418
3.420
3.421
3.422
3.424
3.425
0.4
3.331
3.336
3.341
3.345
3.349
3.353
3.357
3.360
3.363
3.366
3.369
3.372
3.374
3.377
3.379
3.381
3.383
3.385
3.387
3.388
3.390
3.392
3.393
3.395
3.396
0.5
3.279
3.286
3.292
3.297
3.303
3.308
3.312
3.316
3.321
3.324
3.328
3.331
3.335
3.338
3.340
3.343
3.346
3.348
3.351
3.353
3.355
3.357
3.359
3.361
3.363
3.364
0.6
3.216
3.225
3.233
3.240
3.247
3.254
3.260
3.265
3.270
3.275
3.280
3.284
3.289
3.292
3.296
3.300
3.303
3.306
3.309
3.312
3.315
3.317
3.320
3.322
3.325
3.327
3.329
0.7
3.140
3.152
3.162
3.172
3.181
3.189
3.197
3.204
3.211
3.218
3.224
3.229
3.235
3.240
3.244
3.249
3.253
3.257
3.261
3.265
3.268
3.272
3.275
3.278
3.281
3.283
3.286
3.289
0.8
3.047
3.062
3.076
3.089
3.100
3.111
3.122
3.131
3.140
3.148
3.156
3.164
3.171
3.177
3.183
3.189
3.195
3.200
3.205
3.209
3.214
3.218
3.222
3.226
3.230
3.233
3.237
3.240
3.243
0.9
2.933
2.952
2.970
2.987
3.003
3.017
3.030
3.043
3.054
3.065
3.075
3.085
3.094
3.103
3.111
3.118
3.125
3.132
3.138
3.144
3.150
3.158
3.161
3.166
3.171
3.175
3.179
3.184
3.188
3.191
3.195
1.0
2.793
2.818
2.841
2.863
2.883
2.902
2.919
2.936
2.951
2.965
2.978
2.990
3.002
3.013
3.023
3.033
3.042
3.051
3.059
3.067
3.075
3.082
3.088
3.095
3.101
3.107
3.112
3.118
3.123
3.128
3.132
3.137
3.141
1.1
2.624
2.656
2.685
2.713
2.739
2.763
2.785
2.806
2.825
2.843
2.860
2.876
2.891
2.906
2.919
2.932
2.943
2.955
2.965
2.975
2.985
2.994
3.003
3.011
3.019
3.026
3.033
3.040
3.046
3.053
3.058
3.064
3.070
3.075
1.2
2.427
2.465
2.501
2.535
2.567
2.597
2.624
2.651
2.675
2.698
2.720
2.740
2.759
2.777
2.794
2.810
2.825
2.839
2.853
2.866
2.878
2.890
2.901
2.911
2.921
2.930
2.940
2.948
2.956
2.964
2.972
2.979
2.986
2.993
1.3
2.206
2.250
2.292
2.332
2.369
2.405
2.438
2.469
2.499
2.527
2.554
2.579
2.603
2.625
2.646
2.666
2.685
2.703
2.720
2.736
2.752
2.766
2.780
2.793
2.806
2.818
2.829
2.840
2.851
2.861
2.870
2.879
2.888
2.896
1.4
1.970
2.019
2.065
2.109
2.151
2.191
2.229
2.265
2.300
2.333
2.364
2.394
2.422
2.449
2.475
2.499
2.522
2.544
2.565
2.585
2.604
2.622
2.639
2.655
2.671
2.686
2.700
2.714
2.727
2.739
2.751
2.762
2.773
2.784
1.5
1.727
1.778
1.827
1.874
1.919
1.962
2.004
2.044
2.083
2.120
2.155
2.189
2.221
2.252
2.282
2.310
2.337
2.363
2.388
2.411
2.434
2.456
2.476
2.496
2.515
2.533
2.551
2.567
2.583
2.598
2.613
2.627
2.641
2.653
1.6
1.489
1.539
1.588
1.635
1.682
1.727
1.771
1.814
1.855
1.895
1.933
1.970
2.005
2.040
2.073
2.104
2.135
2.164
2.192
2.219
2.245
2.271
2.295
2.318
2.340
2.361
2.382
2.401
2.420
2.438
2.456
2.473
2.489
2.505
1.7
1.270
1.316
1.361
1.407
1.452
1.496
1.540
1.583
1.625
1.666
1.706
1.744
1.782
1.818
1.854
1.888
1.921
1.953
1.984
2.014
2.042
2.070
2.097
2.123
2.148
2.172
2.196
2.218
2.240
2.261
2.281
2.301
2.320
2.338
1.8
1.085
1.123
1.162
1.202
1.242
1.282
1.323
1.363
1.403
1.443
1.482
1.521
1.559
1.596
1.632
1.667
1.702
1.736
1.768
1.800
1.831
1.861
1.890
1.918
1.945
1.972
1.998
2.023
2.047
2.070
2.093
2.115
2.136
2.157
1.9
H1317_CH28.qxd
10/18/07
12:01 PM
Page 303
303
0.0
1.512
1.727
1.966
2.210
2.442
2.653
2.839
3.000
3.140
3.261
3.366
3.458
3.539
3.611
3.674
3.731
3.782
3.828
3.870
3.908
3.943
3.975
4.004
4.031
4.056
4.079
4.101
4.121
4.140
4.157
4.174
4.189
4.204
4.218
4.231
4.243
Ku
1.4
1.2
1.0
0.8
0.6
0.4
0.2
0.0
0.2
0.4
0.6
0.8
1.0
1.2
1.4
1.6
1.8
2.0
2.2
2.4
2.6
2.8
3.0
3.2
3.4
3.6
3.8
4.0
4.2
4.4
4.6
4.8
5.0
5.2
5.4
5.6
Table 2
1.584
1.813
2.065
2.320
2.560
2.774
2.961
3.123
3.261
3.381
3.485
3.576
3.654
3.724
3.786
3.842
3.891
3.936
3.976
4.013
4.046
4.077
4.105
4.131
4.155
4.177
4.197
4.217
4.234
4.251
4.267
4.281
4.295
4.308
4.321
4.332
0.1
1.632
1.871
2.134
2.400
2.648
2.869
3.060
3.224
3.364
3.484
3.588
3.678
3.757
3.826
3.887
3.942
3.990
4.034
4.073
4.109
4.142
4.172
4.199
4.224
4.247
4.269
4.288
4.307
4.324
4.340
4.355
4.369
4.383
4.395
4.407
4.418
0.2
1.655
1.899
2.170
2.446
2.704
2.934
3.133
3.303
3.447
3.570
3.676
3.768
3.847
3.917
3.978
4.033
4.081
4.125
4.164
4.199
4.231
4.261
4.287
4.312
4.335
4.356
4.375
4.393
4.410
4.425
4.440
4.454
4.467
4.479
4.490
4.501
0.3
1.653
1.895
2.169
2.454
2.726
2.969
3.179
3.358
3.510
3.639
3.749
3.844
3.926
3.997
4.060
4.115
4.164
4.208
4.248
4.283
4.315
4.344
4.371
4.396
4.418
4.439
4.458
4.476
4.492
4.508
4.522
4.535
4.548
4.560
4.571
4.581
0.4
1.626
1.861
2.131
2.422
2.708
2.968
3.194
3.387
3.550
3.688
3.805
3.905
3.991
4.066
4.131
4.189
4.239
4.285
4.325
4.361
4.394
4.423
4.450
4.475
4.498
4.518
4.537
4.555
4.571
4.587
4.601
4.614
4.627
4.638
4.649
4.659
0.5
1.579
1.803
2.061
2.349
2.646
2.926
3.173
3.385
3.564
3.715
3.843
3.951
4.044
4.124
4.193
4.253
4.307
4.354
4.396
4.433
4.467
4.498
4.525
4.550
4.573
4.594
4.614
4.631
4.648
4.663
4.677
4.691
4.703
4.715
4.725
4.736
0.6
1.516
1.726
1.966
2.241
2.540
2.837
3.109
3.345
3.546
3.715
3.858
3.978
4.080
4.167
4.243
4.308
4.365
4.416
4.460
4.500
4.535
4.567
4.596
4.622
4.645
4.667
4.687
4.705
4.722
4.737
4.752
4.765
4.778
4.789
4.800
4.810
0.7
1.636
1.856
2.108
2.395
2.699
2.993
3.259
3.488
3.681
3.844
3.981
4.096
4.194
4.278
4.351
4.414
4.468
4.517
4.559
4.597
4.631
4.662
4.689
4.714
4.737
4.757
4.776
4.794
4.809
4.824
4.838
4.851
4.862
4.873
4.884
0.8
1.965
2.225
2.518
2.824
3.116
3.378
3.603
3.793
3.900
4.087
4.208
4.296
4.378
4.449
4.511
4.564
4.611
4.653
4.690
4.723
4.752
4.779
4.803
4.825
4.845
4.863
4.879
4.895
4.909
4.922
4.934
4.945
4.956
0.9
1.822
2.052
2.314
2.608
2.914
3.206
3.468
3.693
3.883
4.043
4.177
4.290
4.386
4.468
4.539
4.600
4.653
4.700
4.741
4.777
4.810
4.839
4.865
4.888
4.910
4.929
4.947
4.963
4.978
4.992
5.004
5.016
5.027
1.0
1.885
2.114
2.373
2.665
2.970
3.264
3.529
3.758
3.952
4.115
4.252
4.367
4.472
4.549
4.620
4.682
4.736
4.783
4.824
4.861
4.893
4.922
4.948
4.972
4.993
5.012
5.029
5.045
5.060
5.073
5.086
5.097
1.1
1.928
2.152
2.405
2.690
2.993
3.290
3.561
3.797
3.998
4.168
4.311
4.431
4.532
4.619
4.693
4.757
4.812
4.860
4.903
4.940
4.973
5.002
5.029
5.052
5.074
5.093
5.110
5.126
5.141
5.154
5.166
1.2
1.952
2.169
2.412
2.687
2.984
3.283
3.561
3.808
4.020
4.200
4.352
4.479
4.587
4.678
4.756
4.824
4.882
4.932
4.976
5.015
5.049
5.080
5.107
5.131
5.152
5.172
5.190
5.206
5.220
5.233
1.3
1.960
2.167
2.398
2.658
2.945
3.243
3.529
3.789
4.015
4.209
4.372
4.510
4.627
4.725
4.809
4.881
4.944
4.997
5.044
5.085
5.122
5.153
5.181
5.207
5.229
5.249
5.267
5.284
5.299
1.4
2.149
2.366
2.609
2.881
3.172
3.463
3.736
3.979
4.189
4.369
4.521
4.649
4.758
4.850
4.929
4.996
5.055
5.106
5.150
5.189
5.223
5.253
5.280
5.303
5.325
5.344
5.361
1.5
2.119
2.322
2.547
2.798
3.075
3.364
3.646
3.907
4.137
4.336
4.506
4.650
4.771
4.875
4.963
5.038
5.103
5.159
5.208
5.250
5.288
5.321
5.350
5.376
5.399
5.418
1.6
2.269
2.476
2.705
2.961
3.238
3.522
3.796
4.047
4.269
4.460
4.623
4.762
4.880
4.980
5.066
5.139
5.202
5.257
5.305
5.346
5.383
5.415
5.443
5.468
1.7
2.399
2.609
2.840
3.095
3.370
3.648
3.916
4.160
4.376
4.563
4.723
4.859
4.976
5.075
5.159
5.232
5.295
5.349
5.396
5.437
5.474
5.505
1.8
2.511
2.719
2.949
3.201
3.471
3.745
4.007
4.247
4.461
4.647
4.806
4.943
5.059
5.159
5.244
5.318
5.381
5.436
5.483
5.525
1.9
H1317_CH28.qxd
10/18/07
12:01 PM
Page 304
304
0.0
4.255
4.266
4.276
4.286
4.296
4.305
4.313
4.322
4.330
4.337
4.344
4.351
4.358
4.365
4.371
4.377
4.382
4.388
4.393
4.398
4.403
4.408
Ku
5.8
6.0
6.2
6.4
6.6
6.8
7.0
7.2
7.4
7.6
7.8
8.0
8.2
8.4
8.6
8.8
9.0
9.2
9.4
9.6
9.8
10.0
10.2
10.4
10.6
10.8
11.0
11.2
11.4
11.6
11.8
12.0
12.2
4.343
4.354
4.364
4.373
4.382
4.391
4.399
4.406
4.414
4.421
4.428
4.434
4.441
4.447
4.452
4.458
4.463
4.468
4.473
4.478
4.483
4.487
0.1
Table 2 Continued.
4.429
4.439
4.448
4.457
4.466
4.474
4.481
4.489
4.496
4.503
4.509
4.515
4.521
4.527
4.532
4.538
4.543
4.548
4.552
4.557
4.561
4.565
0.2
4.511
4.521
4.530
4.538
4.547
4.554
4.562
4.569
4.576
4.582
4.588
4.594
4.600
4.605
4.611
4.616
4.621
4.625
4.630
4.634
4.638
4.642
0.3
4.591
4.600
4.609
4.618
4.626
4.633
4.640
4.647
4.654
4.660
4.666
4.672
4.677
4.682
4.687
4.692
4.697
4.701
4.705
4.710
4.714
4.717
4.721
0.4
4.669
4.678
4.695
4.703
4.710
4.717
4.724
4.730
4.736
4.742
4.747
4.753
4.758
4.762
4.767
4.772
4.776
4.780
4.784
4.788
4.791
4.795
4.798
0.5
4.745
4.754
4.763
4.771
4.778
4.785
4.792
4.799
4.805
4.811
4.817
4.822
4.827
4.832
4.837
4.841
4.845
4.850
4.854
4.857
4.861
4.865
4.868
4.871
4.874
0.6
4.820
4.829
4.837
4.845
4.853
4.860
4.867
4.873
4.879
4.885
4.890
4.896
4.901
4.905
4.910
4.914
4.918
4.923
4.926
4.930
4.934
4.937
4.940
4.943
4.947
4.949
0.7
4.893
4.902
4.911
4.919
4.926
4.933
4.940
4.946
4.952
4.958
4.963
4.969
4.974
4.978
4.983
4.987
4.991
4.995
4.999
5.002
5.006
5.009
5.012
5.015
5.018
5.021
5.024
0.8
4.966
4.975
4.983
4.991
4.999
5.006
5.013
5.019
5.025
5.031
5.036
5.041
5.046
5.051
5.055
5.059
5.063
5.067
5.071
5.074
5.078
5.081
5.084
5.087
5.090
5.092
5.095
5.098
0.9
5.037
5.046
5.055
5.063
5.071
5.078
5.085
5.091
5.097
5.103
5.109
5.114
5.118
5.123
5.127
5.132
5.135
5.139
5.143
5.146
5.149
5.153
5.156
5.158
5.161
5.164
5.166
5.169
5.171
5.173
1.0
5.108
5.117
5.126
5.135
5.143
5.150
5.157
5.164
5.170
5.175
5.181
5.186
5.191
5.195
5.200
5.204
5.208
5.211
5.215
5.218
5.222
5.225
5.228
5.230
5.233
5.236
5.238
5.240
5.243
5.245
5.247
5.249
1.1
5.177
5.188
5.197
5.206
5.214
5.222
5.229
5.236
5.242
5.248
5.253
5.259
5.263
5.268
5.272
5.276
5.280
5.284
5.287
5.291
5.294
5.297
5.300
5.303
5.305
5.308
5.310
5.312
5.314
5.316
5.318
5.320
5.322
1.2
5.246
5.257
5.267
5.276
5.285
5.293
5.301
5.308
5.314
5.320
5.326
5.331
5.336
5.341
5.345
5.349
5.353
5.357
5.361
5.364
5.367
5.370
5.373
5.375
5.378
5.380
5.383
5.385
5.387
5.389
5.391
5.393
5.394
1.3
5.312
5.325
5.336
5.346
5.356
5.364
5.372
5.380
5.387
5.393
5.399
5.404
5.410
5.414
5.419
5.423
5.427
5.431
5.434
5.437
5.440
5.443
5.446
5.449
5.451
5.454
5.456
5.458
5.460
5.462
5.464
5.465
5.467
1.4
5.376
5.390
5.403
5.414
5.425
5.434
5.443
5.451
5.459
5.466
5.472
5.478
5.483
5.488
5.493
5.497
5.501
5.505
5.509
5.512
5.515
5.518
5.521
5.523
5.526
5.528
5.530
5.532
5.534
5.536
5.538
5.539
5.541
1.5
5.436
5.452
5.467
5.480
5.492
5.503
5.513
5.522
5.530
5.538
5.545
5.551
5.557
5.562
5.567
5.572
5.576
5.580
5.584
5.587
5.590
5.593
5.596
5.599
5.601
5.603
5.605
5.607
5.609
5.611
5.613
5.614
5.616
1.6
5.491
5.511
5.529
5.542
5.557
5.569
5.581
5.591
5.601
5.609
5.617
5.624
5.631
5.637
5.642
5.647
5.652
5.656
5.660
5.663
5.667
5.670
5.673
5.675
5.678
5.680
5.682
5.684
5.686
5.687
5.689
5.690
5.692
1.7
5.533
5.558
5.581
5.600
5.618
5.634
5.648
5.658
5.669
5.679
5.688
5.696
5.704
5.710
5.717
5.722
5.727
5.732
5.736
5.740
5.744
5.747
5.750
5.753
5.755
5.757
5.760
5.762
5.763
5.765
5.767
5.768
5.769
1.8
5.561
5.593
5.621
5.646
5.669
5.688
5.706
5.722
5.736
5.749
5.760
5.771
5.775
5.783
5.790
5.797
5.803
5.808
5.813
5.817
5.821
5.825
5.828
5.831
5.834
5.836
5.838
5.840
5.842
5.844
5.845
5.847
5.848
1.9
H1317_CH28.qxd
10/18/07
12:01 PM
Page 305
305
H1317_CH28.qxd
306
10/18/07
12:01 PM
Page 306
P.00135 = 1.486
Step 4. Determine the standardized value that will represent a probability of 0.99865 for
obtaining a value less than P.99865. Interpolation will yield a more precise result.
For + Sk, use Table 2.
For Sk, use Table 1.
Sk :
0.60
Ku :
0.40
0.25
0.20
0.66
0.70
2.926
2.837
3.069
3.173
3.109
P.99865 = 3.069
Step 5. Look up the standardized median M using Table 3.
Sk :
0.60
Ku :
0.40
0.25
0.20
0.66
0.70
0.161
0.212
0.175
0.141
0.183
Lp = X SP.00135
Lp = 6.046 1.253(1.486)
Lp = 4.18
Step 7. From the P.99865, calculate the percentile point in units in which the process is being
measured. This is the Up value.
Up = X + SP.99865
Up = 6.046 + 1.253(3.069)
Up = 9.891
M = X SM
If Sk is negative, add SM to X.
M = 6.046 1.253(0.175)
M = 5.82
0.0
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
Ku
1.4
1.2
1.0
0.8
0.6
0.4
0.2
0.0
0.2
0.4
0.6
0.8
1.0
1.2
1.4
1.6
1.8
2.0
2.2
2.4
2.6
2.8
3.0
3.2
3.4
3.6
3.8
4.0
4.2
4.4
4.6
4.8
5.0
5.2
5.4
5.6
Table 3
0.053
0.039
0.031
0.026
0.023
0.020
0.018
0.017
0.015
0.014
0.013
0.013
0.012
0.011
0.011
0.010
0.010
0.009
0.009
0.009
0.008
0.008
0.008
0.008
0.008
0.007
0.007
0.007
0.007
0.007
0.007
0.006
0.006
0.006
0.006
0.006
0.1
0.111
0.082
0.065
0.054
0.047
0.041
0.037
0.034
0.032
0.029
0.028
0.026
0.025
0.024
0.023
0.022
0.021
0.020
0.020
0.019
0.018
0.018
0.017
0.017
0.017
0.016
0.016
0.015
0.015
0.015
0.015
0.015
0.014
0.014
0.014
0.014
0.2
0.184
0.132
0.103
0.085
0.073
0.064
0.058
0.053
0.049
0.045
0.043
0.040
0.038
0.036
0.035
0.034
0.032
0.031
0.030
0.029
0.029
0.028
0.027
0.027
0.026
0.025
0.025
0.025
0.024
0.024
0.023
0.023
0.023
0.022
0.022
0.022
0.3
0.282
0.196
0.151
0.123
0.104
0.091
0.081
0.073
0.068
0.063
0.059
0.055
0.053
0.050
0.048
0.046
0.044
0.043
0.042
0.040
0.039
0.038
0.037
0.037
0.036
0.035
0.034
0.034
0.033
0.033
0.032
0.032
0.031
0.031
0.030
0.030
0.4
0.424
0.284
0.212
0.169
0.142
0.122
0.108
0.097
0.089
0.082
0.077
0.072
0.068
0.065
0.062
0.060
0.057
0.055
0.054
0.052
0.051
0.049
0.048
0.047
0.046
0.045
0.044
0.043
0.043
0.042
0.041
0.041
0.040
0.040
0.039
0.039
0.5
0.627
0.412
0.297
0.231
0.190
0.161
0.141
0.126
0.114
0.105
0.097
0.091
0.080
0.082
0.078
0.074
0.072
0.069
0.067
0.065
0.063
0.061
0.059
0.058
0.057
0.056
0.054
0.053
0.053
0.052
0.051
0.050
0.049
0.049
0.048
0.047
0.6
0.754
0.591
0.419
0.317
0.254
0.212
0.183
0.161
0.145
0.132
0.122
0.113
0.106
0.100
0.095
0.091
0.087
0.084
0.081
0.078
0.076
0.074
0.072
0.070
0.068
0.067
0.066
0.064
0.063
0.062
0.061
0.060
0.059
0.058
0.057
0.057
0.7
0.727
0.586
0.439
0.343
0.280
0.237
0.206
0.183
0.165
0.151
0.140
0.130
0.122
0.116
0.110
0.105
0.101
0.097
0.094
0.091
0.088
0.085
0.083
0.081
0.079
0.078
0.076
0.075
0.073
0.072
0.071
0.070
0.069
0.068
0.067
0.8
0.598
0.468
0.375
0.311
0.266
0.233
0.208
0.188
0.172
0.159
0.148
0.140
0.132
0.126
0.120
0.115
0.111
0.107
0.104
0.101
0.098
0.095
0.093
0.091
0.089
0.087
0.085
0.084
0.082
0.081
0.080
0.078
0.077
0.9
0.681
0.616
0.504
0.413
0.347
0.299
0.263
0.235
0.213
0.196
0.181
0.169
0.159
0.151
0.143
0.137
0.131
0.126
0.122
0.118
0.114
0.111
0.108
0.105
0.103
0.101
0.099
0.097
0.095
0.093
0.092
0.090
0.089
1.0
0.653
0.633
0.542
0.456
0.388
0.336
0.297
0.266
0.242
0.222
0.206
0.192
0.180
0.171
0.162
0.155
0.148
0.143
0.138
0.133
0.129
0.125
0.122
0.119
0.116
0.113
0.111
0.109
0.107
0.106
0.103
0.101
1.1
0.616
0.638
0.579
0.501
0.433
0.379
0.336
0.301
0.274
0.252
0.233
0.217
0.204
0.193
0.183
0.175
0.167
0.161
0.155
0.150
0.145
0.141
0.137
0.133
0.130
0.127
0.124
0.122
0.119
0.117
0.115
1.2
0.574
0.621
0.605
0.545
0.481
0.425
0.379
0.341
0.310
0.285
0.264
0.246
0.231
0.218
0.207
0.197
0.189
0.181
0.174
0.168
0.163
0.158
0.153
0.149
0.145
0.142
0.139
0.136
0.133
0.131
1.3
0.531
0.582
0.607
0.579
0.527
0.474
0.426
0.385
0.351
0.323
0.299
0.279
0.261
0.246
0.233
0.222
0.212
0.203
0.196
0.188
0.182
0.176
0.171
0.167
0.162
0.158
0.155
0.151
0.148
1.4
0.536
0.579
0.590
0.563
0.520
0.474
0.432
0.396
0.365
0.338
0.315
0.295
0.278
0.263
0.250
0.239
0.228
0.219
0.211
0.204
0.197
0.191
0.186
0.181
0.176
0.172
0.168
1.5
0.489
0.533
0.569
0.576
0.554
0.518
0.480
0.443
0.410
0.381
0.355
0.333
0.313
0.296
0.281
0.268
0.256
0.246
0.236
0.228
0.220
0.213
0.207
0.201
0.196
0.191
1.6
0.484
0.524
0.555
0.564
0.549
0.521
0.488
0.456
0.426
0.398
0.374
0.352
0.333
0.316
0.301
0.288
0.276
0.265
0.255
0.246
0.238
0.231
0.224
0.218
1.7
0.475
0.510
0.540
0.552
0.544
0.524
0.498
0.470
0.443
0.417
0.394
0.373
0.354
0.337
0.322
0.308
0.296
0.285
0.274
0.265
0.257
0.249
1.8
0.461
0.494
0.522
0.538
0.539
0.526
0.508
0.483
0.460
0.437
0.415
0.395
0.376
0.359
0.344
0.330
0.317
0.306
0.295
0.285
1.9
H1317_CH28.qxd
10/18/07
12:01 PM
Page 307
307
0.0
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
Ku
5.8
6.0
6.2
6.4
6.6
6.8
7.0
7.2
7.4
7.6
7.8
8.0
8.2
8.4
8.6
8.8
9.0
9.2
9.4
9.6
9.8
10.0
10.2
10.4
10.6
10.8
11.0
11.2
11.4
11.6
11.8
12.0
12.2
0.006
0.006
0.006
0.006
0.006
0.006
0.005
0.005
0.005
0.005
0.005
0.005
0.005
0.005
0.005
0.005
0.005
0.005
0.005
0.005
0.005
0.005
0.1
Table 3 Continued.
0.014
0.014
0.013
0.013
0.013
0.013
0.013
0.013
0.013
0.012
0.012
0.012
0.012
0.012
0.012
0.012
0.012
0.012
0.012
0.012
0.012
0.011
0.2
0.022
0.021
0.021
0.021
0.021
0.021
0.020
0.020
0.020
0.020
0.020
0.019
0.019
0.019
0.019
0.019
0.019
0.019
0.019
0.019
0.018
0.018
0.3
0.030
0.029
0.029
0.029
0.028
0.028
0.028
0.028
0.027
0.027
0.027
0.027
0.027
0.026
0.026
0.026
0.026
0.026
0.026
0.025
0.025
0.025
0.4
0.038
0.038
0.037
0.037
0.037
0.036
0.036
0.036
0.035
0.035
0.035
0.034
0.034
0.034
0.034
0.033
0.033
0.033
0.033
0.033
0.032
0.032
0.032
0.032
0.5
0.047
0.046
0.046
0.045
0.045
0.044
0.044
0.044
0.043
0.043
0.043
0.042
0.042
0.042
0.041
0.041
0.041
0.040
0.040
0.040
0.040
0.040
0.039
0.039
0.039
0.6
0.056
0.055
0.055
0.054
0.054
0.53
0.053
0.052
0.052
0.051
0.051
0.050
0.050
0.050
0.049
0.049
0.049
0.048
0.048
0.048
0.047
0.047
0.047
0.047
0.046
0.046
0.7
0.066
0.065
0.064
0.063
0.063
0.062
0.061
0.061
0.060
0.060
0.059
0.059
0.058
0.058
0.057
0.057
0.057
0.056
0.056
0.055
0.055
0.055
0.054
0.054
0.054
0.054
0.053
0.8
0.076
0.075
0.074
0.073
0.073
0.072
0.071
0.070
0.070
0.069
0.068
0.068
0.067
0.067
0.066
0.066
0.065
0.065
0.064
0.064
0.063
0.063
0.063
0.062
0.062
0.061
0.061
0.061
0.9
0.087
0.086
0.085
0.084
0.083
0.082
0.081
0.080
0.079
0.079
0.078
0.077
0.076
0.076
0.075
0.075
0.074
0.073
0.073
0.072
0.072
0.071
0.071
0.071
0.070
0.070
0.069
0.069
0.069
0.068
1.0
0.100
0.098
0.097
0.096
0.094
0.093
0.092
0.091
0.090
0.089
0.088
0.087
0.086
0.086
0.085
0.084
0.084
0.083
0.082
0.082
0.081
0.080
0.080
0.079
0.079
0.078
0.078
0.078
0.077
0.077
0.076
0.076
1.1
0.113
0.111
0.110
0.108
0.107
0.105
0.104
0.103
0.101
0.100
0.099
0.098
0.097
0.096
0.095
0.094
0.094
0.093
0.092
0.091
0.091
0.090
0.089
0.089
0.088
0.088
0.087
0.087
0.086
0.086
0.085
0.085
0.084
1.2
0.128
0.126
0.124
0.122
0.120
0.118
0.117
0.115
0.114
0.113
0.111
0.110
0.109
0.108
0.107
0.106
0.105
0.104
0.103
0.102
0.101
0.100
0.099
0.099
0.098
0.097
0.097
0.096
0.095
0.095
0.094
0.094
0.093
1.3
0.145
0.142
0.140
0.137
0.135
0.133
0.131
0.129
0.128
0.126
0.124
0.123
0.121
0.120
0.119
0.118
0.116
0.115
0.114
0.113
0.112
0.111
0.110
0.109
0.109
0.108
0.107
0.106
0.105
0.104
0.104
0.104
0.103
1.4
0.164
0.161
0.158
0.155
0.152
0.150
0.147
0.145
0.143
0.141
0.139
0.137
0.135
0.134
0.132
0.131
0.129
0.128
0.127
0.125
0.124
0.123
0.122
0.121
0.120
0.119
0.118
0.117
0.116
0.116
0.115
0.114
0.113
1.5
0.186
0.182
0.178
0.175
0.171
0.168
0.165
0.162
0.160
0.157
0.155
0.153
0.151
0.149
0.147
0.145
0.143
0.142
0.140
0.139
0.137
0.136
0.135
0.133
0.132
0.131
0.130
0.129
0.128
0.127
0.126
0.125
0.124
1.6
0.212
0.207
0.202
0.197
0.193
0.189
0.185
0.182
0.179
0.176
0.173
0.170
0.168
0.165
0.163
0.161
0.159
0.157
0.155
0.153
0.152
0.150
0.149
0.147
0.146
0.144
0.143
0.142
0.141
0.139
0.138
0.137
0.136
1.7
0.242
0.235
0.229
0.223
0.218
0.213
0.209
0.205
0.201
0.197
0.193
0.190
0.187
0.184
0.181
0.179
0.176
0.174
0.172
0.170
0.168
0.166
0.164
0.162
0.160
0.159
0.157
0.156
0.154
0.153
0.152
0.150
0.149
1.8
0.277
0.268
0.261
0.254
0.247
0.241
0.236
0.230
0.226
0.221
0.217
0.213
0.209
0.205
0.202
0.199
0.196
0.193
0.190
0.188
0.185
0.183
0.181
0.179
0.177
0.176
0.173
0.171
0.169
0.168
0.166
0.165
0.163
1.9
H1317_CH28.qxd
10/18/07
12:01 PM
Page 308
308
H1317_CH28.qxd
10/18/07
12:01 PM
Page 309
309
Step 9. Calculate the process capability indices based on normal distribution compensated
for skewness and kurtosis.
A. Capability ratio, Cr and Cp
USpec LSpec
Up Lp
1
Cr =
= 0.48
Cp
Cp =
Cp =
14.0 2.0
9.891 4.180
14.0 5.82
USpec M
=
= 2.01
C pu =
9.891 5.82
Up M
C pl =
C pu
C pk = 2.01
Data
10
Mean
LSL
USL
+3sp
3sp
6
0
2
Samples: 22
Mean:
6.04546
Std Dev: 1.2527
4
Cpk = 2.01
8
10
Spec Lim: (2, 14)
12
14
H1317_CH28.qxd
310
10/18/07
12:01 PM
Page 310
Had the Cpk and Cr been determined without adjusting for skewness and kurtosis, the
results would have been:
C pl =
X LSpec
3s
USpec X
3s
= C pl = 1.08
C pu =
C pk
C pl =
6.046 2.000
3(1.253)
C pl = 1.08
C pu =
14.000 6.046
3(1.253)
C pu = 2.12
and
Cr =
6s
USpec LSpec
7.518
= 0.63.
12.00
Data
10
Mean
LSL
USL
+3sp
3sp
6
0
2
Samples: 22
Mean:
6.04546
Std Dev: 1.2527
4
Cpk = 1.08
Cr = 0.63
8
10
Spec Lim: (2, 14)
12
14
H1317_CH28.qxd
10/18/07
12:01 PM
Page 311
311
for the nonnormal distribution can be used in place of the average. The median can be
determined by locating the point at which 50 percent of the data are below (or above).
The points at which specified proportions of the data are below the pth percentile are
designated as Xp, where p = the proportion less than the value X.
X0.00135 is the data value that 0.135 percent of the data are below, and X0.99865 is the
data value for which 99.865 percent of the data are below. The point located at X0.50000
always represents the median and for a normal distribution will also represent the average.
Calculations for the adjusted Cpk will be made using the following substitutions:
1. In the equation for the traditional C pl =
For:
X
3S
Use:
X0.5000
X0.5000 X0.00135
( X LSpec)
3S
( USpec X )
3S
Use:
X0.5000
X0.99865 X0.5000
Bibliography
Bothe, D. R. 1997. Measuring Process Capability. New York: McGraw-Hill.
Clements, J. A. 1989. Process Capability Calculations for Non-Normal Distributions.
ASQC Quality Progress (September): 95101.
Kotz, S., and N. L. Johnson. 1993. Process Capability Indices. New York: Chapman and
Hall.
H1317_CH28.qxd
10/18/07
12:01 PM
Page 312
H1317_CH29.qxd
10/15/07
2:31 PM
Page 313
Nonparametric Statistics
Many of the traditional statistical techniques used in quality are based on specific distributions such as the normal distribution. Statistical techniques that are not dependent on a
particular distribution (distribution free) but rather on a relative ranking are referred to as
nonparametric methods. There are several such methods available. The methods discussed
here are the Wilcoxon-Mann-Whitney test and the Kruskall-Wallis rank sum test.
313
H1317_CH29.qxd
314
10/15/07
2:31 PM
Page 314
Formula
Time, days
Rank
A
A
B
B
B
A
B
A
B
A
B
B
B
26
28
30
41
46
50
50
68
70
72
73
78
80
1
2
3
4
5
6.5
6.5
8
9
10
11
12
13
Formula A
26
28
50
68
72
(1)
(2)
(6.5)
(8)
(10)
Formula B
30
41
46
50
70
73
78
80
(3)
(4)
(5)
(6.5)
(9)
(11)
(12)
(13)
If more than 20 percent of the observations are involved in ties, this procedure should
not be used.
Step 3. Let n1 = the smaller sample = 5 (na)
n2 = the larger sample = 8 (nb)
n = n1+ n2 = 13
Step 4. Look up R (n1,n2) in Table 1. This table gives critical values for n1 and n2 to 15.
For larger values of n1 and n2, critical values can be estimated by
n1
n n (n + n + 1)
(n1 + n2 + 1) Z 1 2 1 2
,
2
12
where: Z = 1.28 for = 0.10
Z = 1.65 for = 0.05
Z = 1.96 for = 0.025.
Step 5a. If the two sample sizes are equal or if nb is smaller, compute RB (the sum of ranks
for formula B). If RB R (n1,n2), conclude that the average of formula A exceeds formula
B; otherwise, there is no reason to believe that the average for formula A is greater than the
average for formula B.
Step 5b. If na is the smaller (as in this case), compute RA (the sum of ranks for formula A),
and compute RA = nA (n + 1) RA. If RA R(n1, n2), conclude that formula A exceeds formula B; otherwise, there is no reason to believe that the two formulas are different.
Since the smaller sample is associated with formula A (na is the smaller sample), RA (the
sum of ranks for formula A) is calculated as follows:
H1317_CH29.qxd
10/15/07
2:31 PM
Page 315
Nonparametric Statistics
315
n1 (Smaller sample)
n2
10
11
12
13
14
15
0.10
0.05
0.025
0.10
0.05
0.025
0.10
0.05
0.025
0.10
0.05
0.025
0.10
0.05
0.025
0.10
0.05
0.025
0.10
0.05
0.025
0.10
0.05
0.025
0.10
0.05
0.025
0.10
0.05
0.025
0.10
0.05
0.025
0.10
0.05
0.025
0.10
0.05
0.025
7
6
7
6
8
7
6
9
8
7
10
8
7
11
9
8
11
10
8
12
10
9
13
11
9
14
11
10
15
12
10
16
13
11
16
13
11
13
11
10
14
12
11
15
13
12
16
14
13
17
15
14
19
16
14
20
17
15
21
18
16
22
19
17
23
20
18
25
21
19
26
22
20
20
19
17
22
20
18
23
21
20
25
23
21
27
24
22
28
26
23
30
27
24
32
28
26
33
30
27
35
31
28
37
33
29
30
28
26
32
29
27
34
31
29
36
33
31
38
35
32
40
37
34
42
38
35
44
40
37
46
42
38
48
44
40
41
39
36
44
41
38
46
43
40
49
45
42
51
47
44
54
49
46
56
52
48
59
54
50
61
56
52
55
51
49
58
54
51
60
56
53
63
59
55
66
62
58
69
64
60
72
67
62
75
69
65
70
66
62
73
69
65
76
72
68
80
75
71
83
78
73
86
81
76
90
84
79
87
82
78
91
86
81
94
89
84
98
92
88
102
96
91
106
99
94
106
100
96
110
104
99
114
108
103
118
112
106
123
116
110
127
120
115
131
125
119
136
129
123
141
133
127
149
142
136
154
147
141
159
152
145
174
166
160
179
171
164
200
192
184
10
11
12
13
14
15
RA = 1 + 2 + 6.5 + 8 + 10 = 27.5,
RA = n A (n + 1) RA = 5(13 + 1) 27.5 = 42.5,
R(n1, n2),
H1317_CH29.qxd
10/15/07
316
2:31 PM
Page 316
where:
= 0.05
n1 = 5
n2 = 8
R0.05(5,8) = 23.
Design 2
Design 3
1.7
1.9
6.1
12.5
16.5
25.1
30.5
42.1
82.5
13.6
19.8
25.2
46.2
46.2
46.2
13.4
20.9
25.1
29.7
46.9
(1)*
(2)
(3)
(4)
(7)
(10.5)
(14)
(15)
(20)
(6)
(8)
(12)
(16.5)
(16.5)
(16.5)
(5)
(9)
(10.5)
(13)
(18)
Procedure:
Step 1. Choose a level of significance, or alpha risk . For this example, let = 0.10.
Step 2. Look up the chi-square for 21 for K 1 degrees of freedom (K = 3).
20.90,2 = 4.62
Step 3. We have n1, n2, n3, . . . , n observations on each product design 1, 2, 3, . . . , K.
N = n 1 + n 2 + n3 + . . . + n K
Assign a rank to each observation according to its size in relation to all n observations.
That is, assign rank 1 to the smallest, rank 2 to the next largest, and so on. In cases of ties,
assign to each the rank that would have been assigned had the observations differed
H1317_CH29.qxd
10/15/07
2:31 PM
Page 317
Nonparametric Statistics
317
slightly. If more than 20 percent of the observations are involved in ties, this procedure
should not be used.
Step 4. Calculate Ri (the sum of the ranks of the observations on the ith product) for each
of the products.
R1 = 76.5
R2 = 78.0
R3 = 55.5
R2
12
i 3( N + 1),
N ( N + 1) ni
12
(2280.30) 63,
420
H = 2.15.
H=
Step 6. If H > 21, conclude that the averages of the K designs differ; otherwise, there is
no reason to believe that the averages differ.
When using this procedure, there should be a minimum of five observations for each
sample.
Bibliography
Kohler, H. 1988. Statistics for Business and Economics. 2nd edition. Glenview, IL: Scott,
Foresman and Company.
Siegel, S. 1956. Nonparametric Statistics. New York: McGraw-Hill.
H1317_CH29.qxd
10/15/07
2:31 PM
Page 318
H1317_CH30.qxd
10/17/07
2:37 PM
Page 319
Normal Distribution
Distributions are mathematical functions that describe data. Graphically they can describe
unique shapes. One of the most frequently used is the normal distribution.
The normal distribution is used exclusively with the application of statistical process control (SPC) of variables data and is assumed when working with process capability indices.
The pattern or histogram of normally distributed data is shaped like a bell and,
hence, is sometimes referred to as the bell curve. One of the subjective tests for a normal distribution is to construct a frequency histogram and look for a bell-shaped curve.
There are other tests for a normal distribution, such as performing a normal probability
plot or performing a chi-square goodness-of-fit test (see the module Testing for a Normal Distribution).
All distributions have the following two defining characteristics:
1. A location statistic that defines the central tendency of the data (examples are the
average, median, and mode)
2. A variation statistic that defines the amount of dispersion or variation of the data
(examples are standard deviation and range)
The central tendency of data can be described by one of the following statistics:
( X X )2
n 1
X = average of individuals
n = sample size
= sum
319
H1317_CH30.qxd
10/17/07
320
2:37 PM
Page 320
Approximately 99.7 percent of the data are contained within the limits of X 3S
While the limits of 1S, 2S, and 3S are important from a practical perspective such
as in the application of SPC, there are other times we might want to apply other multiples.
Using the probability density function for the normal distribution, we can determine the
probability of events occurring given an average and standard deviation. The following
examples illustrate this use of the normal distribution.
Example 1:
Given an average of X = 23.0 and a standard deviation of S = 3.1, what is the probability
of getting a value greater than 28.58?
Step 1. Sketch out the problem.
Draw a normal distribution. The vertical line dividing the distribution in half represents the
average. A vertical line drawn in the tail area represents the point of interest; in this case
it is 28.58. The standard deviation is not drawn in, but it can be recorded near the average.
The exact location of the point of interest is not critical, but it should be on the appropriate side of the average line. It is drawn so that there is some tail area remaining outside the
limit. There is an assumption that there is a measured quantity greater than (for this case)
the point of interest.
X = 23.00
S = 3.1
28.58
H1317_CH30.qxd
10/17/07
2:37 PM
Page 321
Normal Distribution
321
X X
28.58 23.00
, ZU =
= ZU = 1.80.
S
3.1
x.x 0
x.x 1
x.x 2
x.x 3 . . . x.x 9
Z
4.0
3.9
3.8
3.7
.
.
.
.
1.8
0.03593
The proportion of the distribution of data that are greater than 28.58 is 0.03593.
An alternative way of expressing this is 3.593 percent. This proportion may also be
expressed as a rate greater than 28.58 by taking the reciprocal and rounding to the nearest integer.
1
= 27.83 28
0.03593
This implies that 1 out of 28 will exceed 27.83.
Example 2:
The specification for the amount of current drawn for a small electric motor under test is
3.0 0.50. Thirty-five motors are tested with an average draw of 2.82 amps and a standard
deviation of 0.45. What percentage of the motors fail to meet the specification requirement? What is the combined rate of nonconformance?
H1317_CH30.qxd
322
10/17/07
2:37 PM
Page 322
2.5
X = 2.82
S = 0.45
3.5
x.x 0
x.x 1
Z
4.0
3.9
3.8
3.7
3.6
.
.
.
.
.
0.7
0.23885
x.x 2
x.x 3 . . . x.x 9
x.x 0
x.x 1
Z
4.0
3.9
3.8
3.7
3.6
.
.
.
.
.
.
1.5 0.06552
x.x 2
x.x 3 . . . x.x 9
H1317_CH30.qxd
10/17/07
2:37 PM
Page 323
Normal Distribution
323
A ZL of 0.71 indicates that a proportion of 0.23885 (23.885 percent) will be below the
lower specification of 2.50. The rate of nonconformance below the lower specification is
1
= 4.186 = 4. 1 out of 4 will fall below the lower specification.
0.23885
A ZU of 1.51 indicates that a proportion of 0.06552 (6.552 percent) will be greater than
the upper specification of 3.50. The rate of nonconformance above the upper specification is
1
= 15.263 = 15. 1 out of 15 will exceed the upper specification.
0.06552
The total exceeding the specification is
1
1
=
= 3. 1 out of 3 will not meet the specification.
0.23885 + 0.06552 0.30437
Example 3:
Samples of cat food are analyzed, and the average sodium chloride content is 110.5 parts
per thousand with a standard deviation of 15.0. What proportion of the product is expected
to have between 80 and 90 parts per thousand sodium chloride?
Step 1. Sketch out the problem.
In this example, we have two points of interest. Both are below the average; therefore,
we will designate the Z-score for the 80.0 point of interest as ZLL and the Z-score for the
90.0 point of interest as ZL.
80
90
X = 110.5
S = 15.0
H1317_CH30.qxd
324
10/17/07
2:37 PM
Page 324
X XLL
S
ZLL =
110.5 80.0
15.0
ZLL = 2.03
X XL
S
ZL =
110.5 90.0
15.0
Z L = 1.37
B. Calculate ZL.
ZL =
x.x 0
x.x 1
x.x 2
Z
4.0
3.9
3.8
3.7
3.6
3.5
.
.
.
.
2.0
For Z = 1.37
x.x 3
0.02118
x.x0
x.x1
x.x2 . . . x.x7
Z
4.0
3.9
3.8
3.7
3.6
3.5
.
.
.
.
1.3 0.08534
The proportion of the data less than 80.0 is 0.02118, and the proportion of the data that
are less than 90.0 is 0.08534. Therefore, the proportion between these two limits is
0.08534
0.02118
0.06416.
The percentage of data between 90.0 and 80.0 is 6.416.
H1317_CH30.qxd
10/17/07
2:37 PM
Page 325
Normal Distribution
325
X 3S = 16
X = 40
X 3S = 64
,
n
SX =
8
,
5
S X = 3.8.
= Averages of n = 5
= Individuals
16
29.2 X = 40 50.8
64
H1317_CH30.qxd
326
10/17/07
2:37 PM
Page 326
H1317_CH31.qxd
10/15/07
2:35 PM
Page 327
p Chart
Control charts are selected according to the type of data taken to monitor a process. Quality data or measurements fall into one of the following two broad categories:
1. Variable data are data obtained in which the initial observation is a measurement
that can take on any value within the limits that it can occur and within the limits
of our ability to make the measurement. Examples of variables are temperature,
length, pressure, and mass.
2. Attribute data are obtained from an initial observation of a yes or a no. Attribute data
are normally associated with the compliance to a requirement. Meeting the requirement yields a yes response, and failing a requirement yields a no response. Variables
data can be converted to attribute data by comparison to a requirement. An example
of attribute data is the percentage of observations that do not conform to a requirement. Frequently used attributes are the percentage of customers not satisfied with
the quality of a service and the percentage of voters voting for a specific candidate.
Once we have decided on the type of data being collected, we can narrow the field of
available control charts from which to select. Assuming that the data are attribute in nature,
we will have the following four basic types available:
1.
2.
3.
4.
p chart
np chart
c chart
u chart
327
H1317_CH31.qxd
328
10/15/07
2:35 PM
Page 328
a fixed period of time such as a shift. The number of units produced during this
period may vary from day to day. We may, on the other hand, elect to inspect a fixed
or constant number of units for a specified period of time.
2. Attribute type: Attribute data can be related to the total number of defective units
(nonconforming) within a sample or the number of defects (nonconformities) in a
sample. For example, if we inspect 50 printed circuit boards and find three missing
components, two reversed polarities, and three cracks, we would report eight
defects. From this information, we do not know the number of defective boards. All
of the eight defects could have been found on one board, or eight defective boards
could have had one defect each. A defective is a unit of inspection that does not conform to a quality requirement. A defect is the specific incident that caused the unit
of inspection to be defective.
Chart name
c chart
u chart
np chart
p chart
Sample size
Constant
May be variable
Constant
May be variable
Attribute type
Defect
Defect
Defective
Defective
If we can define constant as actually being a variable sample size where the amount
of variation is defined as zero, then the c chart and np chart can be eliminated from our
field. The p chart and u chart can perform very well in place of the c chart and np chart.
We do not have to vary the sample size for the p chart and u chart, but we may do so
if we want. A p chart with a constant sample size performs exactly as an np chart, and a u
chart with a constant sample size performs exactly as a c chart. This module discusses the
application and principles of the p chart.
Control charts for variables are based on the normal distribution as a statistical model.
Attribute data are binary in nature in that the initial response of an inspection activity
yields only two possible outcomes: the unit of inspection either conforms to a requirement
or it does notyes or no. The response is discrete. The distribution on which the p chart
is based is the binomial distribution.
Where the sample size n is relatively small in proportion to the population N such that
n/N is less than 0.10 and the product of the proportion defective and the sample size is
greater than five (np > 5), the binomial distribution approximates the normal distribution.
The average of the binomial distribution is np, and the standard deviation is npq , where
n = sample size, p = proportion nonconforming, and q = 1 p.
As with all attribute control charts, the construction and use of a p chart has the following operations or steps:
1. Collection of historical data to characterize the process
2. Calculation of a location statistic (an average)
3. Determination of statistical upper and lower control limits based on 3 standard
deviations
4. Construction of the control chart and plotting of historical data
5. Continuation of monitoring the data, looking for signs of a process change in the
future
H1317_CH31.qxd
10/15/07
2:35 PM
Page 329
p Chart
329
9(1 p)
,
p
H1317_CH31.qxd
330
10/15/07
2:35 PM
Page 330
Sample #1
Sample #2
Sample #3
Sample #4
8/12/95
8/15/95
8/18/95
8/21/95
n = 50
np = 4
p = 0.08
n = 50
np = 2
p = 0.04
n = 50
np = 6
p = 0.12
n = 50
np = 3
p = 0.06
Sample #5
Sample #6
Sample #7
Sample #8
8/23/95
8/25/95
8/28/95
9/2/95
n = 50
np = 2
p = 0.04
n = 50
np = 0
p = 0.00
n = 50
np = 5
p = 0.10
n = 50
np = 3
p = 0.06
Sample #9
Sample #10
Sample #11
Sample #12
9/5/95
9/8/95
9/11/95
9/14/95
n = 50
np = 7
p = 0.14
n = 50
np = 1
p = 0.02
n = 50
np = 4
p = 0.08
Sample #14
Sample #15
Sample #16
9/17/95
9/20/95
9/23/95
9/25/95
n = 50
np = 3
p = 0.06
n = 50
np = 6
p = 0.12
n = 50
np = 3
p = 0.06
n = 50
np = 4
p = 0.08
n = 50
np = 4
p = 0.08
Sample #13
4 + 2 + 6 + 3 ++ 4
50 + 50 + 50 + 50 + + 50
p=
57
800
p = 0.071
Note that p is reported one more decimal place than the individual sample proportion
defective. This is done to minimize the probability of a single point falling exactly on the
average.
H1317_CH31.qxd
10/15/07
2:35 PM
Page 331
p Chart
331
p(1 p)
.
n
If we had taken a sample size of 118, the expected LCL would have been 0.000072.
As a practical matter, the sample size should be somewhat larger than the absolute minimum of 118 (say 125), in which case the LCL would have been 0.002.
Step 4. Construct the control chart, sketch in the average and control limits, and plot the
data.
H1317_CH31.qxd
332
10/15/07
2:35 PM
Page 332
Traditionally, averages are drawn as solid lines (), and control limits are
drawn as broken lines (- - - - - - - - - -). A vertical plotting scale should be chosen such that
part of the chart is available for those future points that might fall outside the control limits. A vertical wavy line has been drawn to separate the history from the future data. Only
the process average proportional defective p is drawn out into the future area of the control chart.
UCL = 0.18
History
0.15
P
0.10
P = 0.071
0.05
0.0
1
13
17
21
25
29
33
37
Step 5. Continue to monitor the process, looking for indications that a change has
occurred.
Data will be collected into the future. Evidence supporting the conclusion that a change
has taken place will be in the form of statistically rare patterns referred to as SPC (statistical process control) detection rules. These detection rules have the following two characteristics:
1. All of these SPC patterns or rules are statistically rare for a normal, stable process
2. Most of these SPC detection rules indicate a direction for the process change
Rule 1: A lack of control or process change is indicated whenever a single point falls
outside the UCL or LCL.
Rule 2: A lack of control or process change is indicated whenever seven consecutive
points fall on the same side of the process average.
Rule 3: A lack of control or process change is indicated whenever seven consecutive
points are steadily increasing or decreasing.
Anytime a process change is indicated, as evident by violation of one of the SPC
detection rules, an investigation should be undertaken to identify an assignable cause.
There is a small probability that one of these rules will be invoked when, in fact, no change
in the process has occurred. This probability, however, is so small that we assume the violation truly represents a process change.
H1317_CH31.qxd
10/15/07
2:35 PM
Page 333
p Chart
333
We will continue to monitor the example process to see if a change has occurred relative to our historical period from 8/12 to 9/25. The p chart does allow for a change in the
sample size. To demonstrate the effects on our chart, we will change the sample size on
the future data from 50 to 150.
A change in sample size will result in a change in the value for three standard deviations, and since we add and subtract three standard deviations to and from the historical
process average proportional defective to determine the UCL and LCL, we will expect to
see a change in these limits. The larger the sample size, the closer the control limits will
be to the average. This decrease in standard deviation as we increase the sample size
increases our ability to detect smaller process changes by increasing the sample size.
The new control limits using a sample size of 150 and the historical average proportion of 0.071 are given by
0.071 3
0.071(1 0.071)
150
0.071 30.063.
UCL = 0.134
LCL = 0.008
We will sketch in the new control limits and extend the historical average, 0.071. We
do not recalculate the average using the future data point, because we want to detect a
process change relative to the historical average.
After sketching in the new control limits for n = 150 and extending the average, we
plot the future data points, looking for a violation of one of the three SPC detection rules.
Future data:
Sample #17
Sample #18
Sample #19
Sample #20
9/28/95
10/1/95
10/4/95
10/7/95
n = 150
np = 18
p = 0.12
n = 150
np = 4
p = 0.03
n = 150
np = 9
p = 0.06
n = 150
np = 15
p = 0.10
Sample #21
10/10/95
n = 150
np = 14
p = 0.09
Sample #22
Sample #23
Sample #24
10/13/95
10/17/95
10/20/95
n = 150
np = 14
p = 0.09
n = 150
np = 0
p = 0.00
n = 150
np = 3
p = 0.02
H1317_CH31.qxd
334
10/15/07
2:35 PM
Page 334
Notice that rule one has been violated on sample #23, suggesting that the process is
performing better than expected (proportional defective has decreased) relative to the historical average of 0.071.
Completed p chart:
UCL = 0.18
History
0.15
UCL = 0.134
P
0.10
P = 0.071
0.05
LCL = 0.008
0.0
1
13
17
21
25
29
33
37
P (1 P )
n
P (1 P )
n
LCL = 0.00
UCL = P 3
UCL = 0.00358
In this case, if we had one defective in 250, the control chart would be out of control with
a calculated p value of 0.0040, which exceeds the UCL of 0.00358. This process would
always be out of control.
H1317_CH31.qxd
10/15/07
2:35 PM
Page 335
p Chart
335
The application of a p chart for this case is awkward and is not appropriate for two
reasons:
1. No LCL is available
2. Only samples yielding zero defective units will maintain an in-control condition,
and we will soon violate the rule of seven points in a row below the average
Another drawback in using the p chart with very low proportional defective is the slow
response to a large shift in the process average.
Consider the previous example, where the process average defective was 300 ppm
(0.0003), the sample size was n = 250, and the UCL was 0.00358. In this case, finding one
defective in a sample of n = 250 leads to an out-of-control condition. If the process proportion defective were to double to 600 ppm, the probability of getting one defective in a
sample of 250 is
n!
x n x
Px =1 =
P Q
x !(n x )!
Px = (250)(0.0006)(0.9994)249 = 0.1292.
This means that, on average, 1/0.1292, or about 8, samples must be inspected before a
defective is found and the process is judged to be out of control.
Rather than focus on the number of defective units found, an alternative approach
would be to focus on the cumulative count of nondefective items until a defective item is
found. The probability of the nth item being defective is given by
Pn = (1 p ) n 1 p
n=
n = 1, 2 , 3, 4 , (a geometric distribution)
1
(n = the average run length given a proportion defective of p).
p
Pn = 1 1 np +
p + .
!
2
( )
= 1 e .
n
n
Solving for n:
n = n ln[1 Pn ].
By substituting 0.5 for Pn in this equation, we can solve for the median n as follows:
n = n ln[1 0.50],
n = 0.693n .
H1317_CH31.qxd
10/15/07
336
2:35 PM
Page 336
Control chart for cumulative counts of conforming items, the Sp control chart:
Average = 0.693 n
LCL = n ln 1
2
and since is small
LCL = n
2
UCL = n ln 1 1
UCL = n ln
2
If is set to 0.0027 as with traditional Shewhart control charts, the control limits become
LCL = 0.00135 n ,
UCL = 6.608.1
Sample
number
Number of
units tested
Cumulative
count when
count stops
Number of
defective units
since previous stop
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
500
500
240
800
1000
500
900
488
750
1000
850
1000
630
500
1000
770
380
500
1000
1240
800
1800
2300
3200
3688
750
1750
2550
3550
4180
500
1500
2270
2650
0
0
1
0
0
0
0
1
0
0
0
0
1
0
0
0
1
1Goh,
Cumulative count
to defective
1240 (A)
3688 (B)
4180 (C)
2650 (D)
T. N. 1987. A Control Chart for Very High Yield Processes. Quality Assurance 13, no. 1 (March): 18-22.
H1317_CH31.qxd
10/15/07
2:35 PM
Page 337
p Chart
337
100,000
UCL = 19,400
10,000
1000
x x
CL = 2037
100
10
LCL = 4
1
1
10
Sample number
15
H1317_CH31.qxd
10/15/07
338
2:35 PM
Page 338
Exercise:
Construct a p chart using the following data:
Set = 0.05
Sample
number
Number of
units tested
Cumulative
count when
count stops
Number of
defective units
since previous stop
1
2
3
4
5
6
7
8
9
10
11
12
13
500
300
175
600
500
300
700
1200
600
550
750
800
358
500
800
175
600
1100
1400
2100
3300
600
1150
750
1550
1908
0
1
1
0
0
0
0
1
0
1
0
0
1
Cumulative count
to defective
800
175
3300
1150
1908
Exponential Model
If we consider the process to be mature and in a steady state of failure rate (but very low),
the failures that do occur may be modeled using the traditional reliability model based on
the exponential distribution. This model is used extensively in reliability predictions.
R=e
where: = mean time between failure or, for this application, average number of units
accumulated when a defective is encountered
n = specific sample size for a given level of reliability
R = the probability of reaching a count of n parts without a failure
By solving this relationship for n at specified levels of reliability, a control chart based
on the cumulative sample size to the point of a single failure occurs. If we select an alpha
risk of = 0.05, the control chart limits and average change as explained in the following
paragraphs.
The LCL is equal to the sample size that represents a reliability of 0.95, or 95 percent
probability that this number of units will be found defect free. The UCL represents the sample size that there will be a 5 percent probability of reaching with no defect being found.
From the exponential relationship for determining reliability, we determine the sample size n for the appropriate reliability.
H1317_CH31.qxd
10/15/07
2:35 PM
Page 339
p Chart
R=e
339
n
ln R =
n = ln R
= Mean number of units before failure or average sample size to failure
The center line for this chart will be the median, which is determined by the point at
which the reliability is 50 percent.
Example:
Hypodermic needles are manufactured in a continuous process. The needles are 100 percent inspected by an automatic visual inspection device. The following 18 samples represent the number of units produced before a defective unit is found.
Step 1. Collect historical data:
Date
Sample size n
Date
Sample size n
11/1
11/4
11/7
11/9
11/11
11/13
11/16
11/19
11/22
120,000
89,000
145,000
78,000
138,000
100,000
180,000
90,000
130,000
11/24
11/28
11/30
12/1
12/3
12/6
12/9
12/12
12/15
160,000
76,000
140,000
110,000
98,000
155,000
80,000
105,000
93,000
Step 2. Select an alpha risk and determine the appropriate reliability. Calculate limits.
For this example, the risk is set at 5 percent ( = 0.05).
The LCL is based on a reliability of 95 percent.
LCL = ln 0.95
LCL = 5948
Center line = Median. Use a reliability of 50 percent for the median.
Median = ln 0..50 Median = 115, 944 ln 0.50 Median = ( 115, 944)( 0.693)
Median = 80, 349
Note: The center line is optional and is not considered in the traditional SPC detection rules,
which use the average and the assumption of a normal or approximate normal distribution.
H1317_CH31.qxd
340
10/15/07
2:35 PM
Page 340
400,000
UCL = 347,832
300,000
200,000
100,000
LCL = 5948
0
Note on interpretation: The only rule for detecting a process change is exceeding the UCL
or LCL. Exceeding the upper limit indicates that the process is performing better than the
historical period. Having a point below the LCL indicates that the process is performing
poorer than the historical period.
A UCL of n = 347,832 is the same as having a failure rate of 1/347,832, or
0.00000287, which is equivalent to a process defective rate of 2.87 ppm. Exceeding this
limit implies that the process has improved.
An LCL of n = 5948 is the same as having a failure rate of 1/5948, or 0.000168, which
is equivalent to a process defective rate of 168 ppm. Having points below this limit implies
that the process has deteriorated.
Step 4. Continue to monitor the process, looking for signs of a change.
Two indicators are used to detect a process change:
1. Seven consecutive points increasing or decreasing
2. A single point outside the UCL or LCL
H1317_CH31.qxd
10/15/07
2:35 PM
Page 341
p Chart
Date
Sample size n
12/18
12/20
12/23
12/26
12/28
12/30
189,000
150,000
220,000
160,000
290,000
360,000
History
400,000
UCL = 347,832
Future
341
Out of control
300,000
200,000
100,000
LCL = 5948
0
The out-of-control point gives us supporting evidence that the process has changed and for
the better.
Data Transformation
Lloyd S. Nelson2 developed a simple method of control charting processes where the nonconforming rate is extremely low (that is, ppm). This method is based on the fact that the
Weibull distribution with a slope parameter () of 3.6 has a skewness of zero and a kurtosis of 2.72. This compares favorably to a normal distribution where the skewness is zero
and the kurtosis is 3.00.
Therefore, the required transformation is simply to raise the cumulative number of
successful samples required to obtain a defective to the 1/3.6 power. This transformed
2Nelson, L. S. 1994. A Control Chart for Parts-per-Million Nonconforming Items. Journal of Quality Technology 26,
no. 3 (July): 239240.
H1317_CH31.qxd
10/15/07
342
2:35 PM
Page 342
value is then treated as a normally distributed variable and is then monitored using the traditional individual/moving range control chart.
Example:
A machine produces 3.5-inch formatted diskettes that are 100 percent certified by writing
and the reading data. Each time a diskette fails, it is rejected and the count for the interval
to the last failure is recorded. The objective is to establish a control chart based on the failure data.
Data:
Sample
number
Cumulative count
to failure, X
Transformed
data, Y = X 0.2777
Moving
range
1
2
3
4
5
6
7
19,400
22,780
14,000
12,700
18,600
25,000
15,100
15.5
16.2
14.2
13.8
15.3
16.7
14.5
0.7
2.0
0.4
1.5
1.4
2.2
17
UCL = 16.54
16
15
14
13
12
Average = 15.17
LCL = 13.80
H1317_CH31.qxd
10/15/07
2:35 PM
Page 343
p Chart
343
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Wheeler, D. J., and D. S. Chambers. 1992. Understanding Statistical Process Control. 2nd
edition. Knoxville, TN: SPC Press.
H1317_CH31.qxd
10/15/07
2:35 PM
Page 344
H1317_CH32.qxd
10/15/07
3:05 PM
Page 345
H1317_CH32.qxd
346
10/15/07
3:05 PM
Page 346
In the special case where the subgroup sizes are all the same (np charts or c charts),
this is the right thing to do. However, for p charts and u charts, this creates a couple of
problems: (1) by giving equal weight to larger and smaller subgroups, the estimates are
statistically biased, and (2) we lose the variable control limits we are accustomed to seeing in p charts and u charts. The XmR charts inherent leveling off of the control limits
(which really should be wider for smaller subgroups and narrower for larger subgroups)
may cause some points to be misclassified (as shown in the following examples).
Laneys p' and u' control charts provide a solution for all the aforementioned problems
encountered with the traditional p and u control charts.
Example 1. Proportion defective:
Consider the data set in Table 1, where
i = subgroup
ni = subgroup size
xi = number of deefective items
pi = xi / ni = proportion defective
p = xi / ni = center line (CL)
= p (1 p ) = estimated population standard deviation
LCLi = p 3 / ni = lower control limit (LCL)
UCLi = p + 3 / ni = upper control limit (UCL)
Table 1 p chart.
ni
xi
pi
CL
LCL
UCL
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
1075
1085
1121
1081
1011
4114
3997
4061
4165
4001
8084
6587
8062
8218
8096
4049
4147
4202
4100
4070
301
395
302
402
207
1100
1046
1340
975
1166
2626
2875
2669
2382
2009
1361
1194
1190
1320
1381
0.28
0.36406
0.2694
0.37188
0.20475
0.26738
0.2617
0.32997
0.23409
0.29143
0.32484
0.43647
0.33106
0.28985
0.24815
0.33613
0.28792
0.2832
0.32195
0.33931
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.26531
0.26551
0.26619
0.26543
0.264
0.28595
0.28564
0.28581
0.28609
0.28565
0.29214
0.29048
0.29212
0.29227
0.29215
0.28578
0.28604
0.28618
0.28592
0.28584
0.34976
0.34957
0.34889
0.34965
0.35108
0.32912
0.32944
0.32926
0.32899
0.32943
0.32294
0.3246
0.32296
0.32281
0.32292
0.32929
0.32904
0.3289
0.32916
0.32924
H1317_CH32.qxd
10/15/07
3:05 PM
Page 347
347
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0.0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
Figure 1 p chart.
A graph of the last four columns gives us the classical p chart shown in Figure 1.
Clearly, overdispersion is present. Converting this to the Xi chart produces Table 2, where
Ri = pi pi 1
(i = 2,..., 20)
R = Ri / 19
CL = p
LCL = CL 2.66 R
UCL = CL + 2.66 R
Table 2 Xi chart.
pi
CL
LCL
UCL
R(p)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
0.28
0.36406
0.2694
0.37188
0.20475
0.26738
0.2617
0.32997
0.23409
0.29143
0.32484
0.43647
0.33106
0.28985
0.24815
0.33613
0.28792
0.2832
0.32195
0.33931
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.12995
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.48513
0.08406
0.09465
0.10248
0.16713
0.06263
0.00568
0.06827
0.09587
0.05733
0.03341
0.11163
0.10541
0.04121
0.0417
0.08799
0.04821
0.00472
0.03875
0.01736
H1317_CH32.qxd
10/15/07
348
3:05 PM
Page 348
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0.0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
Figure 2 Xi chart.
CL
LCL
UCL
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
1.95655
4.03411
2.76687
4.584001
7.0824
5.58164
6.28032
3.097385
10.2712
2.20831
3.370839
22.67474
4.576489
3.4744
11.58
3.942801
2.73779
3.41897
1.999867
4.392585
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
3
Note that the mean of the Z values is assumed to be zero and the standard deviation is one
(so the three sigma control limits are at +/ 3). The completed z chart can be seen in Figure 3.
H1317_CH32.qxd
10/15/07
3:05 PM
Page 349
349
25
20
15
10
5
0
5
10
15
Figure 3 z chart.
This may not seem very helpfulbecause it isnt. We can still clearly see the overdispersion problem. Obviously, +/ 3 is not where the control limits should be. So . . . where
should they be?
It was at this very point that Laneys great inspiration occurred. He says he suddenly remembered Dr. Donald J. Wheelers haunting admonition: Why assume the
variation when you can measure it? [7]. So he decided to repeat the z chart, but this
time, instead of assuming that the standard deviation was one, he would actually measure it using the same methods as in an XmR chart. He reasoned that this would duly
adhere to the basic rule of control charts that the control limits should be based on
short-term variation. He knew that the root-mean-square formula for the variance
would be wrong, since that would represent long-term variation. This gives us Table 4
Table 4 z ' chart.
CL
LCL
UCL
R(z)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
1.95655
4.03411
2.76687
4.584001
7.0824
5.58164
6.28032
3.097385
10.2712
2.20831
3.370839
22.67474
4.576489
3.4744
11.58
3.942801
2.73779
3.41897
1.999867
4.392585
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.6512
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
21.65119
5.990662
6.800979
7.35087
11.6664
1.500762
0.698683
9.377709
13.36854
8.062851
5.579146
19.3039
18.09825
8.050891
8.105579
15.52278
6.680592
0.681179
5.418837
2.392718
H1317_CH32.qxd
350
10/15/07
3:05 PM
Page 350
25
20
15
10
5
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
5
10
15
20
25
Figure 4 z ' chart.
and Figure 4, which are constructed just like the previous XmR chart except that the
values plotted are the Z-scores of the original p values, and the ranges are now the distances between consecutive Z-scores.
As Figure 4 shows, this is much better. Instead of the assumed value of 1, it turns out
that the estimated process standard deviation, by the usual formula for the standard deviation in an XmR chart, is
z = avg( Rz ) / 1.128 = 7.22.
The control limits in the z' chart are 7.22 times wider than they were in the z chart. We
may conclude that there is 7.22 times more short-term variation in our process than the
binomial assumption alone can account for.
This is fine, but many people are confused by z charts; they dont understand the units.
So all that remains is to undo the original z-transformation formula so that everything is
once again depicted on the p-plane. This gives us a modified p chart, which Laney dubbed
the p' chart:
UCLi / LCLi = p 3 / ni z .
Note the addition of the last term. It magnifies our estimate of process variability to
account for the batch-to-batch variation that exists in addition to the usual binomial sampling error. As Roger Hoerl at GE explained it, this merely redefines the rational subgroup for this application.
Our final result is shown in Table 5 and Figure 5.
H1317_CH32.qxd
10/15/07
3:05 PM
Page 351
CL
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
1.9566
4.03411
2.7669
4.584
7.0824
5.5816
6.2803
3.09739
10.271
2.2083
3.37084
22.6747
4.57649
3.4744
11.58
3.9428
2.7378
3.419
1.99987
4.39259
0.28
0.36406
0.2694
0.37188
0.20475
0.26738
0.2617
0.32997
0.23409
0.29143
0.32484
0.43647
0.33106
0.28985
0.24815
0.33613
0.28792
0.2832
0.32195
0.33931
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
0.30754
LCL
0.00285
0.00426
0.00917
0.0037
0.0066
0.15179
0.14953
0.15078
0.15274
0.1496
0.19643
0.18445
0.19628
0.19734
0.19651
0.15054
0.15241
0.15343
0.15152
0.15095
UCL
R(z)
0.61223
0.61082
0.60591
0.61138
0.62172
0.46329
0.46555
0.4643
0.46233
0.46547
0.41865
0.43063
0.4188
0.41774
0.41856
0.46453
0.46267
0.46165
0.46355
0.46413
5.99066
6.80098
7.35087
11.6664
1.50076
0.69868
9.37771
13.3685
8.06285
5.57915
19.3039
18.0983
8.05089
8.10558
15.5228
6.68059
0.68118
5.41884
2.39272
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0.0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
351
H1317_CH32.qxd
10/15/07
352
3:05 PM
Page 352
Table 6 u chart.
CL
LCL
UCL
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
103
102
101
98
101
197
199
200
194
201
499
500
496
500
504
300
305
303
291
292
194
398
204
92
750
598
992
602
791
8
1009
508
498
996
492
1199
599
592
293
591
1.8835
3.90196
2.0198
0.93878
7.42574
3.03553
4.98492
3.01
4.07732
0.0398
2.02204
1.016
1.00403
1.992
0.97619
3.99667
1.96393
1.9538
1.00687
2.02397
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
1.65288
1.6508
1.64868
1.64215
1.64868
1.77091
1.77247
1.77323
1.76854
1.774
1.88546
1.88566
1.88488
1.88566
1.88643
1.82936
1.83142
1.8306
1.82553
1.82597
2.50534
2.50742
2.50954
2.51608
2.50954
2.38731
2.38575
2.38499
2.38968
2.38422
2.27276
2.27256
2.27334
2.27256
2.27179
2.32886
2.3268
2.32762
2.33269
2.33226
H1317_CH32.qxd
10/15/07
3:05 PM
Page 353
10
8
6
4
2
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
Figure 6 u chart.
CL
LCL
UCL
R(u)
1.8835
3.90196
2.0198
0.93878
7.42574
3.03553
4.98492
3.01
4.07732
0.0398
2.02204
1.016
1.00403
1.992
0.97619
3.99667
1.96393
1.9538
1.00687
2.02397
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
3.0896
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
7.24782
2.01847
1.88216
1.08103
6.48697
4.39021
1.94939
1.97492
1.06732
4.03752
1.98224
1.00604
0.01197
0.98797
1.01581
3.02048
2.03273
0.01014
0.94692
1.0171
353
H1317_CH32.qxd
354
10/15/07
3:05 PM
Page 354
10
9
8
7
6
5
4
3
2
1
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
Figure 7 X chart.
CL
LCL
UCL
R(z)
1.3768
12.7677
0.4134
7.829
37.2651
9.30988
28.4286
9.13008
19.302
20.051
0.8841
16.486
16.605
1.3509
17.172
23.034
1.395
1.5128
12.685
0.6534
1.8835
3.90196
2.0198
0.93878
7.42574
3.03553
4.98492
3.01
4.07732
0.0398
2.02204
1.016
1.00403
1.992
0.97619
3.99667
1.96393
1.9538
1.00687
2.02397
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
2.07911
4.8746
4.9086
4.9431
5.0498
4.9431
2.949
2.9236
2.9111
2.9877
2.8987
1.0801
1.077
1.0897
1.077
1.0644
1.9954
1.9619
1.9752
2.0579
2.0508
9.03283
9.06683
9.10134
9.20801
9.10134
7.10719
7.08186
7.06934
7.14592
7.05691
5.23837
5.23521
5.24791
5.23521
5.22266
6.15362
6.12008
6.1334
6.21614
6.20905
14.1445
13.1811
7.41564
45.0941
27.9552
19.1187
19.2985
10.172
39.3533
19.1672
15.6023
0.11877
15.2542
15.8211
40.206
24.429
0.11782
11.1724
12.0318
H1317_CH32.qxd
10/15/07
3:05 PM
Page 355
10
9
8
7
6
5
4
3
2
1
0
355
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
We can see that point #5 is not out of control, even though the X chart suggested that
it was. In this example, z was computed as 16.31. There was over 16 times more (shortrun) variation in the process than the Poisson
assumption alone could have predicted.
Note: There have been cases where z < 1. This may be an indication that the subgroups are not independentthat there is positive autocorrelation in the data.
Conclusion
The p' chart and the u' chart seem to represent the general case for attribute data, where
there is both within-subgroup variation and between-subgroup variation in the data,
whereas the classical p chart and u chart only work in the special case where there is no
between-subgroup variation present. While such a condition might exist in a very carefully
controlled environment, experience teaches that between-subgroup variation is quite common in real-world processes. That being the case, and considering that the p' chart/u' chart
will be virtually identical to the p chart/u chart if the overdispersion is found to be negligible, it is a good idea to always start with the p' or u' versions and see what happens.
Bibliography
1. Breyfogle, Forrest. 1999. Implementing Six Sigma. New York: John Wiley & Sons,
177178.
2. Heimann, Peter A. 1996. Attributes Control Charts with Large Sample Sizes. Journal of Quality Technology 28:451459.
H1317_CH32.qxd
356
10/15/07
3:05 PM
Page 356
3. Jones, G., and K. A. Govindaraju. 2000. Graphical Method for Checking Attribute
Control Chart Assumptions. Quality Engineering 13, no. 1: 1926.
4. Laney, David B. 2002. Improved Control Charts for Attributes. Quality Engineering
14, no. 4: 531537.
5. Laney, David B. 2006. P-Charts and U-Charts Work (But Only Sometimes). Quality Digest, November, 75.
6. Western Electric. 1956. Statistical Quality Control Handbook. Indianapolis, IN: AT&T
Customer Information Center.
7. Wheeler, Donald J. 1993. Understanding VariationThe Key to Managing Chaos.
Knoxville, TN: SPC Press.
H1317_CH33.qxd
10/15/07
4:46 PM
Page 357
Pareto Analysis
N = number of individuals
x = income
A and a = constants.
Since 1951, various questions have been raised to the effect that Pareto was the wrong
man to identify with the concentration of defects. Another candidate who might be associated with the application of defect concentration would be M. O. Lorenz, who depicted
the concentration of wealth in the now-familiar curves, using cumulative percent of population on one axis and cumulative percent of wealth on the other.
The following illustrates the basic procedure for performing a Pareto analysis:
1. Decide the objective of the Pareto analysis (examples include defects by type, dollars of sales by location, defects by location, or numbers of physicians by specialty).
2. Develop a list of the responses to be classified. A list of defects for a paint finish
might include smears, voids, scratches, and runs. A list of geographical locations
might include Seattle, Chicago, New York, and so on.
3. Design a tally sheet to collect the data.
4. Collect data.
5. Arrange the collected data in order of the smallest frequency occurrence first and
the largest frequency last.
6. Determine the percentage of each classification as a percentage of the total.
7. Accumulate the percent distribution in a cumulative manner until a total of 100 percent has been obtained.
357
H1317_CH33.qxd
10/15/07
358
4:46 PM
Page 358
Area of concern
Number of responses
Courteousness
Cost
Procedure explanation
Attractiveness of office
Mouthwash taste
Pain
Magazine selection
Uniforms
Promptness of schedule
Professionalism
Cleanliness
10
35
11
3
1
60
8
2
2
6
12
Total
150
Steps 57. Rearrange the data in descending order, and determine the percentage that each
area is of the total and the cumulative percent total.
Area
Pain
Cost
Cleanliness
Procedure explanation
Courteousness
Magazine selection
Professionalism
Attractiveness of office
Promptness of schedule
Uniforms
Mouthwash taste
Totals:
Number of
responses
Percentage of
total responses
60
35
12
11
10
8
6
3
2
2
1
40.0
23.3
8.0
7.3
6.7
5.3
4.0
2.0
1.3
1.3
0.8
150
100.0
Cumulative
percentage of total
40.0
63.3
71.3
78.6
85.3
90.6
94.6
96.6
97.9
99.2
100.0
From this tabular Pareto analysis, it can be readily seen that 63 percent of Dr. Wises
problems are caused by 18 percent of the problem list.
H1317_CH33.qxd
10/15/07
4:46 PM
Page 359
Pareto Analysis
359
10
15
20
25
30
35
40
45
Pain
Cost
Cleanliness
Procedure explanation
Courteousness
Magazine selection
Professionalism
Office attractiveness
Promptness
Uniforms
Mouthwash
It is clear that Dr. Wise should focus on reducing patient discomfort and lowering his
fees.
In this example, an incident of Magazine is no more important than an incident of
Cleanliness when, in fact, an incident of Cleanliness could result in an infection and possible litigation and a potential malpractice suit.
By assigning weighting factors to different levels of severity, such as critical, major,
and minor, the Pareto analysis can be augmented to reflect the level of impact for the
respective assignments.
This assignment of weighting factors results in a weighted Pareto analysis.
Assume that Dr. Wise has categorized the various areas of concern into three groups
and has assigned a weighting factor of 30, 5, and 1 in order of decreasing severity.
Category
Critical
Major
Minor
Weighting factor
30
5
1
H1317_CH33.qxd
360
10/15/07
4:46 PM
Page 360
With this scale, it will require 30 minor incidents to impact the Pareto analysis as much
as one critical incident. Likewise, it will require six major incidents to impact the analysis
as much as one critical incident.
Each area of concern will be multiplied by the weighting factor, and the resulting product will be used to develop the Pareto analysis.
Area of
concern
Courteousness
Cost
Procedure explanation
Attractiveness of office
Mouthwash taste
Pain
Magazine selection
Uniforms
Promptness of schedule
Professionalism
Cleanliness
Total
Number of
responses
10
35
11
3
1
60
8
2
2
6
12
Category
Factor
Pareto
response
Major
Major
Major
Major
Minor
Major
Minor
Minor
Major
Critical
Critical
5
5
5
5
1
5
1
1
5
30
30
50
175
55
15
1
300
8
2
10
180
360
150
1156
The Pareto responses are reordered in a descending manner, and the percentage of total
and the cumulative percentage of total are determined.
Area of
concern
Cleanliness
Pain
Professionalism
Cost
Procedure explanation
Courteousness
Attractiveness of office
Promptness of schedule
Magazine selection
Uniforms
Mouthwash
Total
Weighted
Pareto response
Percentage
of total
Cumulative
percentage of total
360
300
180
175
55
50
15
10
8
2
1
31.1
26.0
15.6
15.1
4.8
4.3
1.2
0.9
0.6
0.3
0.1
31.1
57.1
72.7
87.8
92.6
96.9
98.1
99.0
99.6
99.9
100.0
1156
100
Note that the number one and number two rankings of concerns are now Cleanliness
and Pain. Of the responses, 72.7 percent are described by 27.3 percent (3 out of 11) of the
concerns list.
Optionally, a Pareto diagram can be constructed.
H1317_CH33.qxd
10/15/07
4:46 PM
Page 361
Pareto Analysis
361
Problem:
Construct a Pareto diagram for the distribution of letters in the following paragraph. What
percentage of the letters are responsible for 80 percent of the total letters in the passage?
Black Huckleberry has fair landscape value and is quite attractive in the fall. It
provides fair cover but no nesting to speak of. Fruit is available for a short period
of time during the summer only and is a favorite of more than 45 species of birds
such as grosbeaks, towhees, bluebirds, robins, chickadees, and catbirds.
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New York:
McGraw-Hill.
Gryna, F. M., R.C.H. Chua, and J. A. De Feo. 2007. Jurans Quality Planning and Analysis
for Enterprise Quality. 5th edition. New York: McGraw-Hill.
Hayes, G. E., and H. G. Romig. 1982. Modern Quality Control. 3rd edition. Encino, CA:
Glencoe.
Juran, J. M. 1960. Pareto, Lorenz, Cournot, Bernoulli, Juran and Others. Industrial
Quality Control 17, no. 4 (October): 25.
H1317_CH33.qxd
10/15/07
4:46 PM
Page 362
H1317_CH34.qxd
10/17/07
2:40 PM
Page 363
Pre-Control
Pre-control is a technique for validating a process to perform at a minimal level of conformance to the specification. F. E. Satterthwaite (1954) initiated the introduction of precontrol in 1954 and identified the development team as consisting of himself, C. W. Carter,
W. R. Purcell, and Dorian Shainin. Shainin has been responsible for the popularization of
pre-control.
Pre-control attempts to control a process relative to the specification requirements,
whereas statistical process control (SPC) monitors the process relative to natural variations
as described by the normal distribution and its associated average and standard deviation.
SPC is truly based on historical characterization of a process and is independent of any
specification requirements.
Description
The specification (bilateral) is divided into four equal parts. The area from the specification nominal to one-half of the upper specification and the area from the nominal to onehalf of the lower specification are called the pre-control limits. The area between the two
pre-control limits is called the green zone. The areas outside the pre-control limits but not
exceeding the upper or lower specification limits are called the yellow zone. Areas outside
the specification limits are called the red zone. Data used in the pre-control technique are
frequently placed on a chart, and the zones are colored according to the aforementioned
scheme, leading to what is referred to as a rainbow chart.
Nominal
Lower
PC
specification limit
Red
Yellow
PC
Upper
limit specification
Green
Green
1/2 print
tolerance
Print tolerance
363
Yellow
Red
H1317_CH34.qxd
364
10/17/07
2:40 PM
Page 364
Example:
What are the pre-control limits if the specification is 4.50 0.40?
1. Divide the tolerance by 4:
0.80
= 0.20.
4.0
2. Subtract the value from nominal for the lower PC limit:
4.50 0.20 = 4.30.
3. Add the value to nominal for the upper PC limit:
4.50 + 0.20 = 4.70.
Thus, the two pre-control limits are located at 4.30 and 4.70, one-half the normal
tolerance.
Note: If the specification is unilateral (single sided), the sole pre-control limit is placed
halfway between the target, or nominal, and the single specification limit.
Pre-Control Rules
1. Initial setup: five consecutive samples are measured. All measurements are
required to fall within the pre-control limits before proceeding.
2. Following the initial setup approval, two consecutive samples are taken from the
process and measured.
a. If both units fall in the green zone, continue.
b. If one unit falls inside the pre-control limits and the other falls in the yellow
zone, continue.
c. If both units fall inside the same yellow zone, the process is stopped and adjustments are made. After the adjustments have been made, the setup approval is
again required. Five consecutive units must fall within the pre-control limits.
d. If both of the units measured fall in opposite yellow zones, then it is likely that
variation has increased. The process is stopped, and efforts are made to reduce
variation. Again, five consecutive measurements are required to fall within the
pre-control limits.
e. A single measurement outside the specification limits (red zone) will result in
stopping the process and adjusting. Following adjustment, setup approval is
required.
After any adjustment or process stoppage, five consecutive measurements are required
to fall inside the pre-control limits.
The statistical foundation of pre-control is based on the assumption that the process
average is equal to the nominal or target value and that normal variation (average 3 standard deviations) is equal to the product or process tolerance. That is, Cpk = 1.00 and
H1317_CH34.qxd
10/17/07
2:40 PM
Page 365
Pre-Control
365
Cp = 1.00. For a normal distribution, approximately 86 percent of the data (12 out of 14)
will fall inside the pre-control limits (the green zone), and approximately 7 percent (1 out
of 14) will fall outside the pre-control limits but inside the specification (the yellow zone).
The probability of getting measurements outside the specification is very low, approximately 3/1000.
The probabilities for various outcomes can be seen in the following table:
Color zones
Decision
Red
Stop
Get help
Stop and
adjust
Continue
Yellow
Green
Yellow
B
A,B
A
B
Lower
specification
Red
Probability
1/14
1/14
1/14
1/14
12/14
1/14
1/14
12/14
12/14
A
A,B
A,B
B
A
A
B
B
A
PC
PC
limit
limit
Nominal (target)
Upper
specification
1/14 = 1/196
1/14 = 1/196
1/14 = 1/196
1/14 = 1/196
12/14 = 144/196
12/14 = 12/196
12/14 = 12/196
1/14 = 12/196
1/14 = 12/196
Total = 196/196
Continue to monitor the process, sampling at a frequency equivalent to six pairs (A,B)
chosen between each adjustment or stoppage.
Time between adjustments, hours
1
2
3
4
5
etc.
H1317_CH34.qxd
366
10/17/07
2:40 PM
Page 366
Plotting the data using traditional SPC methods is vital in detecting process changes,
identifying and removing (special) causes of variation, and distinguishing the differences
in assignable and chance (common) causes.
If we set the specification limits at 2, the pre-control limits become 1. Pre-control
assumes that the process average equals the nominal and that Cp = 1.00. This implies that
= 2/(3 Cp). We may now determine the probability that a pre-control chart scheme will
give a signal (both observations in a yellow zone or one in the red zone) for and degree of
process change and the process Cp. This probability * is the probability of getting a signal and is calculated as follows:
g = Pr[ X in green zone; ]
= Pr[ 1 X +1; ]
(1 )
( 1 )
= Pr
Z
3(1 )Cp
3( 1 )Cp
= Pr
Z
2
2
3( 1 ) Cp
3(1 ) Cp
=
2
2
2
2
* = Pr[signal; ]
= 1 {g ( g + 2 y )}
1/* = average run length (ARL) for a signal given Cp and amount of process shift
where: = amount of process shift from nominal
= cumulative probability density function for the normal distribution
Similarly, the probability of detecting a process change (a single point outside a control limit) using SPC can be determined.
= Pr[sample average exceeds a control limit; ]
where: = process shift from nominal
H1317_CH34.qxd
10/17/07
2:40 PM
Page 367
Pre-Control
367
Table 1 ARLs of pre-control (PC) and average control charts (CC) with subgroup size n = 2 as a
function of level shift and process capability index, Cp.
Cp = 0.75
Cp = 1.00
Cp = 1.30
Percent
defective
ARL
CC
ARL
PC
Percent
defective
ARL
CC
ARL
PC
Percent
defective
ARL
CC
ARL
PC
0.00
0.10
0.20
0.30
0.40
0.50
0.60
0.70
0.80
0.90
1.00
2.445
2.535
2.810
3.274
3.940
4.821
5.935
7.299
8.932
10.850
13.066
370.4
328.9
243.0
164.6
108.8
72.4
48.9
33.7
23.8
17.1
12.6
9.6
9.4
8.7
7.7
6.7
5.7
4.8
4.1
3.5
3.0
2.6
0.270
0.300
0.395
0.567
0.836
1.231
1.791
2.561
3.594
4.948
6.681
370.4
302.0
188.3
108.8
63.4
38.1
23.8
15.4
10.4
7.3
5.3
44.4
41.4
33.9
25.6
18.5
13.2
9.5
6.9
5.2
4.0
3.1
0.009
0.012
0.022
0.044
0.087
0.166
0.307
0.547
0.941
1.565
2.514
370.4
266.1
135.0
66.3
34.2
18.8
11.1
6.9
4.6
3.3
2.4
370.3
310.7
198.1
110.6
59.9
33.1
19.2
11.7
7.5
5.1
3.6
Cp = 1.75
c
Percent
defective
ARL
CC
ARL
PC
0.00
0.10
0.20
0.30
0.40
0.50
0.60
0.70
0.80
0.90
1.00
0.000
0.000
0.000
0.000
0.001
0.004
0.012
0.032
0.082
0.194
0.433
370.4
214.6
82.8
33.7
15.4
7.9
4.5
2.9
2.0
1.6
1.3
13273.1
8200.0
2847.0
890.7
297.6
110.5
45.9
21.3
11.0
6.3
3.9
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New York:
McGraw-Hill.
Ledolter, J., and A. Swersey. 1997. An Evaluation of Pre-Control. Journal of Quality
Technology 29, no. 2 (April): 163171.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New York:
John Wiley & Sons.
Satterthwaite, F. E. 1954. A Simple, Effective Process Control Method, Report 54-1.
Boston: Rath & Strong.
Wheeler, D. J., and D. S. Chambers. 1992. Understanding Statistical Process Control.
2nd edition. Knoxville, TN: SPC Press.
H1317_CH34.qxd
10/17/07
2:40 PM
Page 368
H1317_CH35.qxd
10/15/07
3:12 PM
Page 369
The relationship of the process specification requirements and the actual performance of
the process can be described by one of several index numbers: Cpk, Cr, and Cp. Each of
these index numbers represents a single value that serves as a measure of how well the
process can produce parts or services that comply with the specification requirement. All
process capability indices are applicable only to variables measurements and assume that
the data are normally distributed.
Cpk
The Cpk index uses both the average and standard deviation of the process to determine
whether the process is capable of meeting the specification. For processes where the average is closer to one specification than the other (process not centered), which is more common than being centered, or where there is only a single specification limit (unilateral), the
following table gives the expected level of nonconformance in parts per million (ppm).
Cpk value
Nonconforming, ppm
Rate of nonconformance
1.60
1.50
1.40
1.30
1.20
1.10
1.00
0.90
0.80
0.70
1
3
13
48
159
483
1350
3467
8198
17865
1/1,000,000
1/333,333
1/76,923
1/20,833
1/6,289
1/2070
1/741
1/288
1/122
1/56
Cpk is defined as
Cpk = min (Cpu, Cpl),
where:
( USL X )
3
( X LSL )
C pl =
3
C pu =
369
H1317_CH35.qxd
370
10/15/07
3:12 PM
Page 370
R
d2
or =
S
c4
X = 4.505
|
|
|
|
|
|
|
|
LSL = 4.49
Target = 4.500
USL = 4.510
Steps 3 and 4. Determine the location of the average and select the appropriate relationship for calculating the Cpk.
The average is closer to the upper specification; therefore, the defining relationship for Cpk is
Cpk = Cpu =
H1317_CH35.qxd
10/15/07
3:12 PM
Page 371
371
This process exhibits a high degree of capability, yielding approximately three ppm as a
defective rate.
Example 2:
The minimum latching strength for a component is 7500 pounds. There is no maximum
requirement or target. The average and standard deviation for 130 measurements are
X = 6800
|
|
|
|
LSL = 7500
C pk = C pl =
(6800 7500)
= 1.87
(3)(125)
The process is extremely poor; the process average is below the lower specification
limit. Note that the Cpk can be negative.
C pk 2
1
+
.
9n 2n 2
H1317_CH35.qxd
372
10/15/07
3:12 PM
Page 372
Risk, %
Confidence, %
Z/2
.1
1.0
5.0
10.0
99.9
99.0
95.0
90.0
3.29
2.58
1.96
1.64
3.09
2.33
1.64
1.28
We are assuming a 5 percent risk that our interval will not contain the true Cpk; therefore, the level of confidence is 95 percent, and the Z/2 value is 1.96.
C pk Z / 2
C pk 2
1
(1.155)2
1
+
= 1.15 1.96
+
= 1.15 0.26
9n 2n 2
(9)(45) (2)(45) 2
H1317_CH35.qxd
10/15/07
3:12 PM
Page 373
373
This problem can also be phrased: I do not know the true Cpk, but I am 90 percent
confident that it is not less than _____.
C pk Za
C pk 2
(1.33)2
1
1
+
= 1.33 1.28
+
= 1.33 0.13 = 1.20
(9)(100) (2)(100) 2
9n 2 n 2
An appropriate statement would be: I do not know the true Cpk, but I am 90 percent
confident that it is not less than 1.20.
Determination of the appropriate sample size for a single-sided confidence limit for
Cpk can be determined given a level of confidence and a predesired amount of error.
The single-sided error for Cpk is given by
C pk 2
1
.
+
9 n 2n 2
Za
2 + 9C pk 2 2
18n + n 18 +
+ E 2 = 0,
E2
2
Z 2
Z
2
2 + 9C pk 2
b = 18 +
E2
Z2
c=
2
E2
Z2
b b 2 4 ac
.
2a
Notice that the b term contains the Cpk. A prior estimate for the Cpk may be used as a starting point.
H1317_CH35.qxd
374
10/15/07
3:12 PM
Page 374
C = 90%
C pk = 1.20
E = 0.166
Z = 1.28
a = 18
2
2 + 9C pk
b = 18 +
E2
Z
c=
n=
b = 907.35
2
= 118.9
E2
Z2
b b 2 4 ac
2a
Cr and Cp Indices
There are two indices that do not use the process average to determine a process capability;
rather, they measure the proportion of the specification tolerance that is being consumed
by normal variation. These index numbers are Cr (capability ratio), which is sometimes
reported as a %Cr, and Cp.
Normal variation is defined as six standard deviations.
%Cr is the percentage of the tolerance that is being consumed by normal variation.
Cp is the reciprocal of Cr:
Cp =
1
6S
6S
, Cr =
, %Cr =
100.
Cr
USL LSL
USL LSL
H1317_CH35.qxd
10/15/07
3:12 PM
Page 375
375
Cr and Cp require that the specification be bilateral (have both an upper and lower
limit).
Example 1:
The specification for a part is 12.00 0.38, and the standard deviation is S = 0.08. What
is the %Cr?
%Cr =
6S
100
USL LSL
%Cr =
(6)(0.08)
100
12.38 11.62
%Cr = 63.2%
Example 1:
A Cp of 2.32 has been calculated based on 29 observations. What is the 95 percent confidence interval for this estimate?
Cp = 2.32
n = 29
Confidence = 0.95
Risk, = 0.05
Look up the two required 2 values as follows:
For the upper bound, 2n1,/2 = 228.025 = 44.4607
For the lower bound, 2n1,1/2 = 228.0.975 = 15.3079
Cp
2 n 1, / 2
2 n 1,1 / 2
> True Cp > Cp
n 1
n 1
44.4607
15.3079
> True Cp>2.32
28
28
2.92 > Truee Cp > 1.72
2.332
H1317_CH35.qxd
376
10/15/07
3:12 PM
Page 376
The confidence interval for the %Cr can be readily determined by simply taking the
inverse of the confidence interval for the Cp and reversing the terms. That is, the lower
confidence limit for the Cp is the upper confidence limit for the Cr.
In the previous example, the 95 percent confidence lower bound was 1.72.
1
= 0.58 = upper confidence bound for the %Cr
1.72
and
1
= 0.34 = lower confidence bound for the %Cr
2.92
If tables of the chi-square distribution are not available, it is necessary to use approximate formulas. In the absence of such tables, one commonly used approximation is
2
v ,
1
Z
v +
,
2
2
v,
1 1.96
(28 +
2
2
2 28,0.025 = 43.96.
n 1
n 1, / 2
n 1
n 1,1 / 2
Cpm Index
One of the more recent quality process capability indices is Cpm. A discussion of this index
is reported by Chan, Cheng, and Spiring (July 1988).
H1317_CH35.qxd
10/15/07
3:12 PM
Page 377
377
The Cpm index takes into consideration the proximity to the target value of a specification, unlike the Cpk, which only relates the process variation and average to the nearest
specification limit.
C pm =
USL LSL
6
( Xi T )2
n 1
T = target or nominal
where: =
C pm =
(USL LSL)
6 + ( T )
2
=
Cr 1 +
( T 2
2
The specification is 15 + 6/5. The average X = 17.0, and the standard deviation S = 1.2.
What is Cpm?
Cr =
6S
= 0.65 C pm =
USL LSL
1
Cr 1 +
( T
2
=
0.65 1 +
(17.0 15.0)2
(1.2)2
The following illustrates the movement of Cpk and Cpm for a constant specification
requirement of 28.00 + 3/4, where the standard deviation remains a constant 0.43 and the
process average changes from 29 to 24. Cr will remain constant (Cr = 0.369).
H1317_CH35.qxd
10/15/07
378
3:12 PM
Page 378
Target = 28.0
Cpk = 2.33
Cpm = 1.07
22
23
24
25
26
27
28
LSL
29
30
31
32
USL
X = 29.0
S = 0.43
Cpk = 2.71
Cpm = 1.48
22
23
24
25
26
27
28
29
30
31
LSL
32
USL
X = 28.5
Cpk = 2.33
Cpm = 2.71
22
23
24
25
26
27
28
29
LSL
30
31
32
USL
X = 28.0
Target = 28.0
H1317_CH35.qxd
10/15/07
3:12 PM
Page 379
Cpk = 2.71
Cpm = 1.76
22
23
379
Target = 28.0
24
25
26
LSL
27
28
29
30
31
32
USL
X = 27.5
Cpk = 1.55
Cpm = 1.07
22
23
24
25
26
27
28
29
30
31
LSL
32
USL
X = 27.0
Cpk = 1.16
Cpm = 0.57
22
23
24
25
26
LSL
27
28
29
30
31
32
USL
X = 26.5
H1317_CH35.qxd
10/15/07
380
3:12 PM
Page 380
2
1 (C pk )
+
9n 2n 2
Z = Z 0.10 = 1.28
C pk = 1.75
1.28 = 0.10
n = 45
1
(1.75)2
+
=1.75 0.25 =1.50
9(45) 2(45) 2
An appropriate statement for this problem would be: I do not know the true Cpk, but
I am 90 percent confident that it is not less than 1.50. In other words, if you obtained a
Cpk of 1.75 with n = 45 you would be 90 percent confident that the true Cpk would not be
less than 1.50.
If we relabel the LCL as CpkD and the Cpk in the right side of our LCL equation as CpkR
and solve for CpkD, we can determine the Cpk we need to observe from sample data (CpkR)
in order to achieve a desired CpkD.
C pkD = C pkR Z
2
1 (C pkR )
+
9n 2n 2
1
3n ( 2 n + 2 + Z )
2
(6nC
+ 6 n C pkD
2
pkD
4 nZ 8 n Z + 2 nZ + 18 Z C pkD ( n n ) 2 n Z + 4 n Z
2
Or alternatively,
Z2 (9nC2pkD + 2n 2 Z2 )
2n 2
.
C pkR =
C
pkD
2
18n(n 1)
2n 2 Z
Example using the latter equation:
You want to achieve a Cpk of 1.33. With n = 30 samples and a level of confidence of
90 percent, what Cpk should you have?
CpkR= ?????
CpkD= 1.33
n = 30
and Z = 1.28
).
H1317_CH35.qxd
10/15/07
3:12 PM
Page 381
381
Z2 (9nC2pkD + 2n 2 Z2 )
2n 2
C pkR =
C
pkD
2
18n(n 1)
2n 2 Z
58
(1.28)2 (477.60 + 56.3616)
1
.
33
C pkR =
540(29)
56.3616
= (1.029)(1.33 0.236)
CpkR = 1.611 and 1.125
The CpkR must always be greater than the CpkD; therefore, the answer is 1.611.
If you want to be 90 percent confident that your Cpk is > 1.33 and you measure 30 parts
to get your calculated Cpk, you need to get a calculated Cpk of 1.61.
Table A.14 in the appendix gives CpkR to yield a CpkD for n = 30 to 1000 and levels of
confidence from 90 percent to 99 percent.
Bibliography
Bissel, A. F. 1990. How Reliable Is Your Capability Index? Journal of Applied Statistics
39: 331340.
Chan, L. K., S. W. Cheng, and F. A. Spiring. 1988. A New Measure of Process Capability: Cpm. Journal of Quality Technology 20, no. 3: 162175.
Chou, Y., D. D. Owen, and S. A. Borrego. 1990. Lower Confidence Limits on Process
Capability Indices. Journal of Quality Technology 22, no. 3: 98100.
Kushler, R. H., and P. Hurley. 1992. Confidence Bounds for Capability Indices. Journal
of Quality Technology 24, no. 4: 118195.
H1317_CH35.qxd
10/15/07
3:12 PM
Page 382
H1317_CH36.qxd
10/18/07
2:55 PM
Page 383
0 = Y 1 X
and
SS xy
,
1 =
SS x
( X i ) 2
where: SSx = (Xi X )2 = Xi2
n
Once 0 and 1 have been computed, we substitute their values into the equation of a
line to obtain the following least squares prediction equation:
Y = 0 + 1 X .
It should be noted that rounding errors can significantly affect the answer you obtain
in calculating the SSx and SSxy. It is recommended that calculations for all sum of squares
(SS) be carried to six or more significant figures in the calculation.
Example:
Data have been collected in hours for the life of a deburring tool as a function of the speed in
RPMs. Using the least squares method, what is the linear equation relating tool life to speed?
383
H1317_CH36.qxd
10/18/07
384
2:55 PM
Page 384
Data table:
Yi
Tool life, hours
Xi
Speed, rpm
Xi 2
Yi 2
XiYi
180
220
130
110
280
200
50
40
1350
1230
1380
1440
1200
1300
1500
1540
1,822,500
1,512,900
1,904,400
2,073,600
1,440,000
1,690,000
2,250,000
2,371,600
243,000
270,600
179,400
158,400
336,000
260,000
75,000
61,600
32,400
48,400
16,900
12,100
78,400
40,000
2,500
1,600
Sum 1,210
10,940
15,065,000
1,584,000
232,300
( X i ) 2
(10 , 940 ) 2
= 15, 065, 000
= 104 , 550
8
n
( X i )( Yi )
(10 , 940 )(1210 )
SS xy = X i Yi
= 1, 584 , 000
= 70 , 675
8
n
Y 1210
X i 10 , 940
Y= i =
= 151.25
X=
=
= 1367.5
8
8
n
n
SS x = X i 2
Hence,
SS xy
1 =
SS x
70, 675
1 =
= 0.676,
104, 550
Y t
H1317_CH36.qxd
10/18/07
2:55 PM
Page 385
where:
385
Example:
What is the 95 percent confidence interval for the estimated mean tool life with a speed of
1400 rpm using the least squares model of Y = 1075.67 0.676(X)?
Y = 1075.67 0.676(1400) = 129.27
C = 0.95; therefore /2 = 0.025
t
, n 2 = t0.025,6 = 2.447
n=8
n2=6
2
S = standard deviation of Y for a given value of X
S = SS y 1 SS xy
(Yi)2
(1210)2
= 49, 287.5
= 232, 300
8
n
70, 675
= 0.676
= 70, 675 and 1 =
104, 550
SS y = Yi2
S xy
Y t
129.27 (2.447)(38.87)
1 (1400 1367.5)2
+
= 129.27 34.96
8
104, 550
We are 95 percent confident that, based on the least squares model of Y = 1075.67
0.676(X) with a speed of 1400 rpm, the mean tool life will be between 94.31 hours and
164.23 hours.
H1317_CH36.qxd
386
10/18/07
2:55 PM
Page 386
1 1
,
S
1
where S 1 = the standard deviation of 1 (also called the standard error of 1) and
S =
1
S
,
SS x
S =
( 70, 675)2
104, 550
,
82
49, 287.5
S = 15.87,
S =
1
S
15.87
=
= 0.049.
SS x
104, 550
Assuming that the error variable is normally distributed, the test statistic t follows
students t = distribution with n 2 degrees of freedom.
H1317_CH36.qxd
10/18/07
2:55 PM
Page 387
387
Can we conclude at the 1 percent level of significance that tool wear and speed are linearly related?
We proceed as follows:
Ho: 1 = 0
Ha: 1 0
Decision rule:
Reject the null hypothesis Ho if
t > t / 2, n 1 = t 0.001, 7 = 3.499
t=
1 1
S
t=
0.676 0
= 13.795
0.049
We reject the null hypothesis and conclude that 1 is not equal to zero. A linear correlation exists.
Measuring the Strength of the Linear Relationship
In testing 1, the only question addressed was whether there was enough evidence to allow
us to conclude that a linear relationship exists. In many cases, it is useful to measure the
strength of the linear relationship. We can accomplish this by calculating the coefficient of
correlation, .
Its range is
1 +1.
Because is a population parameter, we must estimate from the sample data. The sample coefficient of correlation is abbreviated r and is defined as follows:
r=
r=
SS xy
(SS x )(SS y )
70, 675
,
(104, 550)(49, 287.5)
r = 0.985.
Testing the Coefficient of Correlation
If = 0, the values of x and y are uncorrelated, and the linear model is not appropriate. We
can test to determine if x and y are correlated by testing the following hypothesis:
Ho: = 0
Ha: 0
H1317_CH36.qxd
388
10/18/07
2:55 PM
Page 388
t=
where Sr =
1 r2
.
n2
r
,
Sr
This test statistic is valid only when testing = 0. Consequently, the test statistic is
simplified to
t=
r
.
Sr
t=
r
Sr
Decision rule:
Reject the null hypothesis if t > t/2,n1 = t0.001,7 = 3.499 (assuming that = 0.01).
Value of the test statistic:
Since r = 0.985,
Sr =
1 r2
,
n2
Sr =
1 0.970
= 0.0707,
82
t=
r
,
Sr
t=
0.9985
= 13.932.
0.0707
Conclusion
13.932 > 3.499; therefore, we reject Ho and conclude that 0.
H1317_CH36.qxd
10/18/07
2:55 PM
Page 389
389
There is sufficient evidence to conclude that x and y are correlated and, as a result, a
linear relationship exists between them.
The results are identical for the tests of 1 and . Because in both cases we are testing
to determine whether a linear association exists, the results must agree. In practice, only
one of the tests is required to establish whether a linear relationship exists.
Bibliography
Petruccelli, J. D., B. Nandram, and M. Chen. 1999. Applied Statistics for Engineers and
Scientists. Upper Saddle River, NJ: Prentice Hall.
Sternstein, M. 1996. Statistics. Hauppauge, NY: Barrons Educational Series.
Walpole, R. E., and R. H. Myers. 1993. Probability and Statistics for Engineers and Scientists. 5th edition. Englewood Cliffs, NJ: Prentice Hall.
H1317_CH36.qxd
10/18/07
2:55 PM
Page 390
H1317_CH37.qxd
10/15/07
3:20 PM
Page 391
Reliability
Reliability can be defined as the probability that a product, system, service, component,
and so on, will perform its intended function adequately for a specified period of time,
operating in a defined operating environment without failure. Components of this definition are explained as follows:
Probability of success
Durability
Dependability
Quality over time
Availability to perform a function
Typical reliability statements include This car is under warranty for 40,000 miles or
3 years, whichever comes first or This mower has a lifetime guarantee.
Concept
Quality = Does it perform its intended function?
Reliability = How long will it continue to do so?
The reliability or probability of mission success can be determined for individual components of a complex system, or they can be determined for the entire system.
Reliability can be described mathematically by one of several distribution functions,
including but not limited to:
1. Normal distribution
2. Exponential distribution
3. Weibull distribution
391
H1317_CH37.qxd
392
10/15/07
3:20 PM
Page 392
Hours at failure
600
1000
850
1100
960
Failure rate, =
Number of failures
5
=
= 0.0011
Total test time
4510
Mean time between failure (MTBF), : The average time or cycles a unit runs before
failure. MTBF can be written as a function of failure rate, 1/.
For the previous example,
MTBF =
4510
= 902 hours.
5
The failure rate for most things changes over time. The relationship of the failure rate
with respect to time is depicted by the bathtub curve.
Consider the following data giving the failure rates for an automatic dishwasher. The
failure rates were determined on dishwashers at different ages, ranging from one hour old
to 20 years old.
H1317_CH37.qxd
10/15/07
3:20 PM
Page 393
Reliability
393
At one hour old, the dishwasher has a failure rate that is relatively high at one failure
per 714 units, or a failure rate of 0.0014. After the units have reached an age of one day,
the failure rate decreases to one failure in 10,000, or a failure rate of 0.0001. By an age of
one month, the dishwasher has a failure rate that is one in 555,555, or 0.0000018.
During this initial period of the life of the dishwasher (the period from one hour to one
month), the failure rate steadily decreased. This period, where the failure rate is decreasing, is referred to as the burn-in period, or the period of infant mortality. Failures in this
area are typically due to poor workmanship or defective components and raw materials.
0.01
Failure rate at one hour old is 1/714
0.001
0.0001
0.00001
0.000001
0.0000001
1 hr.
1 mo.
1 day
4 mo.
1 yr.
5 yr.
10 yr.
15 yr.
20 yr. 24 yr.
Age
During the period from four months to about 12 years, the failure rate stabilized to a
constant failure rate of slightly less than one failure per million units1/1,000,000 or
0.000001.
This portion of the life of the dishwasher is called the useful life and is characterized
by a constant failure rate.
H1317_CH37.qxd
394
10/15/07
3:20 PM
Page 394
0.01
0.001
0.0001
0.00001
0.000001
0.0000001
1 hr.
1 mo.
1 day
4 mo.
1 yr.
5 yr.
10 yr.
15 yr.
20 yr. 24 yr.
Age
During the period from 12 years to 24 years, the failure rate increased. The group of
dishwashers that are 24 years old are failing at a rate of one in 100, or a failure rate of 0.01.
The area of increasing failure is the burnout period. It is during this time that individual
components are wearing out as a result of of metal fatigue, stress, oxidation, and corrosion.
0.01
Infant
mortality
Useful
life
Burnout
0.001
0.0001
0.00001
0.000001
0.0000001
1 hr.
1 mo.
1 day
4 mo.
1 yr.
5 yr.
Age
10 yr.
15 yr.
20 yr. 24 yr.
H1317_CH37.qxd
10/15/07
3:20 PM
Page 395
Reliability
395
Summary
The period life of a product where the failure rate is decreasing is defined as the infant
mortality or burn-in phase. Where the failure rate is constant defines the period of useful
life. Where the failure rate is increasing, the product is in a burnout phase.
When the product is in a state of constant failure, rate predictions regarding reliability
may be made using the exponential distribution or exponential model.
Exponential Distribution
Given a constant failure rate or MTBF , the probability of mission success or reliability based on the exponential model is given by
Rt = e
or
Rt = e t ,
t = 400
= 660
Rt = e 0.61
Rt = 0.54 or 54%
There is only a 54 percent chance that the motor will survive 400 hours of running
time.
Supplemental problem:
The failure rate during the useful life of a product is 0.00018. What is the probability that
the unit will survive 1500 hours of service?
Note: There is an assumption that the failure rate is calculated based on the same time
unit as the reliability calculation objective. That is, the failure rate of 0.00018 means that
one failure is occurring every 5556 hours.
Answer: 76 percent
Determination of Failure Rate, k
Failure rates are determined by testing units until failure. There are three modes of testing:
1. Complete data: All units are tested until all units fail.
2. Time-censored data (type I testing): Multiple units are started at the same time, and
the test is discontinued upon reaching a predetermined time.
3. Failure-censored data (type II testing): Multiple units are started, and the test is discontinued upon reaching a predetermined number of failures.
H1317_CH37.qxd
396
10/15/07
3:20 PM
Page 396
number of failures
.
total test time
1. Complete data
Eight electric motors are tested. All fail at the following times. What is the estimate for the
failure rate, ?
Failure times, hours: 900, 600, 1100, 1350, 1000, 750, 1800, 2890
=
8
10 , 390
= 0.00077
i =1
0
1
2
20
Time, hours
60
80
40
100
120
140
x
x
Unit
4
number
5
+
x
x = failure
+ = still running
H1317_CH37.qxd
10/15/07
3:20 PM
Page 397
Reliability
397
r
n
+ ( n r )T
i =1
3
( 28 + 49 + 110 ) + ( 7 3 )140
r
n
+ ( n r )T
i =1
H1317_CH37.qxd
10/15/07
398
3:20 PM
Page 398
Unit
5
4
7
3
2
1
8
9
6
Cycles on failure
250
800
1100
1800
4800
12,600
12,500
12,600
12,600
=
Status
F
F
F
F
F
F
+
+
+
+ ( n r )T
F = failed
+ = still running at termination
6
21, 350 + ( 3 )(12 , 600 )
= 0.000101
i =1
1
= 9901
0.000101
2 r + 2 ,
2
2
2r
2r
where: = estimated failure rate
r = number of failures
= risk ( = 1.00 confidence)
2 = chi-square.
Example:
Ten units have been tested for 250 days. At the end of the 250-day period, seven units fail.
What is the estimated failure rate, and what is the 90 percent confidence interval for the
estimate?
Data:
Unit no.
1
2
3
4
5
6
7
8
9
10
Day at failure
145
200
250+
188
250+
250+
220
120
227
150
7
= 0.0035
2000
H1317_CH37.qxd
10/15/07
3:20 PM
Page 399
Reliability
399
2
2
2 r ,1
2 r + 2, 2
2
2r
2r
where: r = 7
= 0.10
= 0.0035
2 2 r ,1
2 2 r + 2,
2
2
2r
2r
214,0.95
216,0.05
14
14
Looking up the appropriate chi-square value using a chi-square table, we get the following:
214 ,0.95 = 6.57
216,0.05 = 26.3
26.3
6.57
< True < 0.0035
0.0035
14
14
0.0164 < True < 0.0066.
An appropriate statement would be: We do not know the true failure rate, but we are
90 percent confident that it is between 0.0164 and 0.0066.
Confidence Interval for k from Failure-Censored (Type II) Data
Failure rates that are determined from reliability tests that are terminated when a specified number of failures has been reached are referred to as failure-censored or failuretruncated tests. The confidence interval for these types of data is determined using the
following relationship:
2
2
2 r ,1 2
2 r , 2
2 r
2r
Note that the only difference between this relationship and the time-censored confidence interval is in the upper limit equation, where the degrees of freedom for the chisquare is 2r for the failure-censored data and 2r + 2 for the time-censored data.
Once the confidence interval for has been obtained, the confidence interval for the
MTBF can be determined by simply taking the reciprocal of the limits for the failure and
reversing the order of the intervals. The MTBF resulting from taking the reciprocal of the
lower limit for the failure value yields the upper confidence limit for the MTBF.
Example:
A sample of eight panel bulbs is tested, and after five failures, the estimate for is determined to be 4.6 104 failures per hour. Determine the 95 percent confidence interval for
this failure-censored data and the corresponding interval for the MTBF.
H1317_CH37.qxd
400
10/15/07
3:20 PM
Page 400
where: T = total test time on all units when the test is terminated
r = total number of failed units.
Example:
A set of seven units is tested until all seven units fail. The total time when the seventh unit
fails is 6000 hours. Calculate the single-sided lower confidence limit at 95 percent confidence for the failure rate.
= 0.05
2 1 ,2 r
< True
2T
H1317_CH37.qxd
10/15/07
3:20 PM
Page 401
Reliability
401
Note that since we are asking for only the lower confidence limit, the risk is not divided
by 2. All of the risk is assigned to the single-sided confidence limit.
21,2r = 20.95,14 = 6.57
2 1 ,2 r
6.57
=
= 0.00055
2T
12 , 000
This lower 95 percent confidence limit for the failure rate is the reciprocal of the upper
95 percent confidence limit for the MTBF, .
The upper 95 percent confidence limit for the MTBF is 1818.
Success Run Theorem
There are some cases where no failures occur during a reliability demonstration test. For
those cases, a concept known as the success run theorem may be used to report the given
level of reliability. The relationship of the reliability, level of confidence, and number of
failure-free samples is given in the formula
R = (1 C)1/n,
where R is the reliability, C is the level of confidence, and n is the sample size or number
of failure-free test items.
Most often we use the success run theorem to determine the number of failure-free test
items required to demonstrate a minimum reliability at a given level of confidence. By
rearranging the original formula, we may solve for n.
n=
ln(1 C )
ln R
It is important to realize that time is not the criterion. Items are simply tested to a predetermined time. If the items operate to the prescribed time, success is demonstrated.
Sample application:
Success for an electric relay is defined as 150,000 cycles without failure. How many relays
must pass this test in order to demonstrate a reliability of 95 percent with a confidence of
90 percent?
ln(1 C )
ln R
ln(1 0.90 )
n=
ln 0.95
n=
n=
2.303
= 45
0.051
Table 1 may be used to determine the appropriate sample size for several levels of reliability and confidence. The intersection of the row where R = 0.95 with the column where
C = 0.90 gives a sample size of n = 45. Forty-five relays must successfully operate for
150,000 cycles to demonstrate a reliability of 95 percent with a level of confidence of 90
percent. We are 90 percent confident that the reliability is not less than 95 percent.
H1317_CH37.qxd
402
10/15/07
3:20 PM
Page 402
Confidence level, C
Minimum
reliability, R
0.80
0.80
0.85
0.90
0.95
0.99
0.995
0.999
8
10
16
32
161
322
1609
0.85
9
12
19
37
189
379
1897
0.90
11
15
22
45
230
460
2303
0.95
0.99
14
19
29
59
299
598
2995
21
29
44
90
459
919
4603
0.995
0.9999
24
33
51
104
528
1058
5296
31
43
66
135
688
1379
6905
Reliability Modeling
The reliability of more complex systems can be modeled using the reliability of individual components and block diagrams. Components can be diagramed using series and parallel configurations.
Series Systems
The system reliability for components that function in series is determined by the product
of the reliability of each component.
RS = R1 R2 R3 . . . RK
The system reliability for a series of components is always less reliable than the least
reliable component.
Using a reliability block diagram (RBD), calculate the reliability of the following
system:
R = 0.90
R = 0.95
R = 0.99
H1317_CH37.qxd
10/15/07
3:20 PM
Page 403
Reliability
403
If the failure rate or MTBF is given, the reliability must be determined in order to perform the system reliability model. Failure rates and MTBFs cannot be combined directly.
Example:
Find the system at t = 200 hours for a series system configuration given that one component has a failure rate of 0.002 and the other component has an MTBF of 700 hours.
Given = 700 hours, the reliability at t = 200 is
Rt = 200 = e
where: t = 200
= 700.
Rt = 200 = e ( 0.29 )
Rt = 200 = 0.75
Given = 0.002, the reliability at t = 200 is
Rt = 200 = e t ,
where: t = 200
= 0.002.
Rt = 200 = e ( 0.002 200 )
Rt = 200 = 0.67
The RBD can now be constructed for the individual components:
R = 0.75
R = 0.67
RS = 0.75 0.67,
RS = 0.50.
Parallel Systems
The reliability of systems that have components that function in parallel is determined by
subtracting the product of the component unreliabilities (1 R) from 1. The system reliability of a parallel system will always be more reliable than the most reliable component.
H1317_CH37.qxd
404
10/15/07
3:20 PM
Page 404
R1
R2
Rk
If all of the components are of equal reliability, then the system reliability is given by
RS = 1 (1 R)n,
where n = number of components in parallel.
Example:
What is the system reliability for three components having individual reliabilities of 0.85,
0.90, and 0.80?
R = 0.85
RS = 1 0.003
RS = 0.997
R = 0.80
Complex Systems
Block diagrams of more complex arrangements utilizing a mixture of both serial and parallel components can be developed. Parallel components are reduced to an equivalent single component and the completion of the reliability calculation based on a serial model.
H1317_CH37.qxd
10/15/07
3:20 PM
Page 405
Reliability
405
Example:
Determine the reliability of the following RBD:
R = 0.95
R = 0.90
R = 0.95
R = 0.88
Reduce the two parallel components to the equivalent of one component as follows:
R = 1 [(1 0.95)(1 0.88)],
R = 1 [(0.05)(0.12)],
R = 1 0.006,
R = 0.994.
Determine the system reliability as a series configuration:
RS = 0.90 0.994 0.95,
RS = 0.85.
Shared-Load Parallel Systems
In the previous calculation, it is assumed that only one of the parallel components is
required for success and that the failure of one of the parallel components does not
increase the load on the remaining parallel component. Problems where the failure rate of
the surviving component is different than it would have been in the active redundancy state
are more complex.
Shared-load problems with two parallel components where the failure rate changes
with the number of surviving components are determined by
21
Rt = e 2 1 +
21 2
t
( e 1 e 2 1 t ) ,
H1317_CH37.qxd
406
10/15/07
3:20 PM
Page 406
Example:
A mobile refrigeration unit is served by two cooling systems in parallel. Operational data
show that one unit will provide adequate cooling, but the additional stress on the surviving unit increases its failure rate. The failure rate of the single surviving unit 2 is 0.00025
per hour. When cooling is provided by both units operating, the individual failure rate for
each 1 is 0.000075 per hour. What is the 1000-hour reliability for this system?
21 t
Rt = e 2 1 t +
( e 2 e 2 1 t )
21 2
where: t = 1000
1 = 0.000075
2 = 0.00025
2(0.000075 )
( 0.00025 )(1000 ) ( 2 )( 0.000075 )(1000 )
e
Rt = e2 ( 0.000075 )(1000 ) +
)
(e
(2 )(0.000075 ) (0.00025 )
Rt = e0.15 + (1.5 )(e0.25 e0.75 )
Rt = 0.993
H1317_CH37.qxd
10/15/07
3:20 PM
Page 407
Reliability
407
Bibliography
Dovich, R. A. 1990. Reliability Statistics. Milwaukee, WI: ASQC Quality Press.
Krishnamoorthi, K. S. 1995. Reliability Methods for Engineers. Milwaukee, WI: ASQC
Quality Press.
OConnor, P. D. T. 1995. Practical Reliability Engineering. 3rd edition. New York: John
Wiley & Sons.
H1317_CH37.qxd
10/15/07
3:20 PM
Page 408
H1317_CH38.qxd
10/15/07
3:24 PM
Page 409
Sequential Simplex
Optimization
Theoretical Optimum
The intersection of the two broken lines in Figure 1 shows the location of the result of a single
experiment performed with the conditions or combinations of temperature (70C) and time (30
minutes) for a chemical reaction. The response for this one set of conditions is 68 percent, as
indicated by the contour of the experimental space, which is, in reality, unknown. The response
surface is unknown, and the shape can only be determined by a series of experiments. One
experiment, as is the case with Figure 1, gives no information regarding the response surface.
Shape is a differential quality, dy/dx. In order to define a shape or surface, there must
be a dx. This requires at least two different values for variable x. If we are to understand
how a factor affects a response, we must vary it.
100
Temperature (C)
90
80
100
90
70
80
70
60
60
50
40
30
50
10
20
30
40
50
Time (minutes)
60
70
80
H1317_CH38.qxd
10/15/07
410
3:24 PM
Page 410
100
Temperature (C)
90
80
100
90
70
80
70
60
60
50
40
30
50
10
20
30
40
50
Time (minutes)
60
70
80
H1317_CH38.qxd
10/15/07
3:24 PM
Page 411
411
100
Temperature (C)
90
80
70
100
90
80
70
60
60
50
40
30
50
10
20
30
40
50
60
70
80
Time (minutes)
H1317_CH38.qxd
412
10/15/07
3:24 PM
Page 412
3. A face is that part of the simplex that remains after one of the vertexes is removed.
4. A hyperface is the same as a face, but it is used when the simplex is of four dimensions or greater.
5. A centroid is the center of mass of all the vertexes in that simplex.
6. The centroid of the remaining hyperface is the center of mass of the vertexes that
remain when one vertex is removed from the simplex.
Simplex Calculations
Problem
Two variables will be considered for the optimization of a process. The response variable
will be percent yield. The two independent variables or factors will be time (X1) and temperature (X2). An initial simplex will be formed by three vertexes around an initial set of
conditions located at 60C and 20 minutes. At this operating condition, the yield is 36 percent. Figure 4 shows the worksheet for the first simplex. The three vertexes are labeled W
for the vertex giving the worst response, B for the best response, and N for the next best
response. When two factors are considered, there are always three vertexes describing the
simplex.
This initial simplex is defined by the following three vertexes. The yield for each of
these conditions is determined and ranked as best (B), worst (W), and next best (N).
Simplex 1 coordinates:
Time
20
23
17
Temperature
64
58
58
Yield
42
37
31
Rank
B
N
W
Factor
X1
X2
Response
Rank
Coordinates of
retained vertexes
20
64
42
23
58
37
43
122
P = /k
21.5
61
17
58
(P W )
4.5
R = P + (P W )
26
64
31
54
H1317_CH38.qxd
10/15/07
3:24 PM
Page 413
413
Having established an initial set of conditions, running the experiments, and ranking the
three responses according to best, next best, and worst, the reflected vertex is calculated
using the worksheet. This reflected vertex for simplex #1 is time (X1) = 26 minutes and
temperature (X2) = 64C. The result of this experiment yields 54 percent.
Of these current four responses, we discard the current worst of 31 percent and retain
the remaining vertexes. These responses are reranked B, N, and W, and the data are
recorded on the simplex #2 worksheet (Figure 5). The new reflected vertex is determined
and run. The resulting response is 50 percent.
Factor
X1
X2
Response
Rank
Coordinates of
retained vertexes
26
64
54
20
64
42
46
128
P = /k
23
64
23
58
(P W )
R = P + (P W )
23
70
37
50
Discard the current worst vertex, rerank the remaining three, and record on the worksheet
for simplex #3 (Figure 6).
Factor
X1
X2
Response
Rank
Coordinates of
retained vertexes
26
64
54
23
70
50
49
134
P = /k
24.5
67
20
64
(P W )
4.5
R = P + (P W )
29
70
42
66
H1317_CH38.qxd
414
10/15/07
3:24 PM
Page 414
The response for this reflected vertex is 66 percent. Discarding the current worst vertex
corresponding to the yield of 42 percent and reranking generates simplex #4 (Figure 7).
Factor
X1
X2
Response
Rank
Coordinates of
retained vertexes
29
70
66
26
64
54
55
134
P = /k
27.5
67
23
70
(P W )
4.5
R = P + (P W )
32
64
50
57
The response for the reflected vertex of 32 minutes and 64C is 57 percent. Discard the
current worst, rerank, and calculate the reflected vertex for simplex #5 (Figure 8).
Factor
X1
X2
Response
Rank
Coordinates of
retained vertexes
29
70
66
32
64
57
61
134
P = /k
30.5
67
26
64
(P W )
4.5
R = P + (P W )
35
70
54
80
H1317_CH38.qxd
10/15/07
3:24 PM
Page 415
415
Factor
X1
X2
Response
Rank
Coordinates of
retained vertexes
35
70
80
29
70
66
64
140
P = /k
32
70
32
64
(P W )
R = P + (P W )
32
76
57
80
Discarding the current worst and reranking will establish simplex #7 (Figure 10). The
reflected vertex generated in simplex #6 gave a yield of 80 percent, the same as the original best for simplex #6. The choice for the vertex for the best and next best does not matter as long as both of the 80 percent yields are used.
Factor
X1
X2
Response
Rank
Coordinates of
retained vertexes
35
70
80
32
76
80
67
146
P = /k
33.5
73
29
70
(P W )
4.5
R = P + (P W )
38
76
66
98
In a similar manner of rejecting the current worst and reranking the remaining three, we
generate the remaining simplexes 8, 9, and 10. See Figures 11, 12, and 13.
H1317_CH38.qxd
416
10/15/07
3:24 PM
Page 416
Factor
X1
X2
Response
Rank
Coordinates of
retained vertexes
38
76
98
32
76
80
70
152
P = /k
35
76
35
70
(P W )
R = P + (P W )
35
82
89
X1
X2
Response
Rank
Coordinates of
retained vertexes
38
76
98
35
82
89
73
158
P = /k
36.5
79
32
76
(P W )
4.5
R = P + (P W )
41
82
80
Factor
80
95
H1317_CH38.qxd
10/15/07
3:24 PM
Page 417
417
Factor
X1
X2
Response
Rank
Coordinates of
retained vertexes
38
76
98
41
82
95
79
158
P = /k
39.5
79
35
82
(P W )
4.5
R = P + (P W )
44
76
89
100
The resulting vertex for simplex #10 gives a yield of 100 percent, and the optimization is
complete. See the following figure for the complete map of the 10 simplexes (or simplices).
100
Temperature (C)
90
10
80
9
7
70
3
60
100
8
90
80
70
60
50
40
30
50
10
20
30
40
Time (minutes)
50
60
70
80
H1317_CH38.qxd
418
10/15/07
3:24 PM
Page 418
See the following table for a summary of the simplex movements toward optimization:
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
Vertex
X1 (time)
X2 (temp.)
Initial
Initial
Initial
Reflected
Reflected
Reflected
Reflected
Reflected
Reflected
Reflected
Reflected
Reflected
Reflected
20
23
17
26
23
29
32
35
32
38
35
41
44
64
58
58
64
70
70
64
70
76
76
82
82
76
42 Best
37 Next best
31 Worst
54
50
66
57
80
80
98
89
95
100
Bibliography
Walters, F. H., L. R. Parker, Jr., S. L. Morgan, and S. N. Deming. 1991. Sequential Simplex Optimization. Boca Raton, FL: CRC Press.
H1317_CH39.qxd
10/15/07
4:49 PM
Page 419
Sequential Simplex
Optimization, Variable Size
The basic simplex optimization utilizes a simplex that is of a fixed size throughout the
optimization process; that is, the size of the simplex is not allowed to change, regardless
of the degree of improvement. Variable size simplexes (or simplices) allow the simplex to
expand or contract as a function of the amount of improvement. A simplex is a geometric
figure made up of coordinates that define a vertex.
For two variables, the simplex is a triangle made up of three vertexes (or vertices). For
three variables, the simplex is a tetrahedron. In both cases, we determine the response for
all vertexes, reject the poorest response, and carry the remaining vertexes and a generated
new vertex to the next simplex. This is continued until the goal or objective is reached or
until no further improvement is possible.
Traditionally, there has not been a mechanism by which the rankings have been tested
to see if a statistically significant improvement has been made. The differences observed
might simply be error in the experiment.
Consider the following example, which demonstrates the concept of a variable sample
size simplex and a method to evaluate statistical significance of the improvement.
A clear resin has been developed as a protective coating for biological specimens. The
resin is sprayed on and cured for a length of time at a specified temperature. The clarity,
which is measured by a bioplasmic reversing polarizer, is the most important. The greater
the number, the better the clarity. The current procedure calls for a temperature (X1) of 20
and a curing time (X2) of 20 minutes. The resulting clarity index is 51. A clarity index of
>85 is desirable. The best competitor claims an index of 75.
We begin by altering the conditions around the current practice to form a triangle. This
is simplex I. We want to alter the conditions enough to bring out the effects of the variable
but not enough to blow up the plant! We will evaluate the responses from our initial simplex (I), which is defined by the following coordinates:
Vertex
number
X1
X2
Responses
Average
Standard
deviation
Rank
1
2
3
10
30
20
20
20
30
50.0, 46.0
54.3, 55.7
57.0, 58.0
48.0
55.0
57.5
2.83
0.99
0.71
W
N
B
419
H1317_CH39.qxd
420
10/15/07
4:49 PM
Page 420
We have ranked the responses for the three vertexes according to best (B), next best
(N), and worst (W).
The question is, Is there a statistically significant difference between the B and N averages? We will answer this question by performing a simple t-test. If the calculated t-score
is greater than the critical t-score, the difference is statistically significant. For sample
averages derived from two observations, the t-calculated value is given by
tcalc =
XB XN
S2
100
50
X2, time
80
60
70
60
60
40
50
90
20
40
0
0
20
40
60
X1, temperature
Figure 1 Simplex I on surface response map.
80
100
120
H1317_CH39.qxd
10/15/07
4:49 PM
Page 421
421
Having ranked our vertices and tested for statistical significance in the difference
between the B and the N, we now transfer the data to the simplex worksheet. The average
of the coordinates for the B and N ranks is labeled as P . We calculate the P W values,
where W represents the coordinates for the W ranked vertex. The reflected R coordinates
are calculated as R = P + ( P W ). The R vertex is evaluated and recorded on the first
worksheet.
Factor
Simplex no. I
X1
X2
Response
Rank
20
30
57.5
30
20
55
25
25
10
20
48
15
40
30
65
Cw
Cr
55
35
80
We now need to review the responses for the B, N, R, and W to determine our next
step. The W from the current worksheet is not carried forward to the next worksheet.
Referring to Figure 2:
Is R > B? . . . Yes; evaluate the extended reflection (E)
E = R + (P W )
H1317_CH39.qxd
422
10/15/07
4:49 PM
Page 422
Initial
simplex
Continue experimentation
Rank trials as
B, N, and W
Evaluate
R
R>B
No
No
R>N
Yes
Yes
No
R>W
Yes
Evaluate
E
No
E>R
Yes
Use simplex
B..NE
Use simplex
B..NR
Use simplex
B..NCr
Have
objectives
been met?
Use simplex
B..NCw
No
Yes
Discontinue
Figure 2 Simplex decision map.
From the current initial simplex worksheet, the B, N, and E coordinates will be carried forward. They will be reranked, and the W from this worksheet will not be carried forward. Remember, a W value is never carried forward.
H1317_CH39.qxd
10/15/07
4:49 PM
Page 423
120
100
50
80
60
X2, time
70
60
60
40
50
90
20
90
80
40
0
0
20
40
60
80
100
120
X1, temperature
Figure 3 Surface response map showing the first two simplexes, I and 1.
Factor
Simplex no. I
X1
X2
Response
Rank
55
35
80
20
30
57.5
37.5
32.5
30
20
55
7.5
12.5
45
45
75
Cw
Cr
E
See Figure 4 for the responses after generating the second simplex.
423
H1317_CH39.qxd
424
10/15/07
4:49 PM
Page 424
120
100
50
80
60
X2, time
70
60
60
2
40
50
90
20
40
0
0
20
40
60
80
100
120
X1, temperature
Figure 4 Surface response map for simplexes I, 1, and 2.
Is R > N? Yes
H1317_CH39.qxd
10/15/07
4:49 PM
Page 425
Factor
Simplex no. 2
X1
X2
Response
Rank
55
35
80
45
45
75
50
40
20
30
57.5
30
10
80
50
82
Cw
Cr
110
Is R > B? Yes
Evaluate E
60
Is E > R? No
65
Factor
Simplex no. 3
X1
X2
Response
Rank
80
50
82
55
35
80
67.5
42.5
45
45
75
22.5
2.5
90
40
87
Cw
Cr
112.5
37.5
79
425
H1317_CH39.qxd
426
10/15/07
4:49 PM
Page 426
Is R > B? Yes
Evaluate E
Is E > R? No
Factor
Simplex no. 4
X1
X2
Response
Rank
90
40
87
80
50
82
85
45
55
35
80
30
10
115
55
65
15
70
40
90
Cw
Cr
E
Is R > B? No
Is R > N? No
Is R > W? No
The use of the Cw (contracted toward the W vertex) indicates that an improvement can
be found midway between the W vertex and the midpoint of the line described by vertices
N and B.
The generalized relationship of the various expansions and contractions can be seen in
the following diagram:
X2
N
Cr
Cw
W
X1
H1317_CH39.qxd
10/15/07
4:49 PM
Page 427
427
Factor
Simplex no. 5
X1
X2
Response
Rank
70
40
90
90
40
87
80
40
80
50
82
10
80
30
90
Cw
Cr
80
Is R > B? No
R=B
20
82
The completed surface response map can be seen in Figure 5, which shows all the simplexes, I through 5.
H1317_CH39.qxd
428
10/15/07
4:49 PM
Page 428
120
100
50
80
60
X2, time
70
60
60
5
3
40
50
90
20
80
90
40
0
0
20
40
60
X1, temperature
Figure 5 Completed simplex with response contours.
80
100
120
H1317_CH40.qxd
10/18/07
12:35 PM
Page 429
Short-Run Attribute
Control Chart
Traditional control charts for attributes include the p chart and np chart for defectives
where the sample size is variable and constant, respectively, and the u chart and c chart for
defects where the sample size is variable and constant, respectively.
In all of these cases, it is assumed that the process average is unique to a particular
process or product and is used to provide a historical basis for detecting a process change.
For example, if the process is one that photocopies 8.5 11 paper on one side where the
typical run size is 100 sheets, the average number of defects u might be 0.02. This information could be used to construct a u chart.
UCL = U + 3
U
n
LCL = U 3
and
U
n
Xi X
,
U U
U
n
429
H1317_CH40.qxd
430
10/18/07
12:35 PM
Page 430
The U value is unique to a specific product. The U value for 8.5 11 paper was
0.02, and for two-sided 11 17 paper, the U value might be 0.12.
If the U value is known for each product, then the Z-scores may be determined and
plotted on a universal chart. Since the objective of all control charts is to determine
whether the process average changes relative to the historical process average or target
process average, we will refer to the process average as the target average or Ut.
Development of the short-run u chart is as follows:
LCL < U < UCL
U 3
3
U
<U <U +3
n
U
n
U
U
< U U < +3
n
n
3 <
U U
U
n
< +3.
U
n
Each U has a unique value of Ut. Ut may come from:
1. Previous control charts for that product or item
2. An assumed level of defects/unit
Case Study
The Ionic Corporation manufactures depth-finding units and has several models that are
produced in small lots. The product line is broad, ranging from relatively simple units for
H1317_CH40.qxd
10/18/07
12:35 PM
Page 431
431
consumer use to sophisticated commercial units that are much more complex. Data from
the final inspection audits for four production models give the following historical process
average defects/unit Ut.
Part
A
B
C
D
Target Ut
0.20
0.10
0.13
0.17
As a quality engineer, you would like to establish a u chart using the short-run technique. You will use a short-run method because there is no real long run of any specific
model, as would be required for a traditional u chart, and the process for manufacturing is
more or less the same. Only the complexity of the unit produced changes.
Step 1. Collect historical data.
All of the units produced are inspected and tested to a rigorous format. The sample size
represents 100 percent of the units produced. The following information is available for
the past 13 production runs:
Date
Product
Sample size n
Number of defects c
11/2
11/3
11/4
11/4
11/5
11/5
11/6
11/6
11/7
11/8
11/8
11/9
11/9
A
C
D
A
C
D
A
D
D
B
C
A
B
18
23
7
11
38
16
7
23
11
42
17
13
26
5
2
1
2
4
3
0
2
3
2
3
1
2
U Ut
Ut
n
where:
n = sample size
U t = target average defects/uniit for specific part or model
H1317_CH40.qxd
432
10/18/07
12:35 PM
Page 432
First sample:
Part = A
U t = 0.20
n = 18
c=5
U=
Z=
c 5
=
= 0.28
n 18
U Ut
Z=
Ut
n
0.28 0.20
= +1.28
0.28
18
Second sample:
Part = C
U t = 0.13
n = 23
c=2
U=
Z=
c 2
=
= 0.09
n 23
U Ut
Ut
n
Z=
0.09 0.13
= 0.53
0.13
23
Date
Product
Sample
size n
Number of
defects c
Defects per
unit U
Z-score
11/2
11/3
11/4
11/4
11/5
11/5
11/6
11/6
11/7
11/8
11/8
11/9
11/9
A
C
D
A
C
D
A
D
D
B
C
A
B
18
23
7
11
38
16
7
23
11
42
17
13
26
5
2
1
2
4
3
0
2
3
2
3
1
2
0.28
0.09
0.14
0.18
0.11
0.19
0.00
0.09
0.27
0.05
0.28
0.08
0.08
+1.32
0.53
+0.16
0.15
0.34
+0.19
1.18
0.93
+0.80
1.02
+1.72
+0.97
0.32
H1317_CH40.qxd
10/18/07
12:35 PM
Page 433
433
+1
x
x
2
3
Product:
Date:
A C D A C D A D D B C A B
11/2 11/4 11/5 11/6 11/7 11/8 11/9
Step 4. Continue to collect and plot data, looking for evidence of a process change.
Date
Product
Sample
size n
Number of
defects c
Defects per
unit U
Z-score
11/10
11/10
11/11
11/11
11/12
11/12
11/13
11/14
11/14
C
A
B
C
D
B
A
B
C
40
19
23
15
30
25
9
18
22
5
8
6
3
8
2
3
4
9
0.13
0.42
0.26
0.20
0.27
0.08
0.33
0.22
0.41
0.00
+2.14
+2.43
+0.75
+1.33
0.32
+0.87
+1.61
+3.64
History
+3
+2
+1
0
1
x
x
x
x
x
x
x
x
2
3
Product:
Date:
A C D A C D A D D B C A B C A B C D B A
11/2 11/4 11/5 11/6 11/7 11/8 11/9 11/10 11/11 11/12
B C
11/14
A process change is indicated with sample #22 occurring on 11/14 with product C.
H1317_CH40.qxd
10/18/07
434
12:35 PM
Page 434
Short-Run p Chart
The same principle used to develop the u chart is used for the p chart, only the standard
deviation calculation changes. The calculation for the plotting characteristic Z using the
appropriate estimate for the standard deviation is given by
Z=
P Pt
Pt (1 Pt )
n
A value for the target process average proportion Pt must be given for each product.
Example:
A manufacturer of noninterruptible power supplies has an operation that makes wiring harnesses that are used on its product lines. There are several different styles of harnesses, and
the department manager wants to use a short-run p chart to monitor the process. In some
cases, the lot size for orders may be as high as 100, but there are many runs of different
styles. The harness production line consists of four workers who normally work in other
departments. This production is assembled once or twice a week to produce harnesses that
will be used for the next production period. The manager wants only one chart to monitor
the department performance. Given the following information, develop the short-run p chart.
Model number
A
B
C
D
E
Process average
proportional defective, Pt
0.03
0.07
0.11
0.05
0.09
Production data:
Date
Product
Lot size
Number defective
Proportional defective P
4/5
4/5
4/5
4/11
4/11
4/11
4/11
4/15
4/15
4/15
4/15
4/19
4/19
A
D
B
C
B
E
D
E
C
B
D
C
E
45
123
88
28
35
58
75
95
42
65
27
55
40
1
8
4
2
3
3
1
11
7
5
1
1
4
0.022
0.065
0.045
0.071
0.086
0.052
0.013
0.116
0.167
0.077
0.037
0.018
0.100
H1317_CH40.qxd
10/18/07
12:35 PM
Page 435
435
Sample #2
P Pt
Z=
Pt (1 Pt )
n
0.022 0.020 t
= 0.31
0.03(1 0.03 )
45
Z=
0.022 0.020
= 0.76
0.03(1 0.03 )
45
Z=
Continue to calculate the Z-scores for the remaining data and record.
Date
Product
Lot size
Number defective
Proportional defective P
ZP
4/5
4/5
4/5
4/11
4/11
4/11
4/11
4/15
4/15
4/15
4/15
4/19
4/19
A
D
B
C
B
E
D
E
C
B
D
C
E
45
123
88
28
35
58
75
95
42
65
27
55
40
1
8
4
2
3
3
1
11
7
5
1
1
4
0.022
0.065
0.045
0.071
0.086
0.052
0.013
0.116
0.167
0.077
0.037
0.018
0.100
0.31
0.76
1.08
0.66
0.37
1.01
1.47
0.89
1.18
0.22
0.31
2.18
0.22
History
+3
+2
+1
0
1
2
x
x
x
x
x
x
x
x
3
Product:
Date:
A D B C B E D E C B D C E
4/5
4/5
4/11 4/11 4/15 4/15 4/19
H1317_CH40.qxd
436
10/18/07
12:35 PM
Page 436
Continue to collect data, plot points, and look for signs of a process change:
Date
Product
Lot size
Number defective
Proportional defective P
ZP
4/20
4/20
4/21
4/22
4/23
B
E
A
C
B
85
30
75
32
48
10
4
7
3
4
0.118
0.133
0.093
0.094
0.083
+1.73
+0.82
+3.20
0.29
+0.35
History
+3
+2
+1
0
1
2
x
x
x
x
x
x
x
x
x
x
3
Product:
Date:
A D B C B E D E C B D C E B E A C B
4/5
4/5
4/11 4/11 4/15 4/15 4/19 4/20 4/22
Completed short-run p chart
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Wheeler, D. J. 1991. Short Run SPC. Knoxville, TN: SPC Press.
H1317_CH41.qxd
10/15/07
3:48 PM
Page 437
Short-Run Average/Range
Control Chart
The traditional implementation of statistical process control (SPC) requires that a minimal
number of sample observations be obtained to characterize the process or set up a historical reference against which future changes in the process are monitored. The recommended number of subgroups required for such historical characterization varies. For the
average/range control chart, a minimum of 25 subgroups is suggested, each of a subgroup
size of four or five. In many manufacturing processes, the opportunity to measure 125
units of production would be impractical, if not impossible, due to low production volumes
in a just-in-time operation.
By consolidating the measurements of a universal characteristic that are common to
several different products into one population, we may monitor this characteristic as if it
originated from a single process.
The concept for short-run average/range charts is based on a normalization of the statistic we will use to monitor changes in the location (central tendency) and variation. We
will be measuring the amount of change for a specific characteristic relative to the
expected value for that characteristic. This will be done for several different products, all
having the same characteristic. Each of the products, however, will have a unique expected
value for both the location statistic and the variation statistic. These historical expected
values are established with historical data.
The following derivations are made to generate the new short-run plotting characteristic that will replace the traditional average and range used in the X /R chart.
H1317_CH41.qxd
438
10/15/07
3:48 PM
Page 438
XX
< + A2
R
Rather than plotting the average X , as in the case of a normal average/range chart, we
will plot
XX ,
R
where X and R will have unique values for each product.
The control limits will be based on A2 for this normalized chart.
Control limits for X X are UCL = +A2 and LCL = A2.
R
The value for A2 depends on the subgroup sample size.
R D4 R
<
R
R
R
R
D3 < < D 4
R
<
R
Rather than plotting the range R, we will plot . There will be a unique R value for
R
each product.
Control limits are UCL = D4 and LCL = D3.
Each of the plotting parameters requires a unique value for the X and R for each part
or unit of inspection. The following sections explain the several sources for these values.
H1317_CH41.qxd
10/15/07
3:48 PM
Page 439
439
For X
1. Use X from the previous control chart data
2. Calculate the average from any available data
3. Let the nominal = X
Note: If the nominal is the choice for an estimate of the average, there is a requirment. The
absolute value for the difference in a sample average and the nominal times the sample
standard deviation must be less than a cricital value:
Cd = FcS,
where: Cd = a factor dependent on the sample size m from which the standard deviation was calculated
S = the sample standard deviation.
If Nominal Average FcS*, the average of m observations should be used rather
than the nominal where Fc is a factor determined by the sample size m from which S, the
sample standard deviation, was determined.
Fc factors:
Sample size m
4
5
6
7
8
9
10
11
Fc
Sample size m
Fc
1.1765
0.9535
0.8226
0.7344
0.6699
0.6200
0.5796
0.5463
12
13
14
15
16
17
18
19
0.5185
0.4942
0.4733
0.4547
0.4383
0.4235
0.4101
0.3978
Example 1:
Five measurements are available, and the specification nominal is 10.000. Should we use
the nominal or the average of the five measurements for the target X ?
Data:
10.101, 10.008, 9.993, 10.102, 10.007
The sample average X = 10.042, and the sample standard deviation S = 0.054
m = 5 and the table value for Fc = 0.9535.
Critical difference Cd = FcS = 0.9535 0.054 = 0.0515
The absolute difference between the nominal and the sample average is
10.000 10.042 = 0.0042
H1317_CH41.qxd
440
10/15/07
3:48 PM
Page 440
Since 0.0042 is not greater than the critical difference of 0.0515, we can use the nominal in place of the average of the five measurements.
Rule: If the difference between the average and the nominal is greater than or equal to the
critical difference, use average rather than nominal for the target X .
For R
1. Use R .
2. Calculate R from an assumed Cpk:
R=
where:
(USL Xt )d2
3Cpk
or R =
( Xt LSL )d2
.
3Cpk
R
d2
and
S
.
c4
Therefore,
R S
= ,
d 2 c4
d
R = 2 S.
c4
The value for d2 depends on the selected subgroup sample size. The value for c4
depends on the sample size n from which the sample standard deviation S is calculated.
The values for d2 and c4 can be found in Table A.7 in the appendix.
Case Study
A product is routinely sampled and brought into a testing laboratory for evaluation. The
products are all different models, but the characteristic being measured is common to all
products. Your objective is to establish a short-run control chart. The production sched-
H1317_CH41.qxd
10/15/07
3:48 PM
Page 441
441
ule consists of several product changes and is relatively short for a given product design.
You will be working with three different product models. The subgroup has a sample size
of n = 3.
Model A:
This is a new design, and a temporary specification of 30 24 has been issued. The
expected process capability index (Cpk) is 1.00.
Model B:
The specification is 150 20, and you have test data for five units tested last month.
Model C:
Another new model is added with an expected Cpk of 1.20 and a specification of
400 20.
Step 1. Establish working values for target X and R for each model.
Model A
XA determination:
Based on an expected Cpk = 1.00 and a specification requirement of 30 24, the expected
target average X will be set at the nominal.
X A = 30.0
R A determination:
Use an expected Cpk = 1.00.
We are using a subgroup sample size of n = 3; therefore, d2 = 1.693.
R=
R=
R=
81.264
6.00
R = 13.5
Model B
XB determination:
Five measurements have been obtained and yield an average of X = 160.56 and a sample
standard deviation of S = 4.79. You may consider using the specification nominal of 150.0
for the target X , but you must first test to see if this is statistically acceptable. If it is not,
you will use the average of the five available data points instead of the nominal.
Critical difference, Cd = FcS = (0.9535)(4.79) = 4.57
H1317_CH41.qxd
442
10/15/07
3:48 PM
Page 442
If the absolute difference of the average of the five measurements exceeds the critical difference Cd, then use the average of the five measurements rather than the nominal for X B.
Difference = Nominal Average = 150.0 160.56 = 10.6
10.6 > 4.57; therefore, use X B = 160.56.
RB determination:
The standard deviation S from the five measurements will be used to estimate the expected
range for model B when subgroups of n = 3 are taken.
d2 = 1.693 for subgroup size n = 3
c4 = 0.9400 for a sample standard deviation S calculated from five measurements
S = 4.79
d
R = 2 S
c4
1.693
RB =
4.79
0.9400
R B = 8.63
Model C
Xc determination:
The specification is 400 20, and no prior data are available. Therefore, we will set the
target X to equal the nominal of 400:
XC = 400.0.
Rc determination:
Using the specification of 400 20, the expected Cpk = 1.20, and the subgroup sample size
of n = 3:
R=
( USL LSL)d 2
(420 380)1.693
67.72
, R=
, R=
, R = 9.41.
6C pk
(6)(1..20)
7.20
X
30.00
160.56
400.00
R
13.50
8.63
9.41
These values are recorded in the upper-left portion of the short-run X /R chart form.
H1317_CH41.qxd
10/15/07
3:48 PM
Page 443
443
Sample 1
A
Sample 2
A
Sample 3
A
Sample 4
B
Sample 5
B
Measurement
(1)
(2)
(3)
37.1
37.3
30.1
41.9
43.1
26.1
44.4
34.1
33.1
161.6
157.9
158.4
163.4
160.3
161.7
Average
Range
34.8
7.2
37.0
17.0
37.2
11.3
159.3
3.7
161.8
3.1
Model
Sample 6
B
Sample 7
B
Sample 8
B
Sample 9
C
Sample 10
C
Model
Measurement
(1)
(2)
(3)
163.0
161.7
160.0
161.2
154.2
162.4
153.9
161.9
156.3
411.0
400.0
405.0
395.0
399.0
416.0
Average
Range
161.6
3.0
159.3
8.2
157.4
8.0
405.3
11.0
403.3
21.0
Sample 11
A
Sample 12
B
Sample 13
B
Measurement
(1)
(2)
(3)
44.9
37.6
35.4
165.8
161.7
159.2
161.8
157.6
161.9
Average
Range
39.3
9.5
162.2
6.6
160.4
4.3
Model
H1317_CH41.qxd
444
10/15/07
3:48 PM
Page 444
Step 3. Calculate the average, range, average plot point, and range plot point, and plot the
points on the chart.
Subgroup 1: Model A
The plotting points are calculated, recorded on the chart form, and plotted.
Measurements:
37.1
37.3
30.1
Average X :
34.8
Range R:
7.2
41.9
43.1
26.1
Average X :
37.0
Range R:
17.0
H1317_CH41.qxd
10/15/07
3:48 PM
Page 445
445
Subgroup 3: Model A
Calculate, plot, and connect to subgroup point 3.
Measurements:
44.4
34.1
33.1
Average X :
Range R :
37.2
11.3
161.6
157.9
158.4
Average X :
Range R :
159.3
3.7
H1317_CH41.qxd
10/15/07
446
3:48 PM
Page 446
See the following completed control chart with all of the historical data points plotted.
(X Xt)/Rt
History
UCL = +1.02
+1.0
x
x
x
x
1.0
LCL = 1.02
8 9 10 11 12 13 14 15 16 17 18 19
Sample number
2.57
UCL = 2.57
2.0
R/Rt
x
1.0
x
x
1
x
x
x
2
x
5
x
x
9 10 11 12 13 14 15 16 17 18 19
Step 5. Continue to collect data into the future, looking for signs of a process change.
During the next few days, additional data are collected. Based on these observations, do
you feel that a process change has occurred? If so, was there a change in the average or in
the variation?
H1317_CH41.qxd
10/15/07
3:48 PM
Page 447
447
Sample 14
B
Sample 15
A
Sample 16
A
Sample 17
B
Measurement
(1)
(2)
(3)
150.3
159.4
53.7
26.4
29.2
23.0
26.0
21.0
28.2
149.0
158.2
152.0
390.4
388.7
386.9
Average
Range
154.5
9.1
26.2
6.2
25.1
7.2
153.1
9.2
388.7
3.5
Model
Sample 19
C
Model
Measurement
(1)
(2)
(3)
394.9
392.3
390.2
Average
Range
384.9
9.4
Sample 18
C
Subgroup number
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
Model
Average X
Range R
A
A
A
B
B
B
B
B
C
C
A
B
B
B
A
A
B
C
C
34.8
37.0
37.2
159.3
161.8
161.6
159.3
157.4
405.3
403.3
39.3
162.2
160.4
154.5
26.2
25.1
153.1
388.7
384.9
7.2
17.0
11.3
3.7
3.1
3.0
8.2
8.0
11.0
21.0
9.5
6.6
4.3
9.1
6.2
7.2
9.2
3.5
9.4
X XT
RT
R
RT
0.36
0.52
0.53
0.15
0.14
0.14
0.15
0.37
0.56
0.35
0.69
0.19
0.02
0.70
0.28
0.36
0.86
1.20
1.60
0.53
1.26
0.84
0.43
0.36
0.35
0.95
0.93
1.17
2.23
0.70
0.76
0.50
1.05
0.46
0.53
1.07
0.37
0.99
H1317_CH41.qxd
10/15/07
448
3:48 PM
Page 448
(X Xt)/Rt
History
UCL = +1.02
+1.0
x
x
x
x
1.0
x
x
LCL = 1.02
x
1
9 10 11 12 13 14 15 16 17 18 19
2.57
UCL = 2.57
2.0
R/Rt
x
1.0
x
x
1
x
x
x
2
x
5
x
x
9 10 11 12 13 14 15 16 17 18 19
The control chart indicates an out-of-control condition beginning with sample 18.
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Wheeler, D. J. 1991. Short Run SPC. Knoxville, TN: SPC Press.
H1317_CH42.qxd
10/15/07
3:50 PM
Page 449
Short-Run Individual/
Moving Range Control Chart
Individual/moving range control charts are useful when the opportunity for data is limited
and the data are from a family of products that have the same characteristic but the target,
tolerance, or the average and standard deviation for different products varies. Examples of
this might include the angle of a mounting bracket for several different brackets. One
model might have a mounting angle of 55, and another might have a requirement of 60.
Another example might be a batch manufacturing process for the manufacture of hot
melt adhesive. The measured characteristic will be the melt index. Specifications for the
requirements of four products are as follows:
Product
A
B
C
D
Melt index
2575
88120
1836
133150
deviations are calculated from 3S = 2.66 MR , where MR is the average moving range.
449
H1317_CH42.qxd
10/15/07
450
3:50 PM
Page 450
XX
< + 2.66
MR
XX
Using this transformation, the new plot point for the individuals will be X
, and
MR
the new upper and lower control limits for this plot point will be 2.66.
For each product, there will be a unique average and average moving range. This
unique average and moving range will be identified as the target average XT and the target
moving range MR T.
X1 X T X 2 X T
.
MR T
MR T
The traditional UCL for the moving range is given by UCLR = 3.268MR . Therefore, any
given moving range should be less than the UCL:
Dividing by MR , we have
3.268 >
MR
.
MR
H1317_CH42.qxd
10/15/07
3:50 PM
Page 451
451
But since the individual plot points have already been divided by an expected or target
average moving range MR T, we do not have to divide again. The plot point for the shortrun moving range simply becomes the moving range MR, and the UCL becomes 3.268
with an average moving range of 1.00.
Each product must have an expected average X T. We may let the nominal = X T, or we may
determine it from a collection of data such as a collection of measurements or control
chart data.
fMR
1.200
1.185
1.180
1.169
1.164
1.160
1.156
1.154
1.152
1.150
1.148
1.143
1.140
H1317_CH42.qxd
452
10/15/07
3:50 PM
Page 452
Case Study
The Acme Chemical Co. produces several bioplasmic double regenerative phalangic processing aids (the famous Glycomul line). Each of these products has specific hydroxyl
value. These products are produced in a batch operation requiring approximately four
hours to produce a 2000-gallon batch of product. Specifications for the various products
are as follows:
Product
Glycomul S
Glycomul L
Glycomul T
Specification
235260
330350
5869
MR T = (1.143)(2.19),
MR T = 2.50.
data points give X T = 335.2 and S = 3.79. Calculate MR T from fMR = 1.154 and S = 3.79.
MR T = (1.154)(3.79)
MR T = 4.37
Glycomul T
Since this is a new product, we have no previous data. The development pilot plant trials
indicate that there will be some difficulty in manufacturing this product. We will assume
a Cpk of 1.00 and let the process average equal the specification nominal.
X T = Nominal = 58
(69 58 )
= 63.5
2
For MR T :
MRT =
MRT =
1.128 | 69 58 |
6(1.00 )
MRT = 2.07.
H1317_CH42.qxd
10/15/07
3:50 PM
Page 453
453
Product
Code
XT
MR T
Specification
A
B
C
249.8
335.2
63.5
2.50
4.37
2.07
235260
330350
5869
Glycomul S
Glycomul L
Glycomul T
Date
Product
Hydroxyl
value
Date
Product
Hydroxyl
value
11/3/96
11/3/96
11/3/96
11/3/96
11/4/96
11/5/96
11/5/96
11/5/96
11/6/96
A
A
A
B
B
C
A
A
C
250
248
253
334
336
63
249
253
64
11/6/96
11/7/96
11/7/96
11/7/96
11/8/96
11/8/96
11/8/96
11/9/96
11/9/96
C
C
A
B
A
A
C
C
C
65
62
252
335
247
250
65
60
66
Step 3. Calculate the location and variation plot for each data point. Plot points on the
chart.
Sample #1
Each observation must have a transformed value to plot for the individual. This sample is
the hydroxyl value for a batch of Glycomul S identified as product A. The target average
is 249.8, and the target moving range is 2.50. The calculated short-run plotting value for
this sample is
Individual plot point =
The variation plot point for this first sample is undefined because there is no moving range possible with only one data point. For this reason, the first moving range cell
is blackened in.
H1317_CH42.qxd
454
10/15/07
3:50 PM
Page 454
Sample #2
Another batch of Glycomul S is run with a reported hydroxyl value of 248.0. The target
value is again 249.8 with a target moving range of 2.50. The calculated short-run plot
point is
Individual plot point =
The moving range is the algebraic difference between the largest and smallest consecutive values.
Moving range = 0.08 (0.72) = 0.80
Note: All moving ranges are positive.
Plot the points and connect to the previous point.
Sample #3
A third batch of Glycomul S is run with a reported hydroxyl value of 253.0. The location
plot point is
( X XT ) (253.0 249.8) 3.20
=
=
= 1.28.
2.50
2.50
MRT
Mooving range = 1.28 (0.72) = 2.00
Sample #4
With the fourth sample, we have a product change. The next batch run is Glycomul L,
which has a different set of short-run parameters. For this product, XT = 335.2, and MR T =
4.37. The measured hydroxyl value for this batch is 334.0.
( X X T ) ( 334.0 335.2 ) 1.20
=
=
= 0.27
4.37
4.37
MR T
Moving range = 1.28 ( 0.27 ) = 1.55
Sample #5
Another batch of Glycomul L is run with a reported hydroxyl value of 336.0.
( X X T ) ( 336.0 335.2 ) 0.80
=
=
= 0.18
4.37
4.37
MR T
Moving range = 0.18 ( 0.27 ) = 0.45
H1317_CH42.qxd
10/15/07
3:50 PM
Page 455
455
Sample #6
Glycomul T, a new product that has been produced only in a pilot operation, is run for the
first time. We are using the nominal for the target average X T = 63.5. The expected average
moving range MR T is based on an assumed Cpk of 1.00. The estimate for MR T is 2.07.
( X X T ) ( 63.0 63.5 ) 0.50
=
=
= 0.24
2.07
2.07
MR T
Moving range = 0.18 ( 0.24 ) = 0.42
Continue calculating the plot points and plot the results. The completed control chart
showing all of the plotted historical data points follows.
History
+3.00
UCL = +2.66
+2.00
+1.00
0
1.00
2.00
3.00
LCL = 2.66
1
10
15
20
Individuals
25
30
UCL = 3.268
3.00
2.00
1.00
0.0
1
10
15
20
Moving range
25
30
H1317_CH42.qxd
456
10/15/07
3:50 PM
Page 456
Product
Hydroxyl value
Sample
Product
Hydroxyl value
19
20
21
22
A
A
B
B
248
250
331
329
23
24
24
25
B
C
C
A
333
59
59
243
The completed short-run individual/moving range control chart follows. It appears that
the process has changed with sample number 25.
Step 5. Continue to collect and plot data.
History
+3.00
UCL = +2.66
+2.00
+1.00
0
1.00
2.00
3.00
LCL = 2.66
1
10
15
20
Individuals
25
30
UCL = 3.268
3.00
2.00
1.00
0.0
1
10
15
20
Moving range
25
30
H1317_CH42.qxd
10/15/07
3:50 PM
Page 457
457
Bibliography
Besterfield, D. H. 1994. Quality Control. 4th edition. Englewood Cliffs, NJ: Prentice Hall.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Wheeler, D. J. 1991. Short Run SPC. Knoxville, TN: SPC Press.
H1317_CH42.qxd
10/15/07
3:50 PM
Page 458
H1317_CH43.qxd
10/15/07
6:13 PM
Page 459
A statistical process control (SPC) chart is a set of statistical control limits applied to a set
of sequential data from samples chosen from a process. The data comprising each of the
plotted points are a location statistic such as an individual, an average, a median, a proportion, and so on. If the control chart monitors variables data, then an additional associated chart for the process variation statistic can be utilized. Examples of variation statistics
are the range, standard deviation, or moving range.
By their design, control charts utilize unique and statistically rare patterns that we can
associate with process changes. These relatively rare or unnatural patterns are usually
assumed to be caused by disturbances or influences that interfere with the ordinary behavior of the process. These causes that disturb or alter the output of a process are called
assignable causes. They can be caused by equipment, personnel, or materials.
Most of these patterns can be characterized by:
1. The degree of statistical rarity for processes where no change has occurred (rate of
false alarm)
2. The direction in which the process has changed (increased or decreased)
Not all of the SPC detection rules indicate a direction of process change. For that reason, it is suggested that these rules or patterns be given a low priority for consideration and
application.
Several of these patterns, or SPC detection rules, will be discussed. These SPC rules
will be numbered as they are presented, and the example chart diagrams will be that of an
average. The detection rules, however, apply to all control charts (individual/moving
range, p chart, and so on). The example control chart diagrams will have the following
defined areas:
459
H1317_CH43.qxd
460
10/15/07
6:13 PM
Page 460
Rule 1. A single point outside the control limits; false alarm rate is 1/370.
This rule pattern will occur once in 370 times even if the process is in control and no shift
has occurred. The control limits are based on an average three standard deviations. With
a normal distribution, the expected proportion falling inside the control limits is 0.9973,
and the expected proportion falling outside the control limits is 0.0027. The 0.0027 proportion outside the control limits is equivalent to 1/370. The occurrence of rule #1 is frequently referred to as out of control when in fact it is a rule violation.
x
A
B
X
Rule 2. Seven consecutive points on the same side of the average; false alarm rate is 1/64.
The probability of getting seven points on one side of the average is (1/2)7 = 0.0078. The
probability of getting seven consecutive points above the average or below the average is
2(1/2)7 = 0.0156.
Seven consecutive points above or below the average is equivalent to a rate of 1/64.
The number of consecutive points in a row varies from reference to reference. All are correct; only the false alarm rate will change. See the following sources for various selections
for the number of points on one side of the average.
Number of
consecutive points
1/64
Introduction to Statistical
Quality Control, 2nd edition,
Douglas C. Montgomery
1/128
ISO-8258
1/256
Reference
B
X
C
C
B
A
x
x
x
x
x
x
H1317_CH43.qxd
10/15/07
6:13 PM
Page 461
461
Rule 3. Seven consecutive points increasing (or decreasing), indicating a trend; false alarm
rate is 1/360.
The probability of getting seven consecutive points increasing (or decreasing) is 2/6! =
0.00277. This is equivalent to a rate of 1/360.
A
B
X
x
x
x
x
C
B
x
x
Rule 4. Two out of three consecutive points falling in zone A or beyond; false alarm rate
is 1/659.
The probability of getting a single point in zone A or beyond is 0.02275. The probability
of getting two consecutive points in zone A or beyond is (0.02275)2 = 0.000518, which
gives a rate of 1/1923. To reduce this false alarm rate, the rule stipulates that the number
of points in zone A or beyond in three consecutive samples be three. The probability of
getting two out of three consecutive points in zone A or beyond is (0.02275)2(0.97725)(3)
= 0.00152, which is equivalent to a rate of 1/658.
A
B
X
B
A
Rule 5. Four out of five consecutive points falling in zone B or beyond; false alarm rate
is 1/360.
The probability of getting a single point in zone B or beyond is 0.1567. The probability of
getting four consecutive points is (0.1567)4 = 0.000603, equivalent to a rate of 1/1659. The
rule is four out of five in zone B or beyond, not four consecutive points. The probability of
getting a point somewhere other than zone B or beyond is 1 0.1567 = 0.8433. The probability of getting four out of five in zone B or beyond is (0.000634)(0.8433)(5) = 0.0027,
which is equivalent to 1/360.
A
B
X
x
x
C
C
B
A
H1317_CH43.qxd
462
10/15/07
6:13 PM
Page 462
Rule 6. Five consecutive points outside zone C with points on both sides of the average;
false alarm rate is 1/331.
The probability of getting five points outside zone C (both sides of the average) is given by:
(15/16)(0.31738)5 = 0.00302 equivalent = rate of 1/331.
A
B
X
C
B
Rule 7. Fifteen points in zone C above and below the average; false alarm rate is 1/307.
The probability of getting 15 points above and below the average is given by:
(0.68262)15 = 0.00326, which is equivalent to 1/307.
A
B
X
x x
x x
x x x
x x
x
x
x
x x
B
A
Rule 8. Fourteen points in a row alternating up and down; false alarm rate is 1/219.
A
B
X
x
x
x
C
B
A
x
x
x
x
x
x
x
x
H1317_CH43.qxd
10/15/07
6:13 PM
Page 463
463
Bibliography
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Nelson, L. S. 1984. Technical Aids: The Shewhart Control ChartTests for Special
Causes. Journal of Quality Technology 16, no. 4 (October): 237239.
Technical Committee TC 69/5C. 1991. Shewhart Control Charts. Geneva, Switzerland:
International Organization for Standardization.
Wheeler, D. J. 1995. Advanced Topics in Statistical Process Control. Knoxville, TN:
SPC Press.
H1317_CH43.qxd
10/15/07
6:13 PM
Page 464
H1317_CH44.qxd
10/17/07
2:47 PM
Page 465
Taguchi defines quality as the avoidance of loss to society in terms of economic loss or
dollars. This loss is a result of noncompliance to manufacturing or service specification
targets. He also suggests that this loss can be mathematically modeled as a function. The
Taguchi loss function is quadratic in nature in that Taguchi hypothesizes that the loss is
proportional to the square of the deviation from the target value.
The Taguchi loss function is
Cx = K(X Tg)2,
where: Cx = cost, dollars
K = proportionally constant (loss parameter or loss constant)
X = observed or measured value
Tg = Target or nominal value (perfection), sometimes abbreviated .
Loss, dollars
Target
Tg T l
Tg
Tg + T l
The key to utilizing the Taguchi loss function is in the determination of the proportionally constant K.
The specification for a shaft is 1.00 0.05. The cost of nonconformance (Cx) is
$25.00. What is the value of K?
Cx = K ( X Tg)2
K=
Cx
( X Tg)2
465
K=
25.00
= 10, 000
(1.05 1.00)2
H1317_CH44.qxd
466
10/17/07
2:47 PM
Page 466
Knowing K, we may now calculate the loss for any shaft diameter. For example, given
K = 10,000, what is the expected loss when the shaft diameter is 1.08?
Cx = K(X Tg)2
Cx = 10,000(1.08 1.00)2
Cx = 10,000(0.08)2
Cx = $64.00
Even if the shaft is within the specification limits, there will be an associated loss. The
Taguchi philosophy is such that any deviation from perfection (target) will result in a loss
(to society). The loss for a shaft measuring 0.98 is
Cx = $4.00.
Cx = 10,000(0.98 1.00)2,
A customers specification for a particular part is 10.00 0.15 (Tg Tlc), and the associated cost of nonconformance is $35.00, Cc. In an effort to avoid producing nonconforming parts, the manufacturer sets an internal specification of 10.00 0.05 (Tg Tlm).
The internal cost to the manufacturer for failing to meet this requirement is $12.00.
The customer will realize a cost of Cc whenever the product lies at or exceeds the
specification of Tg Tlc. The customers loss will be
Cc = K(Tg + Tlc Tg)2,
Cc = K(Tlc)2.
Likewise, it can be shown that the loss to the manufacturer when exceeding the manufacturing specification is
Cm = K(Tlm)2.
The value for the loss parameter K can be found by solving for K in either relationship.
Loss parameter K =
Cc
Cm
2 =
(Tlc)
(Tlm)2
In this example, where Cm = $12.00 and Tlm = 0.05, the value for the loss parameter K is given by
K=
12.00
= $4800.
(0.05)2
Cc
Cm
2 =
(Tlc)
(Tlm)2
Tlm = Tlc
Cm
Cc
Tlm = 0.15
122
35
Tlm = 0.088.
H1317_CH44.qxd
10/17/07
2:47 PM
Page 467
467
The manufacturers tolerance can be determined when the ratio of the consumers loss
to the manufacturers loss is known.
Example:
Customers tolerance, Tlc = 0.015
Customers loss, Cc = $25.00
Manufacturers loss, Cm = $15.00
Manufacturer's tolerance, Tlm = Tlc
Cm
Cc
15
25
KL =
120.00
( X Tg)2
120.00
(0.2)2
K L = $3000
22.00
(10.0 9.8)2
KU = $550
KL
KU determination:
Loss at upper specification, UCx = $22.00
22.00 = KU (Tg X )2
KU =
22.00
(Tg X )2
KU =
The loss function for the parts measuring less than the target is given by
LCx = 3000(Tg X)2.
H1317_CH44.qxd
468
10/17/07
2:47 PM
Page 468
The loss function for the parts measuring greater than the target is given by
UCx = 550 (X Tg)2.
Using these loss functions, the following losses are determined for various measurement observations:
Measurement
Loss, dollars
Measurement
9.5
9.6
9.7
9.8
9.9
10.0
10.1
10.2
750
480
270
120
30
0
5.5
22
10.3
10.4
10.5
10.6
10.7
10.8
10.9
11.0
LS
Tg
US
Loss, dollars
49.5
88
137.5
198
269.5
352
446
550
800
X
700
600
X
Loss, dollars
500
X
X
400
X
300
X
X
200
$120
100
X
X
$22
0
X
9.4
9.6
9.8
X
X
10.0
X X
10.2
10.4
Measurement
Taguchi loss with differing proportionality constants.
10.6
10.8
11.0
H1317_CH44.qxd
10/17/07
2:47 PM
Page 469
469
8.35
8.4
8.35
8.3
8.4
8.4
8.65
8.6
8.4
8.35
8.4
8.6
8.4
8.5
8.5
8.5
8.2
8.4
8.5
8.5
1
Upper mean square, UMS = ( XU tgt )2
n
where: XU = data below the target or nominal
n = total samples
tgt = target or nominal
UMS =
1
[3(8.6 8.5)2 + 2(8.65 8.5)2 ] = 0.0021
25
1
( X L tgt )2
n
LMS =
H1317_CH44.qxd
470
10/17/07
2:47 PM
Page 470
Bibliography
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Wheeler, D. J. 1995. Advanced Topics in Statistical Process Control. Knoxville, TN: SPC
Press.
H1317_CH45.qxd
10/17/07
2:48 PM
Page 471
Sample
value X
1
2
3
4
5
6
7
14.0
15.8
16.0
17.3
17.7
18.3
19.0
471
H1317_CH45.qxd
472
10/17/07
2:48 PM
Page 472
Sample
number I
Sample
value X
8
9
10
11
12
13
14
15
19.0
20.2
20.4
21.0
21.9
22.3
22.8
24.6
(i 0.5)
100.
n
( i 0.3 )
100
( n + 0.4 )
ic
for the Hazen c = 0.5 and for the Benard c = 0.3 .
n 2c +1
n i + ( 0.5 ) 1/ n ( 2 i n 1)
.
n 1
H1317_CH45.qxd
10/17/07
2:48 PM
Page 473
473
% MR =
( i 0.3 )
( n + 0.4 )
100
% MR =
100
% MR =
(1 0.3 )
(15 + 0.4 )
100
% MR =
100
% MR =
( 0.7 )
(15.4 )
100
% MR = 4.55%
2nd median rank:
% MR =
( i 0.3 )
( n + 0.4 )
( 2 0.3 )
(15 + 0.4 )
(1.7 )
(15.4 )
100
% MR = 11.0%
In a similar manner, continue to calculate the percent median rank for all 15
observations.
Sample
number i
Sample
value X
% median
rank
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
14.0
15.8
16.0
17.3
17.7
18.3
19.0
19.0
20.2
20.4
21.0
21.9
22.3
22.8
24.6
4.55
11.00
17.53
24.03
30.52
37.01
43.51
50.00
56.49
62.99
69.48
75.97
82.47
88.96
95.45
H1317_CH45.qxd
474
10/17/07
2:48 PM
Page 474
70
60
50
40
30
20
10
5
4.55
2
1
0.5
0.2
0.1
0.05
0.01
10 12 14 16 18 20 22 24 26 28 30 32 34 36 38 40
14.0
Value
Median rank
80
H1317_CH45.qxd
10/17/07
2:48 PM
Page 475
The second data point is plotted; X2 = 15.8 and median rank position = 11.00%.
99.99
99.9
99.8
99
98
95
90
80
50
40
30
20
x
xx
11
10
5
2
1
0.5
0.2
0.1
0.05
0.01
10 12 14 16 18 20 22 24 26 28 30 32 34 36 38 40
15.8
Value
Median rank
70
60
475
H1317_CH45.qxd
476
10/17/07
2:48 PM
Page 476
Continue to plot the remaining points in a similar manner. When all of the points have
been plotted, draw the best straight line possible through the points. Be less influenced by
the first and last points of the data, as these are the regions where the tails of the distribution influence the distribution of the normal probability the most. The degree to which a
single straight line can be drawn though all of the points is proportional to the degree that
a normal distribution is present.
The following figure shows the completed normal probability plot. A normal distribution is supported due to the straightness of the fitted line to the plot.
99.99
99.9
99.8
99
98
x
x
x
95
90
80
50
40
30
x
x
x
x
x
20
x
x
10
5
x
2
1
0.5
0.2
0.1
0.05
0.01
10 12 14 16 18 20 22 24 26 28 30 32 34 36 38 40
Value
Completed normal probability plot.
Median rank
70
60
x
x
x
H1317_CH45.qxd
10/17/07
2:48 PM
Page 477
477
The degree to which a straight line can be drawn through all of the points reflects the
degree to which a normal distribution is appropriate to describe the data. This example
exhibits normalcy.
If normal probability paper is not available, the median ranks (not %MR but MR) may
be transformed and plotted using ordinary rectangular graph paper. The transformation
equation is given by
Yi = 4.91[(MR)0.14 (1 MR)0.14].
Normal Probability Plots Using Transformed Data
When normal probability plotting paper is available, one may still test for a normal distribution using conventional rectangular coordinate paper by transposing the median rank
values or Pi values to Yi values according to the following:
Yi = 4.91[(Pi)0.14 (1 Pi)0.14].
No transformation of the individual data values, Xi, is required.
Consider the following example: Fourteen data values, Xi, are given. The Benards median
rank values are determined and transformed.
Order, i
Data
value, Xi
Median
rank, Pi
Transformed
median rank, Yi
1
2
3
4
5
6
7
8
9
10
11
12
13
14
25.08
28.99
29.60
30.54
30.75
32.85
33.51
34.10
34.62
35.90
36.07
36.10
39.36
42.03
0.0486
0.1181
0.1875
0.2569
0.3264
0.3958
0.4653
0.5347
0.6042
0.6736
0.7431
0.8125
0.8819
0.9514
1.661
1.184
0.885
0.651
0.448
0.263
0.087
+0.087
+0.263
+0.448
+0.651
+0.885
+1.184
+1.661
i 0.3
1 0.3
=
= 0.0486
n + 0.4 14 + 0.4
H1317_CH45.qxd
478
10/17/07
2:48 PM
Page 478
Transformation of Pi:
Yi = 4.91[Pi)0.14 (1 Pi)0.14]
Yi = 4.91[(0.0486)0.14 (1 0.0486)0.14] = 1.661
The Yi and Xi values are plotted on rectangular coordinate graph paper. The resulting Yi
values represent the vertical plot positions, and the individual data values, Xi, represent the
horizontal positions. The degree to which a straight line is obtained is proportional to the
degree of normalcy for the data. The completed plot is indicative of normally distributed
data because the line is relatively straight. The Xi intercept is an estimate of the average of
the data, and the slope is inversely proportional to the standard deviation.
+3
+2
+1
10
20
30
40
50
Xi
Yi
Normal probability plot using rectangular coordinates.
H1317_CH45.qxd
10/17/07
2:48 PM
Page 479
479
(Oi Ei )2
Ei
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
10
11
12
53
46
50
49
48
50
52
54
51
48
54
54
51
47
50
50
44
50
50
47
52
53
51
48
43
56
46
49
46
51
54
50
49
50
51
50
46
56
52
56
52
44
47
49
53
51
48
49
47
53
48
50
52
47
50
50
57
51
51
48
50
50
51
53
50
48
45
51
52
47
45
52
51
57
49
47
44
45
48
52
53
51
52
47
48
50
51
47
52
46
46
49
54
48
53
48
51
53
42
50
56
49
51
51
48
49
54
53
51
47
51
53
49
51
50
48
52
50
54
49
50
50
47
47
49
47
51
56
54
49
51
51
48
48
55
47
50
50
49
48
46
46
47
52
49
53
52
51
48
47
47
45
57
48
51
49
49
54
48
48
52
52
48
47
48
50
49
47
56
53
50
50
53
51
46
51
50
47
45
50
50
55
42
55
45
47
57
48
51
50
43
52
55
47
48
58
50
54
54
53
53
49
48
55
45
49
49
54
58
52
48
47
47
53
51
54
49
53
47
48
55
49
45
52
53
49
51
48
49
53
48
50
49
50
50
49
48
53
49
51
52
48
57
51
54
49
53
45
52
51
50
50
51
46
50
48
53
45
50
50
48
55
50
51
50
51
49
51
48
53
47
51
50
44
53
55
54
44
49
50
50
53
51
49
46
53
49
50
42
54
48
52
53
56
49
53
47
56
50
51
H1317_CH45.qxd
480
10/17/07
2:48 PM
Page 480
Step 2. Calculate the average and sample standard deviation of the data.
Interval value
X 2S
X 2S to X 1S
X 1S to X
X to X + 1S
X + 1S to X + 2S
X + 2S
<44
44 to
47 to
50 to
53 to
>56
Frequency
5
27
95
106
52
15
47
50
53
56
Step 4. Determine the expected frequency of occurrence in the defined intervals given the
average and standard deviation for a normal distribution.
If we are dealing with a normal distribution, we can calculate the expected number of data
values falling inside the defined intervals. The expected percentage of the data falling
inside the intervals and the actual observed data values inside the intervals are calculated
as follows:
Interval definition
% of data
for normal distribution
X 2S
X 2S to X 1S
X 1S to X
X to X + 1S
X + 1S to X + 2S
X + 2S
2.5
13.5
34.0
34.0
13.5
2.5
Expected
data points
7.5
40.5
102
102
40.5
7.5
Observed
data points
5
27
95
106
52
15
2 =
(Oi Ei )2
.
Ei
H1317_CH45.qxd
10/17/07
2:48 PM
Page 481
481
Interval definition
X 2S
X 2S to X 1S
X 1S to X
X to X + 1S
X + 1S to X + 2S
X + 2S
Expected E
data points
Observed O
data points
(O E )2
E
7.5
40.5
102
102
40.5
7.5
5
27
95
106
52
15
0.8
4.5
0.5
0.2
2.7
7.5
2 =
(Oi Ei )2
= 16.2
Ei
H1317_CH45.qxd
482
10/17/07
2:48 PM
Page 482
Ku and Sk are based on calculations using moments of the mean. The kth moment of
the mean is given by
( Xi X )k
,
n
k = the kth moment of the mean,,
k =
where: X = average
Xi = an individual observation.
The first moment about the mean is always zero, and the second moment about the
mean adjusted for bias is the variance.
1 = 0
n
= s2
2
n 1
2 = 2
= 2
3
( 2 )
3
23/ 2
4
3.
22
6
n
and SKU =
24
.
n
The confidence intervals for the skewness and kurtosis are given by
Z / 2 6
Sk
n
Z / 2 24 .
and Ku
n
Z/2 = 1.96 for 95 percent confidence and 1.645 for 90 percent confidence.
H1317_CH45.qxd
10/17/07
2:48 PM
Page 483
483
The following data illustrate calculation of Sk and Ku and testing for normalcy of
the data:
(X X )2
Frequency f
6
7
8
9
10
11
12
13
14
15
1
6
15
13
11
8
6
4
2
1
13.69
7.29
2.89
0.49
0.09
1.69
5.29
10.89
18.49
28.09
(X X )3
(X X )4
50.65
19.68
4.91
0.34
0.03
2.20
12.17
35.94
79.51
148.88
187.40
53.14
8.35
0.24
0.01
2.86
27.98
118.59
341.88
789.05
2 =
( X X )2
n
2 =
3 =
( X X )3
n
3 =
4 =
( X X )4
n
4 =
15
10
0
1
10 11 12 13 14 15 16 17 18
H1317_CH45.qxd
484
10/17/07
2:48 PM
Page 484
Sk =
3
4.42
= 0.57
3/ 2 =
2
(3.91)3
67
Ku = 24 3 =
2 3 = 0.29
2
261
.72
67
90 percent confidence interval:
24
0.29 1.645
1.27 to 0.69
67
This distribution is skewed to the right as indicated by the shape of the frequency histogram and the positive Sk. The distribution is also flatter than a perfectly normal distribution as indicated by the slightly negative Ku. The distribution is normal with respect to
the amount of kurtosis since the confidence interval for the kurtosis contains zero (which
is expected for a kurtosis from a normal distribution). However, the confidence interval for
the skewness does not contain zero, and we conclude that the data come from some distribution other than a normal one.
Bibliography
Betteley, G., N. Mettrick, E. Sweeney, and D. Wilson. 1994. Using Statistics in Industry.
New York: Prentice Hall.
Dovich, R. A. 1992. Quality Engineering Statistics. Milwaukee, WI: ASQC Quality Press.
Grant, E. L., and R. S. Leavenworth. 1996. Statistical Quality Control. 7th edition. New
York: McGraw-Hill.
Hayes, G. E., and H. G. Romig. 1982. Modern Quality Control. 3rd edition. Encino, CA:
Glencoe.
Montgomery, D. C. 1996. Introduction to Statistical Quality Control. 3rd edition. New
York: John Wiley & Sons.
Petruccelli, J. D., B. Nandram, and M. Chen. 1999. Applied Statistics for Engineers and
Scientists. Upper Saddle River, NJ: Prentice Hall.
Walpole, R. E., and R. H. Myers. 1993. Probability and Statistics for Engineers and Scientists. 5th edition. Englewood Cliffs, NJ: Prentice Hall.
H1317_CH46.qxd
10/15/07
5:00 PM
Page 485
Weibull Analysis
The Weibull distribution is one of the most used density functions in reliability. It is named
after Waloddi Weibull, who invented it in 1937. Weibulls 1951 paper, A Statistical Distribution Function of Wide Applicability, led to a wide use of the distribution. Weibull
demonstrated that this distribution is very flexible for a wide variety of data distributions.
Though initial reaction to Weibulls distribution varied from skepticism to outright
rejection, his distribution has passed the test of time and today is widely accepted as a
viable model for application in reliability engineering.
The Weibull probability density function is defined as follows:
F (t ) =
( t ) 1
e
Distribution
Exponential
Rayleigh
Lognormal
Normal
The Weibull reliability function describes the probability of survival as a function of time.
Rt = e
When < 1, the rate of change in the reliability is decreasing; when = 1, the failure rate
is constant; and when > 1, the failure rate is increasing. These three areas are frequently
referred to as infant mortality, useful life, and burnout.
485
H1317_CH46.qxd
486
10/15/07
5:00 PM
Page 486
80 5 1.2
230
(t ) 1
Example:
What is the instantaneous failure rate at t = 80, where = 1.2, = 230, and = 5?
h( t ) =
1.2(80 5)0.2
= 0.0042
2301.2
While the Weibull is very versatile, its drawback is the difficulty in estimating its
parameters. There are several methods for estimating the Weibull parameters.
Probability Plotting
Taking the logarithm of the Weibull cumulative distribution twice, the following relationship is obtained:
1
ln ln
= ln (t ) ln .
1 f (t )
This is a linear equation with ln(t) as the independent variable, as the slope, the depen 1
dent variable as ln ln
, and ln as the y-intercept. Plotting ln(t) against
1 f (t )
1
ln ln
, we may graphically determine the Weibull parameters and .
1 f (t )
For an estimate of f(t), we may use Benards median rank.
f (t ) =
i 0.3
n + 0.4
H1317_CH46.qxd
10/15/07
5:00 PM
Page 487
Weibull Analysis
487
i 0.3
i 0.3
=
= 0.067.
n + 0.4 10 + 0.4
i 0.3
7 0.3
=
= 0.644.
n + 0.4 10 + 0.4
Step 3. Determine the natural logarithm for the time to failure for each observation.
Step 4. Determine the natural logarithm of the inverse of 1 MR for each observation.
For the first ordered observation:
1
1
= 2.669
ln ln
.
= ln ln
1 0.067
1 f (t )
For the seventh ordered observation:
1
1
= 0.0323 .
ln ln
= ln ln
1 0.644
1 f (t )
H1317_CH46.qxd
10/15/07
488
5:00 PM
Page 488
Time to failure
f(t)
ln(t)
ln[ln(1/1 f (t ))]
1
2
3
4
5
6
7
8
9
10
87
118
137
175
190
203
221
264
270
333
0.067
0.163
0.259
0.355
0.452
0.548
0.644
0.740
0.837
0.933
4.47
4.77
4.92
5.16
5.25
5.31
5.40
5.58
5.60
5.81
2.669
1.726
1.205
0.824
0.508
0.231
0.032
0.298
0.596
0.994
1
Step 5. Plot the ln (t) on the x-axis and ln ln
on the y-axis.
1 f (t )
1n (t )
1
x x
xx
2
x
4
6
1
1n 1n
1 f (t )
Slope = 2.70
10
12
14
Y0 = 14.66
The Weibull slope, = 2.70. The Weibull characteristic life is calculated from the
equation
=e
y0
H1317_CH46.qxd
10/15/07
5:00 PM
Page 489
Weibull Analysis
489
Order
number, i
Failure time,
hours
%MR
1
2
3
4
5
6
7
8
9
10
11
12
13
28
39
50
60
70
85
100
115
135
160
200
250
370
7.5
12.7
20.1
27.6
35.1
42.5
50.0
57.5
64.9
72.4
79.9
87.3
94.8
A plot of the failure times as a function of the %MR yields a slightly curved line. Discretion must be used in the interpretation of the underlying cause for the nonlinearity. This
could be due to the existence of mixed distributions, or there could be a failure-free time
or a minimum life characteristic.
The minimum life characteristic is time that, when subtracted from all of the failure
times, yields a somewhat straight line. Essentially by subtracting the minimum life characteristic from all of the failure times and by replotting the data, we are reducing the
Weibull plot to a two-parameter model using the characteristic life and Weibull slope.
H1317_CH46.qxd
490
10/15/07
5:00 PM
Page 490
We begin by plotting the original data as a function of the %MR. The data are ordered
in decreasing order. The MR is calculated using Benards approximation. The ith MR is
given by
%MR =
i 0.3
100.
n + 0.4
6.0
4.0
3.0
2.
0
99
4
1. 2
1.
1.0
0.8
0.7
0.6
0.5
95
90
80
x
x
x
x
70
60
x
x
x
x
40
x
30
x
x
20
x
10
5
4
3
600
700
800
900
1000
500
400
300
200
50
60
70
80
90
100
40
30
20
10
50
H1317_CH46.qxd
10/15/07
5:00 PM
Page 491
Weibull Analysis
491
Order number, i
%MR
1
2
3
4
5
6
7
8
9
10
11
12
13
28
39
50
60
70
85
100
115
135
160
200
250
370
7.5
12.7
20.1
27.6
35.1
42.5
50.0
57.5
64.9
72.4
79.9
87.3
94.8
7
18
29
39
49
64
79
94
114
139
179
229
349
492
10/15/07
5:00 PM
Page 492
6.0
4.0
3.0
H1317_CH46.qxd
2.
0
99
4
1. 2
1.
95
= 1.12
1.0
0.8
0.7
0.6
0.5
90
80
70
60
40
L1
50
30
20
10
5
4
3
500
400
300
600
700
800
900
1000
= 112
200
50
60
70
80
90
100
40
30
20
10
Bibliography
Abernethy, R. B. 1994. The New Weibull Handbook. North Palm Beach, FL: Robert B.
Abernethy.
Dodson, B. 2006. The Weibull Analysis Handbook. 2nd edition. Milwaukee, WI: ASQ
Quality Press.
Dovich, R. A. 1992. Reliability Engineering Statistics. Milwaukee, WI: ASQC Quality Press.
OConnor, P. D. T. 1991. Practical Reliability Engineering. New York: John Wiley & Sons.
Weibull, W. 1951. A Statistical Distribution Function of Wide Applicability. Journal of
Applied Mechanics 18:293297.
H1317_CH47.qxd
10/15/07
5:02 PM
Page 493
Type B
1689
901
1068
394
836
891
1592
258
153
528
712
Ranking all the data with the source identified, we get the following table:
Observation
Type
Rank
153
258
394
528
712
836
891
901
1068
1592
1689
B
B
A
B
B
A
B
A
A
B
A
1
2
3
4
5
6
7
8
9
10
11
493
H1317_CH47.qxd
494
10/15/07
5:02 PM
Page 494
Step 1. Construct the null hypothesis (Ho) and the alternative hypothesis (Ha).
For this example, we want to prove that the median life of type A is greater than that of
type B.
Ho : A = B
Ha : A > B
Step 2. Select a level of confidence and risk. Let the risk of our conclusion be 8.9 percent.
(The probability of a type I error is 0.089.)
Step 3. Calculate the test statistic. The test statistic will be the sum of ranks W for the first
sample, NA.
W = 3 + 6 + 8 + 9 + 11 = 37
Step 4. Compare the calculated sum of ranks W with the critical sum of ranks WC. See the
following table, where n2 = larger sample size and n1 = smaller sample size.
The relationships of the calculated W and the critical sum of ranks with respect to rejecting the null hypothesis and accepting the alternative hypothesis are:
If Ha : A B , use the table of lower and upper percentage points
If Ha : A > B , use the table of upper percentage points
If Ha : A < B , use the table of lower percentage points
If W > WC, reject Ho and accept Ha; 37 is not greater than 38, do not reject Ho. At a
risk of 8.9 percent or level of confidence of 90.1 percent, there is not sufficient evidence
to conclude that type A has a median greater than that of type B.
n1
n2
2/0.500
3/0.333 7/0.167
4/0.250 9/0.100
14/0.100
15/0.050
5/0.200 11/0.067
23/0.100
17/0.057
25/0.029
18/0.029
26/0.014
6/0.167
7/0.143
8/0.125
H1317_CH47.qxd
10/15/07
5:02 PM
Page 495
495
Tabled values are limited. WC for larger N values greater than 10 can be approximated by
Lower WC =
n1 (n1 + n2 + 1)
n n (n + n + 1)
Za 1 2 1 2
2
12
Upper WC =
n1 (n1 + n2 + 1)
n n (n + n + 1)
.
+ Za 1 2 1 2
2
12
and/or
If we estimate the upper WC (recall that the table value was 38), we find WC = 37.4.
n1 (n1 + n2 + 1)
n n (n + n + 1)
+ Za 1 2 1 2
2
12
n2 = 6
n1 = 5
a = 0.089
Za = 1.35
5(5 + 6 + 1)
(5)(6)(5 + 6 + 1)
+ 1.35
= 37.4
2
12
Illustrative example 2:
A study has been made comparing the systolic blood of individuals selected from two body
mass index (BMI) groups. The BMI is determined by
BMI =
Weight (lbs.)
703.
Height (in.)2
Underweight
Normal
Overweight
Obese
Group A has BMIs between 25 and 29, and group B has BMIs between 24 and 25.
Data:
Group A: 137, 138, 144, 140, 139, 138, 138, 139, 138, 143, 144, 140, 142, 140, 140
Group B: 139, 138, 134, 136, 137, 139, 138, 133, 133, 137, 135
NA = N1 = 15
NB = N2 = 11
Step 1. Construct the null hypothesis (Ho) and the alternative hypothesis (Ha).
H1317_CH47.qxd
496
10/15/07
5:02 PM
Page 496
Objective: Are the two groups from the same population or from different populations? No
distribution is assumed.
Ho : A = B
Ho : A B
Step 2. Select a level of confidence and risk. The level of confidence is 95 percent, and the
risk is 5 percent. Because this is a two-tailed test, we will place half our risk in each test
statistic, the upper and lower critical Wcs.
Step 3. Calculate a test statistic. The sum of ranks for the first sample will be the test statistic.
Rearranging the data in ranked order, we get the following table:
Systolic
blood
pressure
Group
Order
Rank
133
133
134
135
136
137
137
137
138
138
138
138
138
138
139
139
139
139
140
140
140
140
142
143
144
144
B
B
B
B
B
A
B
B
B
B
A
A
A
A
A
A
B
B
A
A
A
A
A
A
A
A
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
1.5
1.5
3
4
5
7
7
7
11.5
11.5
11.5
11.5
11.5
11.5
16.5
16.5
16.5
16.5
20.5
20.5
20.5
20.5
23
24
25.5
25.5
If multiple values occur, the rank is reported as the median for all the multiple values.
For example, in this study the value 133 occurred with the first- and second-order values.
The corresponding ranks (for each value) are the median of 1 and 21.5.
H1317_CH47.qxd
10/15/07
5:02 PM
Page 497
497
n1 (n1 + n2 + 1)
n n (n + n + 1)
Za 1 2 1 2
2
12
n1 (n1 + n2 + 1)
n n (n + n + 1)
+ Za 1 2 1 2
2
12
Rejection criteria: If the calculated W is less than the lower WC or if the calculated W
is greater than the upper WC, reject Ha and accept Ho.
Lower critical rank, WC =
15(15 + 11 + 1)
(15)(11)(15 + 11 + 1)
1.96
= 149.7
12
2
W = 289
n1
n2
2/0.500
3/0.333 7/0.167
4/0.250 9/0.100
14/0.100
15/0.050
5/0.200 11/0.067
23/0.100
17/0.057
25/0.029
18/0.029
26/0.014
6/0.167
7/0.143
8/0.125
10
10
H1317_CH47.qxd
498
10/15/07
5:02 PM
Page 498
W is not less than the lower WC of 149.7. Therefore, we cannot (at this time) reject Ho.
We now compare W to the upper critical rank value.
Upper critical rank, WC =
15(15 + 11 + 1)
(15)(11)(15 + 11 + 1)
+ 1.96
= 225.3
12
2
W = 289
W is greater than the upper WC of 225.3. Therefore, we can reject Ho in favor of Ha.
We have sufficient evidence to conclude that the two groups do in fact represent two populations of systolic blood pressure.
n1
n2
10
1/0.500
1/0.333 3/0.167
1/0.250 3/0.100
6/0.050
7/0.100
1/0.200 3/0.067
10/0.014
6/0.029
11/0.029
7/0.057
13/0.100
1/0.167
3/0.048
4/0.095
1/0.143
3/0.036
4/0.071
10
H1317_CH48.qxd
10/15/07
5:05 PM
Page 499
The successful implementation of statistical process control (SPC) into the work environment requires that the people most closely related to the process maintain the control chart.
In many cases and for a variety of reasons, these individuals may feel uncomfortable with
the task of plotting data and having to remember a number of the traditional rules that
indicate an out-of-control condition or evidence that a change in the process characteristic
being monitored has occurred.
Zone charting offers an effective way to overcome many of these types of obstacles.
Zone charting uses a summation rule based on the probability of getting values within certain discrete areas as defined by the process average and increments of 1, 2, and 3 standard deviations.
Traditional Shewhart control charts define a location statistic and a variation statistic.
The variation statistic is based on 3 standard deviations and is derived from some type of
range measurement.
The individual/moving range control chart limits for the individual portion are determined by
Upper control limit (UCL) = X + 2.66 MR
and
Lower control limit (LCL) = X 2.66 MR,
where: MR = moving range
X = average of individuals.
The average/range control chart limits for the averages portion is determined by
UCL = X + A2 R
and
LCL = X A2 R ,
where: R = range
X = average of averages
A2 = a factor dependent on the subgroup sample size.
499
H1317_CH48.qxd
500
10/15/07
5:05 PM
Page 500
In both cases, the control chart limits are based on an average response 3 standard
deviations.
The performance of a control chart is determined by the average amount of time
required to detect a change in the process. This performance is measured in terms of the
average run length (ARL). The ARL for a Shewhart control chart using a single detection
rule of a single point outside the upper or lower control limit where a 1 S shift has
occurred is ARL = 42.
By adding additional rules of detection, we can reduce the ARL for a given amount of
shift in the process. For example, by adding the rule of seven points in a row on the same
side of the average, we reduce the ARL of 42 to an ARL of 11 (for a 1 S shift). Other
rules are available, but the task of remembering all of the rules becomes burdensome in
application.
Zone charts are based on the probability of getting values in each of the eight defined
zones or areas. The weighting factors are proportional to the inverse of the probability of
the event occurring. The zone factors are also adjusted to allow for various ARLs for a
given amount of process change.
H1317_CH48.qxd
10/15/07
5:05 PM
Page 501
501
ARLs for selected zone scales deviation from historical average in units of S.
Zone scores
Process change
CRS
0.0
0.5
1.0
1.5
2.0
0
0
0
0
0
0
1
0
1
1
1
1
1
1
1
1
2
2
2
3
3
2
2
2
5
2
4
4
6
6
4
4
5
6
12
4
8
8
8
12
4
4
5
6
12
5
8
8
8
12
49
100
211
320
358
311
41
100
27
62
19
26
45
63
86
44
16
26
13
21
7
8
11
14
17
11
6
8
6
7
4
4
5
6
7
5
4
4
3
4
2.5
2.9
3.3
3.8
3.8
3.6
2.8
2.8
2.4
2.9
370
100
140
33
42
12
13
7
6.5
3
H1317_CH48.qxd
10/15/07
502
5:05 PM
Page 502
The following data will be used to develop an average/range control chart using the
zone format.
Collect Historical Data
#1
#2
#3
#4
#5
#6
7.17
7.14
7.09
7.29
7.13
7.13
7.23
7.23
7.37
7.08
7.15
7.16
7.08
7.18
7.15
7.11
7.25
7.25
7.33
7.15
7.19
7.20
7.28
7.27
7.31
X = 7.19
R = 0.32
X = 7.16
R = 0.20
X = 7.21
R = 0.29
X = 7.14
R = 0.10
X = 7.22
R = 0.22
X = 7.25
R = 0.12
#7
#8
#9
#10
#11
#12
7.01
7.14
7.27
7.23
7.22
7.42
7.22
6.98
7.12
7.24
7.23
7.05
7.34
7.11
7.36
7.14
7.08
7.26
7.25
7.18
7.26
7.36
7.12
7.24
7.16
X = 7.10
R = 0.51
X = 7.17
R = 0.26
X = 7.20
R = 0.44
X = 7.22
R = 0.31
X = 7.18
R = 0.18
X = 7.23
R = 0.24
#13
#14
#15
#16
#17
#18
7.15
7.21
7.30
7.18
7.25
7.37
7.23
7.16
7.13
7.06
7.27
7.06
7.34
7.05
7.19
7.20
7.18
7.00
7.39
7.05
7.28
7.20
7.25
7.41
7.17
X = 7.22
R = 0.15
X = 7.19
R = 0.31
X = 7.18
R = 0.29
X = 7.16
R = 0.39
X = 7.26
R = 0.24
7.14
7.03
7.33
7.35
7.11
7.37
7.05
7.09
7.11
6.86
7.17
7.13
7.15
7.15
7.12
X = 7.14
R = 0.05
Samples 118 will be used to determine the historical characterization of the process.
The location statistic will be the grand average of the averages X .
H1317_CH48.qxd
10/15/07
5:05 PM
Page 503
X=
X1 + X 2 + X 3 + + X k
k
X=
503
X = 7.191
R1 + R2 + R3 + + Rk
k
R=
R = 0.257
The control limits are based on an average 3 standard deviations. Three standard
deviations for the distribution of averages are calculated using the following relationship:
3S X = A2 R ,
3S X = (0.577)(0.257) = 0.148.
Note: The range portion for this example of the zone format average/range chart is not presented, as the range portion is performed exactly as with a traditional range chart. See the
module entitled Average/Range Control Chart.
One standard deviation for the distribution of averages is determined by dividing 3SX
by three.
SX = 0.049
Zones in increments of one standard deviation about the average are determined, and
a zone-weighting scale is assigned for each of the 1, 2, and 3 standard deviation zones.
The following table lists the zone-definition area and the corresponding zone score.
Zone location
>X + 3S
X + 2S to X + 3S
X + 1S to X + 2S
X to X + 1S
X 1S to X
X 1S to X 2S
X 2S to X 3S
<X 3S
>7.338
7.289 to
7.240 to
7.191 to
7.142 to
7.093 to
7.093 to
<7.044
7.338
7.289
7.240
7.191
7.142
7.044
Zone score
8
4
2
1
1
2
4
8
H1317_CH48.qxd
504
10/15/07
5:05 PM
Page 504
These zone values and the zone-weighting factors are recorded on the following zone
format X /R chart:
+3S = 7.338
+2S = 7.289
+1S = 7.240
Avg = 7.191
1S = 7.142
2S = 7.093
3S = 7.044
8
4
2
1
1
2
4
8
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27
Each of the averages is plotted in the appropriate zone location. The following rules
of protocol are used with the zone charting technique:
1. Plotting points are circled with the appropriate zone score placed inside the circle.
2. Continue to accumulate with the current zone score the previous circled zone score
as long as the positions remain on the same side of the average line. Connect the
circled values with a line.
3. If the average line is crossed going from one zone area to another, start the accumulation process over.
4. An indication of an out-of-control condition is evident when a cumulated zone
score of greater than or equal to eight is obtained.
The first few values are plotted to illustrate the concept.
Sample #1
An average of 7.19 is located in zone 1 below the average. A zone score of 1 is placed in
this area:
+3S = 7.338
+2S = 7.289
+1S = 7.240
Avg = 7.191
1S = 7.142
2S = 7.093
3S = 7.044
8
4
2
1
1
2
4
8
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27
H1317_CH48.qxd
10/15/07
5:05 PM
Page 505
505
Sample #2
An average of 7.16 is located in zone 1 below the average. The average has not been
crossed; therefore, we add the previous value of 1 to the current zone area of 1 for a total
zone score of 2. The 2 is circled and connected to the preceding value.
+3S = 7.338
+2S = 7.289
+1S = 7.240
Avg = 7.191
1S = 7.142
2S = 7.093
3S = 7.044
8
4
2
1
1
2
4
8
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27
Sample #3
An average of 7.21 is located in zone 1 above the average. The average has been crossed
going from sample #2 to the current position for sample #3; therefore, the accumulation
process is started over. The correct value for this zone position is 1.
+3S = 7.338
+2S = 7.289
+1S = 7.240
Avg = 7.191
1S = 7.142
2S = 7.093
3S = 7.044
8
4
2
1
1
1
1
2
4
8
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27
Sample #4
An average of 7.14 is located in zone 2 below the average. The average has been crossed
going from the previous value to the current sample #4; therefore, the accumulation
process is started over.
H1317_CH48.qxd
10/15/07
506
5:05 PM
Page 506
+3S = 7.338
+2S = 7.289
+1S = 7.240
Avg = 7.191
1S = 7.142
2S = 7.093
3S = 7.044
8
4
2
1
1
1
1
4
8
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27
+3S = 7.338
+2S = 7.289
+1S = 7.240
Avg = 7.191
1S = 7.142
2S = 7.093
3S = 7.044
History
4
2
1
1
1
1
3
2
4
8
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27
A change has been made in the operation of the process. Based on the next nine sets
of data, do you have evidence to support the conclusion that the process data reflect any
change?
#19
#20
#21
#22
#23
#24
7.20
7.02
7.17
7.05
7.06
7.42
7.22
6.98
7.24
6.94
7.14
7.02
7.27
7.22
7.22
6.94
7.06
7.18
7.10
7.02
7.06
7.04
7.07
7.16
6.92
7.13
7.24
6.97
7.23
7.41
X = 7.10
R = 0.18
X = 7.16
R = 0.48
X = 7.17
R = 0.25
X = 7.06
R = 0.24
X = 7.05
R = 0.24
X = 7.20
R = 0.44
H1317_CH48.qxd
10/15/07
5:05 PM
Page 507
#25
#26
#27
7.04
7.12
7.35
7.11
7.06
7.05
7.11
7.01
7.06
7.07
6.93
7.05
7.17
7.02
7.08
X = 7.14
R = 0.31
X = 7.06
R = 0.10
X = 7.05
R = 0.24
+3S = 7.338
+2S = 7.289
+1S = 7.240
Avg = 7.191
1S = 7.142
2S = 7.093
3S = 7.044
507
History
4
2
1
1
1
1
3
2
1
1
2
8
6 10
8
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27
Bibliography
Jaehn, A. H. 1987. Zone Control Charts: A New Tool for Quality Control. Tappi Journal 87 (February): 159161.
H1317_CH48.qxd
10/15/07
5:05 PM
Page 508
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 509
Appendix: Tables
x.x0
x.x1
x.x2
x.x3
x.x4
x.x5
x.x6
x.x7
x.x8
x.x9
4.0
0.00003
0.00003
0.00003
0.00003
0.00003
0.00003
0.00003
0.00002
0.00002
0.00002
3.9
3.8
3.7
3.6
3.5
0.00005
0.00007
0.00011
0.00016
0.00023
0.00005
0.00007
0.00010
0.00015
0.00022
0.00004
0.00007
0.00010
0.00015
0.00022
0.00004
0.00006
0.00010
0.00014
0.00021
0.00004
0.00006
0.00009
0.00014
0.00020
0.00004
0.00006
0.00009
0.00013
0.00019
0.00004
0.00006
0.00008
0.00013
0.00019
0.00004
0.00005
0.00008
0.00012
0.00018
0.00003
0.00005
0.00008
0.00012
0.00017
0.00003
0.00005
0.00008
0.00011
0.00017
3.4
3.3
3.2
3.1
3.0
0.00034
0.00048
0.00069
0.00097
0.00135
0.00032
0.00047
0.00066
0.00094
0.00131
0.00031
0.00045
0.00064
0.00090
0.00126
0.00030
0.00043
0.00062
0.00087
0.00122
0.00029
0.00042
0.00060
0.00084
0.00118
0.00028
0.00040
0.00058
0.00082
0.00114
0.00027
0.00039
0.00056
0.00079
0.00111
0.00026
0.00038
0.00054
0.00076
0.00107
0.00025
0.00036
0.00052
0.00074
0.00104
0.00024
0.00035
0.00050
0.00071
0.00100
2.9
2.8
2.7
2.6
2.5
0.00187
0.00256
0.00347
0.00466
0.00621
0.00181
0.00248
0.00336
0.00453
0.00604
0.00175
0.00240
0.00326
0.00440
0.00587
0.00170
0.00233
0.00317
0.00427
0.00570
0.00164
0.00226
0.00307
0.00415
0.00554
0.00159
0.00219
0.00298
0.00402
0.00539
0.00154
0.00212
0.00289
0.00391
0.00523
0.00149
0.00205
0.00280
0.00379
0.00509
0.00144
0.00199
0.00272
0.00368
0.00494
0.00140
0.00193
0.00264
0.00357
0.00480
2.4
2.3
2.2
2.1
2.0
0.00820
0.01072
0.01390
0.01787
0.02275
0.00798
0.01044
0.01355
0.01743
0.02222
0.00776
0.01017
0.01321
0.01700
0.02169
0.00755
0.00990
0.01287
0.01659
0.02118
0.00734
0.00964
0.01255
0.01618
0.02068
0.00714
0.00939
0.01223
0.01578
0.02018
0.00695
0.00914
0.01191
0.01539
0.01970
0.00676
0.00889
0.01160
0.01500
0.01923
0.00657
0.00866
0.01130
0.01463
0.01876
0.00639
0.00842
0.01101
0.01426
0.01831
1.9
1.8
1.7
1.6
1.5
0.02872
0.03593
0.04457
0.05480
0.06681
0.02807
0.03515
0.04363
0.05370
0.06552
0.02743
0.03438
0.04272
0.05262
0.06426
0.02680
0.03363
0.04182
0.05155
0.06301
0.02619
0.03288
0.04093
0.05050
0.06178
0.02559
0.03216
0.04006
0.04947
0.06057
0.02500
0.03144
0.03920
0.04846
0.05938
0.02442
0.03074
0.03836
0.04746
0.05821
0.02385
0.03005
0.03754
0.04648
0.05705
0.02330
0.02938
0.03673
0.04551
0.05592
1.4
1.3
1.2
1.1
1.0
0.08076
0.09680
0.11507
0.13567
0.15866
0.07927
0.09510
0.11314
0.13350
0.15625
0.07780
0.09342
0.11123
0.13136
0.15386
0.07636
0.09176
0.10935
0.12924
0.15151
0.07493
0.09012
0.10749
0.12714
0.14917
0.07353
0.08851
0.10565
0.12507
0.14686
0.07215
0.08692
0.10384
0.12302
0.14457
0.07078
0.08534
0.10204
0.12100
0.14231
0.06944
0.08379
0.10027
0.11900
0.14007
0.06811
0.08226
0.09853
0.11702
0.13786
0.9
0.8
0.7
0.6
0.5
0.18406
0.21186
0.24196
0.27425
0.30854
0.18141
0.20897
0.23885
0.27093
0.30503
0.17879
0.20611
0.23576
0.26763
0.30153
0.17619
0.20327
0.23270
0.26435
0.29806
0.17361
0.20045
0.22965
0.26109
0.29460
0.17106
0.19766
0.22663
0.25785
0.29116
0.16853
0.19490
0.22363
0.25463
0.28774
0.16602
0.19215
0.22065
0.25143
0.28434
0.16354
0.18943
0.21770
0.24825
0.28096
0.16109
0.18673
0.21476
0.24510
0.27760
0.4
0.3
0.2
0.1
0.0
0.34458
0.38209
0.42074
0.46017
0.50000
0.34090
0.37828
0.41683
0.45621
0.49601
0.33724
0.37448
0.41294
0.45224
0.49292
0.33360
0.37070
0.40905
0.44828
0.48803
0.32997
0.36693
0.40517
0.44433
0.48405
0.32636
0.36317
0.40129
0.44038
0.48006
0.32276
0.35942
0.39743
0.43644
0.47608
0.31918
0.35569
0.39358
0.43251
0.47210
0.31561
0.35197
0.38974
0.42858
0.46812
0.31207
0.34827
0.38591
0.42466
0.46414
509
H1317_CH49-AP.qxd
510
10/15/07
4:14 PM
Page 510
0.00
0.01
0.02
0.03
0.04
0.05
0.06
0.07
0.08
0.09
4.0
4.1
4.2
4.3
4.4
4.5
4.6
4.7
4.8
4.9
3.18E-05
2.08E-05
1.34E-05
8.62E-06
5.48E-06
3.45E-06
2.15E-06
1.33E-06
8.18E-07
4.98E-07
3.05E-05
1.99E-05
1.29E-05
8.24E-06
5.23E-06
3.29E-06
2.05E-06
1.27E-06
7.79E-07
4.73E-07
2.92E-05
1.91E-05
1.23E-05
7.88E-06
5.00E-06
3.14E-06
1.96E-06
1.21E-06
7.41E-07
4.50E-07
2.80E-05
1.82E-05
1.18E-05
7.53E-06
4.77E-06
3.00E-06
1.87E-06
1.15E-06
7.05E-07
4.28E-07
2.68E-05
1.75E-05
1.13E-05
7.20E-06
4.56E-06
2.86E-06
1.78E-06
1.10E-06
6.71E-07
4.07E-07
2.57E-05
1.67E-05
1.08E-05
6.88E-06
4.35E-06
2.73E-06
1.70E-06
1.05E-06
6.39E-07
3.78E-07
2.47E-05
1.60E-05
1.03E-05
6.57E-06
4.16E-06
2.60E-06
1.62E-06
9.96E-07
6.00E-07
3.68E-07
2.36E-05
1.53E-05
9.86E-06
6.28E-06
3.97E-06
2.48E-06
1.54E-06
9.48E-07
5.78E-07
3.50E-07
2.26E-05
1.47E-05
9.43E-06
6.00E-06
3.79E-06
2.37E-06
1.47E-06
9.03E-01
5.50E-07
3.32E-07
2.17E-05
1.40E-05
9.01E-06
5.73E-06
3.62E-06
2.26E-06
1.40E-06
8.59E-07
5.23E-07
3.16E-07
5.0
5.1
5.2
5.3
5.4
5.5
5.6
5.7
5.8
5.9
3.00E-07
1.80E-07
1.07E-07
6.27E-08
3.66E-08
2.12E-08
1.22E-08
6.98E-09
3.96E-09
2.23E-09
2.85E-07
1.71E-07
1.01E-07
5.95E-08
3.47E-08
2.01E-08
1.16E-08
6.60E-09
3.74E-09
2.11E-09
2.71E-07
1.62E-07
9.59E-08
5.64E-08
3.29E-08
1.90E-08
1.09E-08
6.24E-09
3.53E-09
1.99E-09
2.58E-07
1.54E-07
9.10E-08
5.34E-08
3.11E-08
1.80E-08
1.03E-08
5.89E-09
3.34E-09
1.88E-09
2.45E-07
1.46E-07
8.63E-08
5.06E-08
2.95E-08
1.70E-08
9.78E-09
5.57E-09
3.15E-09
1.77E-09
2.32E-07
1.39E-07
8.18E-08
4.80E-08
2.79E-08
1.61E-08
9.24E-09
5.26E-09
2.97E-09
1.67E-09
2.21E-07
1.31E-07
7.76E-08
4.55E-08
2.64E-08
1.53E-08
8.74E-09
4.97E-09
2.81E-09
1.58E-09
2.10E-07
1.125-07
7.36E-08
4.31E-08
2.50E-08
1.44E-08
8.26E-09
4.70E-09
2.65E-09
1.49E-09
1.99E-07
1.18E-07
6.98E-08
4.08E-08
2.37E-08
1.37E-08
7.81E-09
4.44E-09
2.50E-09
1.40E-09
1.89E-07
1.12E-07
6.62E-08
3.87E-08
2.24E-08
1.29E-08
7.39E-09
4.19E-09
2.36E-09
1.32E-09
6.0
6.1
6.2
6.3
6.4
6.5
6.6
6.7
6.8
6.9
1.25E-09
6.94E-10
3.84E-10
2.11E-10
1.15E-10
6.25E-11
3.38E-11
1.82E-11
9.72E-12
5.18E-12
1.18E-09
6.54E-10
3.61E-10
1.98E-10
1.08E-10
5.88E-11
3.18E-11
1.71E-11
9.13E-12
4.86E-12
1.11E-09
6.17E-10
3.40E-10
1.87E-10
1.02E-11
5.53E-11
2.98E-11
1.60E-11
8.57E-12
4.56E-12
1.05E-09
5.81E-10
3.21E-10
1.76E-10
9.59E-10
5.20E-11
2.81E-11
1.51E-11
8.05E-12
4.28E-12
9.88E-10
5.48E-10
3.02E-10
1.66E-10
9.02E-11
4.89E-11
2.64E-11
1.42E-11
7.56E-12
4.02E-12
9.31E-10
5.16E-10
2.84E-10
1.56E-10
8.49E-11
4.60E-11
2.48E-11
1.33E-11
7.10E-12
3.77E-12
8.78E-10
4.87E-10
2.68E-10
1.47E-10
7.98E-11
4.32E-11
2.33E-11
1.25E-11
6.66E-12
3.54E-12
8.2BE-10
4.59E-10
2.52E-10
1.38E-10
7.51E-11
4.07E-11
2.19E-11
1.17E-11
6.26E-12
3.32E-12
7.81E-10
4.32E-10
2.38E-10
1.30E-10
7.06E-11
3.82E-11
2.06E-11
1.10E-11
5.87E-12
3.12E-12
7.36E-10
4.07E-10
2.24E-10
1.22E-10
6.65E-11
3.59E-11
1.93E-11
1.04E11
5.52E-12
2.93E-12
7.0
7.1
7.2
7.3
7.4
7.5
7.6
7.7
7.8
7.9
2.75E-12
1.45E-12
7.64E-13
4.01E-13
2.10E-13
1.09E-13
5.68E-14
2.94E-14
1.52E-14
7.85E-15
2.58E-12
1.36E-12
7.16E-13
3.76E-13
1.96E-13
1.02E-13
5.32E-14
2.76E-14
1.42E-14
7.35E-15
2.42E-12
1.28E-12
6.72E-13
3.52E-13
1.84E-13
9.58E-14
4.98E-14
2.58E-14
1.33E-14
6.88E-15
2.27E-12
1.20E-12
6.30E-13
3.30E-13
1.72E-13
8.98E-14
4.66E-14
2.42E-14
1.25E-14
6.44E-15
2.13E-12
1.12E-12
5.90E-13
3.09E-13
1.62E-13
8.41E-14
4.37E-14
2.26E-14
1.17E-14
6.02E-15
2.00E-12
1.05E-12
5.54E-13
2.90E-13
1.51E-13
7.87E-14
4.09E-14
2.12E-14
1.09E-14
5.64E-15
1.87E-12
9.88E-13
5.19E-13
2.72E-13
1.42E-13
7.38E-14
3.83E-14
1.98E-14
1.02E-14
5.28E-15
1.76E-12
9.26E-13
4.86E-13
2.55E-13
1.33E-13
6.91E-14
3.58E-14
1.86E-14
9.58E-15
4.94E-15
1.65E-12
8.69E-13
4.56E-13
2.39E-13
1.24E-13
6.47E14
3.36E-14
1.74E-14
8.97E-15
4.62E-15
1.55E-12
8.15E-13
4.28E-13
2.24E-13
1.17E-13
6.06E-14
3.14E-14
1.63E-14
8.39E-15
4.32E-15
8.0
8.1
8.2
8.3
8.4
8.5
8.6
8.7
8.8
8.9
4.05E-15
2.08E-15
1.07E-15
5.48E-16
2.81E-16
1.44E-16
7.34E-17
3.75E-17
1.92E-17
9.79E-18
3.79E-15
1.95E-15
9.99E-16
5.12E-16
2.62E-16
1.34E-16
6.87E-17
3.51E-17
1.79E-17
9.16E-18
3.54E-15
1.82E-15
9.35E-16
4.79E-16
2.45E-16
1.26E-16
6.42E-17
3.28E-17
1.68E-17
8.56E-18
3.31E-15
1.70E-15
8.74E-16
4.48E-16
2.30E-16
1.17E-16
6.00E-17
3.07E-17
1.57E-17
8.00E-18
3.10E-15
1.59E-15
8.18E-16
4.19E-16
2.15E-16
1.10E-16
5.61E-17
2.87E-17
1.47E-17
7.48E-18
2.90E-15
1.49E-15
7.65E-16
3.92E-16
2.01E-16
1.03E-16
5.25E-17
2.68E-17
1.37E-17
7.00E-18
2.72E-15
1.40E-15
7.16E-16
3.67E-17
1.88E-16
9.60E-17
4.91E-17
2.51E-17
1.28E-17
6.54E-18
2.54E-15
1.31E-15
6.69E-16
3.43E-16
1.76E-16
8.98E-17
4.59E-17
2.35E-17
1.20E-17
6.12E-18
2.38E-15
1.22E-15
6.26E-16
3.21E-16
1.64E-16
8.40E-17
4.29E-17
2.19E-17
1.12E-17
5.72E-18
2.22E-15
1.14E-15
5.86E-16
3.00E-16
1.54E-16
7.85E-17
4.01E-17
2.05E-17
1.05E-17
5.35E-18
9.0
9.1
9.2
9.3
9.4
9.5
9.6
9.7
9.8
9.9
10.0
5.00E-18
2.56E-18
1.31E-18
6.67E-19
3.41E-19
1.75E-19
8.94E-20
4.58E-20
2.35E-20
1.21E-20
6.22E-21
4.68E-18
2.39E-18
1.22E-18
6.24E-19
3.19E-19
1.63E-19
8.37E-20
4.29E-20
2.20E-20
1.13E-20
5.82E-21
4.37E-18
2.23E-18
1.14E-18
5.83E-19
2.98E-19
1.53E-19
7.82E-20
4.01E-20
2.06E-20
1.06E-20
5.44E-21
4.09E-18
2.09E-18
1.07E-18
5.46E-19
2.79E-19
1.43E-19
7.32E-20
3.75E-20
1.93E-20
9.90E-21
5.09E-21
3.82E-18
1.95E-18
9.98E-19
5.10E-19
2.61E-19
1.34E-19
6.85E-20
3.51E-20
1.80E-20
9.26E-21
4.77E-21
3.57E-18
1.83E-18
9.33E-19
4.77E-19
2.44E-19
1.25E-19
6.40E-20
3.28E-20
1.69E-20
8.67E-21
4.46E-21
3.34E-18
1.71E-18
8.73E-19
4.46E-19
2.2BE-19
1.17E-19
5.99E-20
3.07E-20
1.58E-20
8.11E-21
4.17E-21
3.13E-18
1.60E-18
8.16E-19
4.17E-19
2.14E-19
1.09E-19
5.60E-20
2.87E-20
1.48E-20
7.59E-21
3.91E-21
2.92E-18
1.49E-18
7.63E-19
3.90E-19
2.00E-19
1.02E-19
5.24E-20
2.69E-20
1.38E-20
7.10E-21
3.66E-21
2.73E-18
1.40E-18
7.14E-19
3.65E-19
1.87E-19
9.56E-20
4.90E-20
2.52E-20
1.29E-20
6.64E-21
3.42E-21
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 511
Appendix: Tables
511
0.95
0.90
0.75
0.50
0.25
0.10
0.05
C
0
1
2
3
4
5
6
7
8
9
10
0.051
0.355
0.818
1.366
1.970
2.613
3.286
3.981
4.695
5.426
6.169
0.105
0.532
1.102
1.745
2.433
3.152
3.895
4.656
5.432
6.221
7.021
0.288
0.961
1.727
2.523
3.369
4.219
5.083
5.956
6.838
7.726
8.620
0.693
1.678
2.674
3.672
4.671
5.670
6.670
7.669
8.669
9.669
10.668
1.386
2.693
3.920
5.109
6.274
7.423
8.558
9.684
10.802
11.914
13.020
2.303
3.890
5.322
6.681
7.994
9.275
10.532
11.771
12.995
14.206
15.407
2.996
4.744
6.296
7.754
9.154
10.513
11.842
13.148
14.434
15.705
16.962
11
12
13
14
15
16
17
18
19
20
6.924
7.690
8.464
9.246
10.035
10.831
11.663
12.442
13.254
14.072
7.829
8.646
9.470
10.300
11.135
11.976
12.822
13.672
14.525
15.383
9.519
10.422
11.329
12.239
13.152
14.068
14.986
15.907
16.830
17.755
11.668
12.668
13.668
14.668
15.668
16.668
17.668
18.668
19.668
20.668
14.121
15.217
16.310
17.400
18.486
19.570
20.652
21.731
22.808
23.883
16.598
17.782
18.958
20.128
21.292
22.452
23.606
24.756
25.902
27.045
18.208
19.442
20.668
21.886
23.098
24.302
25.500
26.692
27.879
29.062
21
22
23
24
25
26
27
28
29
30
14.894
15.719
16.548
17.382
18.218
19.058
19.900
20.746
21.594
22.444
16.244
17.108
17.975
18.844
19.717
20.592
21.469
22.348
23.229
24.113
18.682
19.610
20.540
21.471
22.404
23.338
24.237
25.209
26.147
27.086
21.668
22.668
23.668
24.668
25.667
26.667
27.667
28.667
29.667
30.667
24.956
26.028
27.098
28.167
29.234
30.300
31.365
32.428
33.491
34.552
28.184
29.320
30.453
31.584
32.711
33.836
34.959
36.080
37.198
38.315
30.241
31.416
32.586
33.752
34.916
36.077
37.234
38.389
39.541
40.690
31
32
33
34
35
36
37
23.298
24.152
25.010
25.870
26.731
27.594
28.460
24.998
25.885
26.774
27.664
28.556
29.450
30.345
28.025
28.966
29.907
30.849
31.792
32.736
33.681
31.667
32.667
33.667
34.667
35.667
36.667
37.667
35.613
36.672
37.731
38.788
39.845
40.901
41.957
39.430
40.543
41.654
42.764
43.872
44.978
46.082
41.838
42.982
44.125
45.226
46.404
47.540
48.676
H1317_CH49-AP.qxd
512
10/15/07
4:14 PM
Page 512
` = 0.05
a = 0.10
` = 0.05
a = 0.05
` = 0.05
a = 0.01
np 1
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
44.89
10.946
6.509
4.890
4.057
3.549
3.206
2.957
2.768
2.618
2.479
2.379
2.312
2.240
2.177
2.122
2.073
2.029
1.990
1.954
1.922
1.892
1.865
1.840
1.817
1.795
1.775
1.757
1.739
1.723
1.707
1.692
1.679
1.665
1.653
1.641
1.630
1.619
1.609
1.599
1.590
58.404
13.349
7.699
5.675
4.646
4.023
3.604
3.303
3.074
2.895
2.750
2.630
2.528
2.442
2.367
2.302
2.244
2.192
2.145
2.103
2.065
2.030
1.999
1.969
1.942
1.917
1.893
1.871
1.850
1.831
1.813
1.796
1.780
1.764
1.750
1.736
1.723
1.710
1.698
1.687
1.676
89.781
18.781
10.280
7.352
5.890
5.017
4.435
4.019
3.707
3.462
3.265
3.104
2.968
2.852
2.752
2.665
2.588
2.520
2.458
2.403
2.352
2.307
2.265
2.226
2.191
2.158
2.127
2.098
2.071
2.046
2.023
2.001
1.980
1.960
1.941
1.923
1.906
1.890
1.875
1.860
1.846
0.052
0.355
0.818
1.366
1.970
2.613
3.286
3.981
4.695
5.426
6.169
6.924
7.690
8.464
9.246
10.035
10.831
11.633
12.442
13.254
14.072
14.894
15.719
16.548
17.382
18.218
19.058
19.900
20.746
21.594
22.444
23.298
24.152
25.010
25.870
26.731
27.594
28.460
29.327
30.196
31.066
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 513
Appendix: Tables
513
0.10
0.05
0.025
0.010
0.005
0.0025
1
2
3
4
5
6
7
8
9
10
3.078
1.886
1.638
1.533
1.476
1.440
1.415
1.397
1.383
1.372
6.314
2.920
2.353
2.132
2.015
1.943
1.895
1.860
1.833
1.812
12.706
4.303
3.182
2.776
2.571
2.447
2.365
2.306
2.262
2.228
31.820
6.965
4.541
3.747
3.365
3.143
2.998
2.896
2.821
2.764
63.657
9.925
5.841
4.604
4.032
3.707
3.499
3.355
3.250
3.169
127.32
14.089
7.453
5.598
4.773
4.317
4.019
3.833
3.690
3.581
11
12
13
14
15
16
17
18
19
20
1.363
1.356
1.350
1.345
1.341
1.337
1.333
1.330
1.328
1.325
1.796
1.782
1.771
1.761
1.753
1.746
1.740
1.734
1.729
1.725
2.201
2.179
2.160
2.145
2.131
2.120
2.110
2.101
2.093
2.086
2.718
2.681
2.650
2.624
2.602
2.583
2.567
2.552
2.539
2.528
3.106
3.055
3.012
2.977
2.947
2.921
2.898
2.878
2.861
2.845
3.497
3.428
3.372
3.326
3.286
3.252
3.222
3.197
3.174
3.153
21
22
23
24
25
26
27
28
29
30
1.323
1.321
1.319
1.318
1.316
1.315
1.314
1.313
1.311
1.310
1.721
1.717
1.714
1.711
1.708
1.706
1.703
1.701
1.699
1.697
2.080
2.074
2.069
2.064
2.060
2.056
2.052
2.048
2.045
2.042
2.518
2.508
2.500
2.492
2.485
2.479
2.473
2.467
2.462
2.457
2.831
2.819
2.807
2.797
2.787
2.779
2.771
2.763
2.756
2.750
3.135
3.119
3.104
3.091
3.078
3.067
3.057
3.047
3.038
3.030
40
60
120
1.303
1.296
1.289
1.684
1.671
1.658
2.021
2.000
1.980
2.423
2.390
2.358
2.704
2.660
2.617
2.971
2.915
2.860
1.282
1.645
1.960
2.326
2.576
2.807
H1317_CH49-AP.qxd
514
10/15/07
4:14 PM
Page 514
0.995
0.975
0.950
0.90
0.100
0.050
0.025
0.005
0.0439
0.010
0.072
0.207
0.412
0.676
0.989
1.34
1.73
2.16
2.60
3.07
3.57
4.07
4.60
5.14
5.70
6.26
6.84
7.43
8.03
8.64
9.26
9.89
10.5
11.2
11.8
12.5
13.1
13.8
20.7
28.0
35.5
43.3
51.2
59.2
67.3
0.0398
0.020
0.216
0.484
0.831
1.24
1.69
2.18
2.70
3.25
3.82
4.40
5.01
5.63
6.26
6.91
7.56
8.23
8.91
9.59
10.3
11.0
11.7
12.4
13.1
13.8
14.6
15.3
16.0
16.8
24.4
32.4
40.5
48.8
57.2
65.7
74.2
0.003
0.103
0.352
0.711
1.15
1.64
2.17
2.73
3.33
3.94
4.57
5.23
5.89
6.57
7.26
7.96
8.67
9.39
10.1
10.9
11.6
12.3
13.1
13.8
14.6
15.4
16.2
16.9
17.7
18.5
26.5
34.8
43.2
51.7
60.4
69.1
77.9
0.01
0.21
0.58
1.06
1.61
2.20
2.83
3.49
4.17
4.87
5.58
6.30
7.04
7.79
8.55
9.31
10.09
10.87
11.65
12.44
13.24
14.04
14.85
15.66
16.47
17.29
18.11
18.94
19.77
20.60
29.05
37.69
46.46
55.33
64.28
73.29
82.36
2.71
4.61
6.25
7.78
9.24
10.6
12.0
13.4
14.7
16.0
17.3
18.5
19.8
21.1
22.3
23.5
24.8
26.0
27.2
28.4
29.6
30.8
32.0
33.2
34.4
35.6
36.7
37.9
39.1
40.3
51.8
63.2
74.4
85.5
96.6
107.6
118.5
3.84
5.99
7.82
9.49
11.1
12.6
14.1
15.5
16.9
18.3
19.7
21.0
22.4
23.7
25.0
26.3
27.6
28.9
30.1
31.4
32.7
33.9
35.2
36.4
37.7
38.9
40.1
41.3
42.6
43.8
55.8
67.5
79.1
90.5
101.9
113.1
124.3
5.02
7.38
9.35
11.1
12.8
14.4
16.0
17.5
19.0
20.5
21.9
23.3
24.7
26.1
27.5
28.8
30.2
31.5
32.9
34.2
35.5
36.8
38.1
39.4
40.6
41.9
43.2
44.5
45.7
47.0
59.3
71.4
83.3
95.0
106.6
118.1
129.6
7.88
10.6
12.8
14.9
16.7
18.5
20.3
22.0
23.6
25.2
26.8
28.3
29.8
31.3
32.8
34.3
35.7
37.2
38.6
40.0
41.4
42.8
44.2
45.6
46.9
48.3
49.6
51.0
52.3
53.7
66.7
79.5
92.0
104.2
116.3
128.3
140.2
For number of degrees of freedom > 100, calculate the approximate normal deviation
Z = 2 X 2 ( df ) 1 .
2
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 515
Appendix: Tables
515
A2
d2
D3
D4
A3
c4
B3
B4
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
1.880
1.023
0.729
0.577
0.483
0.419
0.373
0.337
0.308
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
1.128
1.693
2.059
2.326
2.534
2.704
2.847
2.970
3.078
3.173
3.258
3.336
3.407
3.472
3.532
3.588
3.640
3.689
3.735
3.778
3.819
3.858
3.895
3.931
0.076
0.136
0.184
0.223
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
3.267
2.574
2.282
2.114
2.004
1.924
1.864
1.816
1.777
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
2.659
1.954
1.628
1.427
1.287
1.182
1.099
1.032
0.975
0.927
0.886
0.850
0.817
0.789
0.763
0.739
0.718
0.698
0.680
0.663
0.647
0.633
0.619
0.606
0.7979
0.8862
0.9213
0.9400
0.9515
0.9594
0.9650
0.9693
0.9727
0.9754
0.9776
0.9794
0.9810
0.9823
0.9835
0.9845
0.9854
0.9862
0.9869
0.9876
0.9882
0.9887
0.9892
0.9896
0.030
0.118
0.185
0.239
0.284
0.321
0.354
0.382
0.406
0.428
0.448
0.466
0.482
0.497
0.510
0.523
0.534
0.545
0.555
0.565
3.267
2.568
2.266
2.089
1.970
1.882
1.816
1.761
1.716
1.679
1.646
1.618
1.594
1.572
1.552
1.534
1.518
1.503
1.490
1.477
1.466
1.455
1.445
1.438
Note: Factors A2, D3, and D4 are not appropriate (NA) for these sample sizes. Use the average/
standard deviation control chart where n > 10.
E2
2
3
4
5
6
7
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
2.660
1.772
1.457
1.290
1.184
1.109
0.946
0.921
0.899
0.881
0.864
0.849
0.836
0.824
0.813
0.803
0.794
0.785
0.778
0.770
0.763
1.880
1.187
0.796
0.691
0.548
0.508
0.350
0.317
0.308
0.283
0.276
0.258
0.252
0.237
0.232
0.220
0.216
0.206
0.203
0.194
0.191
H1317_CH49-AP.qxd
516
10/15/07
4:14 PM
Page 516
BfMR
BfCR
Mr
Cr
BfMR
BfCR
0.01
0.02
0.03
0.04
0.05
0.06
0.07
0.08
0.09
0.10
0.11
0.12
0.13
0.14
0.15
0.16
0.17
0.18
0.19
0.20
0.21
0.22
0.23
0.24
0.25
.0264
.0488
.0681
.0863
.1040
.1200
.1334
.1497
.1626
.1758
.1872
.1989
.2107
.2227
.2323
.2444
.2541
.2613
.2709
.2803
.2874
.2966
.3034
.3101
.3187
0.26
0.27
0.28
0.29
0.30
0.31
0.32
0.33
0.34
0.35
0.36
0.37
0.38
0.39
0.40
0.41
0.42
0.43
0.44
0.45
0.46
0.47
0.48
0.49
0.50
.3251
.3312
.3372
.3420
.3485
.3538
.3572
.3621
.3668
.3712
.3739
.3778
.3814
.3836
.3867
.3885
.3910
.3925
.3945
.3961
.3970
.3977
.3984
.3989
.3989
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 517
Appendix: Tables
517
Table A.10 Critical values of smaller rank sum for Wilcoxon-Mann-Whitney test.
n1 (smaller sample)
n2
0.10
0.05
0.025
7
6
0.10
0.05
0.025
7
6
13
11
10
0.10
0.05
0.025
8
7
6
14
12
11
20
19
17
0.10
0.05
0.025
9
8
7
15
13
12
22
20
18
30
28
26
0.10
0.05
0.025
10
8
7
16
14
13
23
21
20
32
29
27
41
39
36
0.10
0.05
0.025
11
9
8
17
15
14
25
23
21
34
31
29
44
41
38
55
51
49
0.10
0.05
0.025
11
10
8
19
16
14
27
24
22
36
33
31
46
43
40
58
54
51
70
66
62
10
0.10
0.05
0.025
12
10
9
20
17
15
28
26
23
38
35
32
49
45
42
60
56
53
73
69
65
87
82
78
11
0.10
0.05
0.025
13
11
9
21
18
16
30
27
24
40
37
34
51
47
44
63
59
55
76
72
68
91
86
81
106
100
96
12
0.10
0.05
0.025
14
11
10
22
19
17
32
28
26
42
38
35
54
49
46
66
62
58
80
75
71
94
89
84
110
104
99
127
120
115
13
0.10
0.05
0.025
15
12
10
23
20
18
33
30
27
44
40
37
56
52
48
69
64
60
83
78
73
98
92
88
114
108
103
131
125
119
149
142
136
14
0.10
0.05
0.025
16
13
11
25
21
19
35
31
28
46
42
38
59
54
50
72
67
62
86
81
76
102
96
91
118
112
106
136
129
123
154
147
141
174
166
160
15
0.10
0.05
0.025
16
13
11
26
22
20
37
33
29
48
44
40
61
56
52
75
69
65
90
84
79
106
99
94
123
116
110
141
133
127
159
152
145
179
171
164
10
11
12
13
14
15
200
192
184
H1317_CH49-AP.qxd
518
10/15/07
4:14 PM
Page 518
0.018 0.033 0.046 0.074 0.113 0.143 0.198 0.33 0.53 0.79 1.22 1.90 2.90 4.94 7.12 11.46
1/2
1/3
1/4
1/5
1/7
1/10
1/15
1/25
1/50
1/100
1/200
1540
2550
3340
3960
4950
6050
7390
9110
11730
14320
17420
840
1390
1820
2160
2700
3300
4030
4970
6400
7810
9500
600
1000
1310
1550
1940
2370
2890
3570
4590
5600
6810
375
620
810
965
1205
1470
1800
2215
2855
3485
4235
245
405
530
630
790
965
1180
1450
1870
2305
2760
194
321
420
498
623
762
930
1147
1477
1820
2178
140
232
303
360
450
550
672
828
1067
1302
1583
84
140
182
217
270
335
410
500
640
790
950
53
87
113
135
168
207
255
315
400
500
590
36
59
76
91
113
138
170
210
270
330
400
23
38
49
58
73
89
108
134
175
215
255
15
25
32
38
47
57
70
86
110
135
165
10
16
21
25
31
38
46
57
72
89
106
6
10
13
15
18
22
27
33
42
52
62
5
7
9
11
13
16
19
23
29
36
43
3
5
6
7
8
10
12
14
18
22
26
0.010
0.015
0.025
0.040
0.065
0.10
0.15
0.25
0.40
0.65
1.0
1.5
2.5
4.0
6.5
10.0
AQL(%)
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 519
Appendix: Tables
519
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1.0
0
1
2
3
4
5
6
7
0.905
0.995
1.000
1.000
1.000
1.000
1.000
1.000
0.819
0.983
0.999
1.000
1.000
1.000
1.000
1.000
0.741
0.963
0.996
1.000
1.000
1.000
1.000
1.000
0.670
0.938
0.992
0.999
1.000
1.000
1.000
1.000
0.607
0.910
0.986
0.998
1.000
1.000
1.000
1.000
0.549
0.878
0.977
0.997
1.000
1.000
1.000
1.000
0.497
0.844
0.966
0.994
0.999
1.000
1.000
1.000
0.449
0.809
0.953
0.991
0.999
1.000
1.000
1.000
0.407
0.773
0.937
0.987
0.998
1.000
1.000
1.000
0.368
0.736
0.920
0.981
0.996
0.999
1.000
1.000
1.1
1.2
1.3
1.4
1.5
1.6
1.7
1.8
1.9
2.0
0
1
2
3
4
5
6
7
8
9
0.333
0.699
0.900
0.974
0.995
0.999
1.000
1.000
1.000
1.000
0.301
0.663
0.880
0.966
0.992
0.999
1.000
1.000
1.000
1.000
0.273
0.627
0.857
0.957
0.989
0.998
1.000
1.000
1.000
1.000
0.247
0.592
0.834
0.946
0.986
0.996
0.999
1.000
1.000
1.000
0.223
0.558
0.809
0.934
0.981
0.994
0.999
1.000
1.000
1.000
0.202
0.525
0.783
0.921
0.976
0.992
0.999
1.000
1.000
1.000
0.183
0.493
0.757
0.907
0.970
0.992
0.998
1.000
1.000
1.000
0.165
0.463
0.731
0.891
0.964
0.990
0.997
0.999
1.000
1.000
0.150
0.434
0.704
0.875
0.956
0.987
0.997
0.999
1.000
1.000
0.135
0.406
0.677
0.857
0.947
0.983
0.996
0.999
1.000
1.000
2.1
2.2
2.3
2.4
2.5
2.6
2.7
2.8
2.9
3.0
0
1
2
3
4
5
6
7
8
9
10
11
12
0.123
0.380
0.650
0.839
0.938
0.980
0.994
0.999
1.000
1.000
1.000
1.000
1.000
0.111
0.355
0.623
0.819
0.928
0.971
0.993
0.998
1.000
1.000
1.000
1.000
1.000
0.100
0.331
0.596
0.799
0.916
0.970
0.991
0.997
0.999
1.000
1.000
1.000
1.000
0.091
0.308
0.570
0.779
0.904
0.964
0.988
0.997
0.999
1.000
1.000
1.000
1.000
0.082
0.287
0.544
0.758
0.891
0.958
0.986
0.996
0.999
1.000
1.000
1.000
1.000
0.074
0.267
0.518
0.736
0.877
0.951
0.983
0.995
0.999
1.000
1.000
1.000
1.000
0.067
0.249
0.494
0.714
0.863
0.943
0.979
0.993
0.998
1.000
1.000
1.000
1.000
0.061
0.231
0.470
0.692
0.848
0.935
0.956
0.992
0.998
0.999
1.000
1.000
1.000
0.055
0.215
0.446
0.670
0.812
0.926
0.971
0.990
0.997
0.999
1.000
1.000
1.000
0.050
0.199
0.423
0.647
0.815
0.916
0.967
0.988
0.996
0.999
1.000
1.000
1.000
(continued )
H1317_CH49-AP.qxd
520
10/15/07
4:14 PM
Page 520
3.1
3.2
3.3
3.4
3.5
3.6
3.7
3.8
3.9
4.0
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
0.045
0.185
0.401
0.625
0.798
0.906
0.961
0.986
0.995
0.999
1.000
1.000
1.000
1.000
1.000
0.041
0.171
0.380
0.603
0.781
0.895
0.955
0.983
0.994
0.998
1.000
1.000
1.000
1.000
1.000
0.037
0.159
0.359
0.580
0.763
0.883
0.949
0.980
0.993
0.998
0.999
1.000
1.000
1.000
1.000
0.033
0.147
0.340
0.558
0.744
0.871
0.942
0.977
0.992
0.997
0.999
1.000
1.000
1.000
1.000
0.030
0.136
0.321
0.537
0.725
0.858
0.935
0.973
0.990
0.997
0.999
1.000
1.000
1.000
1.000
0.027
0.126
0.303
0.515
0.706
0.844
0.927
0.969
0.988
0.996
0.999
1.000
1.000
1.000
1.000
0.022
0.116
0.285
0.494
0.687
0.830
0.918
0.965
0.986
0.995
0.998
1.000
1.000
1.000
1.000
0.022
0.107
0.269
0.474
0.668
0.816
0.909
0.960
0.984
0.994
0.998
0.999
1.000
1.000
1.000
0.020
0.099
0.253
0.453
0.648
0.801
0.900
0.955
0.982
0.993
0.998
0.999
1.000
1.000
1.000
0.018
0.092
0.238
0.434
0.629
0.785
0.889
0.949
0.979
0.992
0.997
0.999
1.000
1.000
1.000
4.1
4.2
4.3
4.4
4.5
4.6
4.7
4.8
4.9
5.0
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
0.017
0.085
0.224
0.414
0.609
0.769
0.879
0.943
0.976
0.991
0.997
0.999
1.000
1.000
1.000
1.000
1.000
0.015
0.078
0.210
0.395
0.590
0.753
0.868
0.936
0.972
0.989
0.996
0.999
1.000
1.000
1.000
1.000
1.000
0.014
0.072
0.197
0.377
0.570
0.737
0.856
0.929
0.968
0.987
0.995
0.998
1.000
1.000
1.000
1.000
1.000
0.012
0.066
0.185
0.360
0.551
0.720
0.844
0.921
0.964
0.985
0.994
0.998
0.999
1.000
1.000
1.000
1.000
0.011
0.061
0.174
0.342
0.532
0.703
0.831
0.913
0.960
0.983
0.993
0.998
0.999
1.000
1.000
1.000
1.000
0.010
0.056
0.163
0.326
0.513
0.686
0.818
0.905
0.955
0.981
0.992
0.997
0.999
1.000
1.000
1.000
1.000
0.009
0.052
0.152
0.310
0.495
0.668
0.805
0.896
0.950
0.978
0.991
0.997
0.999
1.000
1.000
1.000
1.000
0.008
0.048
0.143
0.294
0.476
0.651
0.791
0.887
0.944
0.975
0.990
0.996
0.999
1.000
1.000
1.000
1.000
0.007
0.044
0.133
0.279
0.458
0.634
0.777
0.877
0.938
0.972
0.988
0.995
0.998
0.999
1.000
1.000
1.000
0.007
0.040
0.125
0.265
0.441
0.616
0.762
0.867
0.932
0.968
0.986
0.995
0.998
0.999
1.000
1.000
1.000
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 521
Appendix: Tables
521
10
15
20
30
50
100
200
500
1
2
3
4
5
39.9
8.53
5.54
4.54
4.06
49.5
9.00
5.46
4.32
3.78
53.6
9.16
5.39
4.19
3.62
55.8
9.24
5.34
4.11
3.52
57.2
9.29
5.31
4.05
3.45
58.2
9.33
5.28
4.01
3.40
58.9
9.35
5.27
3.98
3.37
59.4
9.37
5.25
3.95
3.34
59.9
9.38
5.24
3.94
3.32
60.2
9.39
5.23
3.92
3.30
61.2
9.42
5.20
3.87
3.24
61.7
9.44
5.18
3.84
3.21
62.3
9.46
5.17
3.82
3.17
62.7
9.47
5.15
3.80
3.15
63.0
9.48
5.14
3.78
3.13
63.2
9.49
5.14
3.77
3.12
63.3
9.49
5.14
3.76
3.11
63.3
9.49
5.13
3.76
3.10
6
7
8
9
10
3.78
3.59
3.46
3.36
3.28
3.46
3.26
3.11
3.01
2.92
3.29
3.07
2.92
2.81
2.73
3.18
2.96
2.81
2.69
2.61
3.11
2.88
2.73
2.61
2.52
3.05
2.83
2.67
2.55
2.46
3.01
2.78
2.62
2.51
2.41
2.98
2.75
2.59
2.47
2.38
2.96
2.72
2.56
2.44
2.35
2.94
2.70
2.54
2.42
2.32
2.87
2.63
2.46
2.34
2.24
2.84
2.59
2.42
2.30
2.20
2.80
2.56
2.38
2.25
2.16
2.77
2.52
2.35
2.22
2.12
2.75
2.50
2.32
2.19
2.09
2.73
2.48
2.31
2.17
2.07
2.73
2.48
2.30
2.17
2.06
2.72
2.47
2.29
2.16
2.06
11
12
13
14
15
3.23
3.18
3.14
3.10
3.07
2.86
2.81
2.76
2.73
2.70
2.66
2.61
2.56
2.52
2.49
2.54
2.48
2.43
2.39
2.36
2.45
2.39
2.35
2.31
2.27
2.39
2.33
2.28
2.24
2.21
2.34
2.28
2.23
2.19
2.16
2.30
2.24
2.20
2.15
2.12
2.27
2.21
2.16
2.12
2.09
2.25
2.19
2.14
2.10
2.06
2.17
2.10
2.05
2.01
1.97
2.12
2.06
2.01
1.96
1.92
2.08
2.01
1.96
1.91
1.87
2.04
1.97
1.92
1.87
1.83
2.00
1.94
1.88
1.83
1.79
1.99
1.92
1.86
1.82
1.77
1.98
1.91
1.85
1.80
1.76
1.97
1.90
1.85
1.80
1.76
16
17
18
19
20
3.05
3.03
3.01
2.99
2.97
2.67
2.64
2.62
2.61
2.59
2.46
2.44
2.42
2.40
2.38
2.33
2.31
2.29
2.27
2.25
2.24
2.22
2.20
2.18
2.16
2.18
2.15
2.13
2.11
2.09
2.13
2.10
2.08
2.06
2.04
2.09
2.06
2.04
2.02
2.00
2.06
2.03
2.00
1.98
1.96
2.03
2.00
1.98
1.96
1.94
1.94
1.91
1.89
1.86
1.84
1.89
1.86
1.84
1.81
1.79
1.84
1.81
1.78
1.76
1.74
1.79
1.76
1.74
1.71
1.69
1.76
1.73
1.70
1.67
1.65
1.74
1.71
1.68
1.65
1.63
1.73
1.69
1.67
1.64
1.62
1.72
1.69
1.66
1.63
1.61
22
24
26
28
30
2.95
2.93
2.91
2.89
2.88
2.56
2.54
2.52
2.50
2.49
2.35
2.33
2.31
2.29
2.28
2.22
2.19
2.17
2.16
2.14
2.13
2.10
2.08
2.06
2.05
2.06
2.04
2.01
2.00
1.98
2.01
1.98
1.96
1.94
1.93
1.97
1.94
1.92
1.90
1.88
1.93
1.91
1.88
1.87
1.85
1.90
1.88
1.86
1.84
1.82
1.81
1.78
1.76
1.74
1.72
1.76
1.73
1.71
1.69
1.67
1.70
1.67
1.65
1.63
1.61
1.65
1.62
1.59
1.57
1.55
1.61
1.58
1.55
1.53
1.51
1.59
1.56
1.53
1.50
1.48
1.58
1.54
1.51
1.49
1.47
1.57
1.53
1.50
1.48
1.46
40
50
60
80
100
2.84
2.81
2.79
2.77
2.76
2.44
2.41
2.39
2.37
2.36
2.23
2.20
2.18
2.15
2.14
2.09
2.06
2.04
2.02
2.00
2.00
1.97
1.95
1.92
1.91
1.93
1.90
1.87
1.85
1.83
1.87
1.84
1.82
1.79
1.78
1.83
1.80
1.77
1.75
1.73
1.79
1.76
1.74
1.71
1.70
1.76
1.73
1.71
1.68
1.66
1.66
1.63
1.60
1.57
1.56
1.61
1.57
1.54
1.51
1.49
1.54
1.50
1.48
1.44
1.42
1.48
1.44
1.41
1.38
1.35
1.43
1.39
1.36
1.32
1.29
1.41
1.36
1.33
1.28
1.26
1.39
1.34
1.31
1.26
1.23
1.38
1.33
1.29
1.24
1.21
(continued )
H1317_CH49-AP.qxd
522
10/15/07
4:14 PM
Page 522
10
11
12
13
14
15
16
17
18
1
2
3
4
5
161
18.5
10.1
7.71
6.61
200
19.0
9.55
6.94
5.79
216
19.2
9.28
6.59
5.41
225
19.2
9.12
6.39
5.19
230
19.3
9.01
6.26
5.05
234
19.3
8.94
6.16
4.95
237
19.4
8.89
6.09
4.88
239
19.4
8.85
6.04
4.82
241
19.4
8.81
6.00
4.77
242
19.4
8.79
5.96
4.74
243
19.4
8.76
5.94
4.70
24.4
19.4
8.74
5.91
4.08
245
19.4
8.73
5.82
4.66
245
19.4
8.71
5.87
4.64
246
19.4
8.70
5.86
4.62
246
19.4
8.69
5.84
4.60
247
19.4
8.68
5.83
4.59
247
19.4
8.67
5.82
4.58
6
7
8
9
10
5.99
5.59
5.32
5.12
4.96
5.14
4.74
4.46
4.26
4.10
4.76
4.35
4.07
3.86
3.71
4.53
4.12
3.84
3.63
3.48
4.39
3.97
3.69
3.48
3.33
4.28
3.87
3.58
3.37
3.22
4.21
3.79
3.50
3.29
3.14
4.15
3.73
3.44
3.23
3.07
4.10
3.68
3.39
3.18
3.02
4.06
3.64
3.35
3.14
2.98
4.03
3.60
3.31
3.10
2.94
4.00
3.57
3.28
3.07
2.91
3.98
3.55
3.26
3.05
2.89
3.96
3.53
3.24
3.03
2.86
3.94
3.51
3.22
3.01
2.85
3.92
3.49
3.20
2.99
2.83
3.91
3.48
3.19
2.97
2.81
3.90
3.47
3.17
2.96
2.80
11
12
13
14
15
4.84
4.75
4.67
4.60
4.54
2.98
3.89
3.81
3.74
3.68
3.50
3.49
3.41
3.34
3.29
3.36
3.26
3.18
3.11
3.06
3.20
3.11
3.03
2.96
2.90
3.01
3.00
2.92
2.85
2.79
2.95
2.91
2.83
2.76
2.71
2.90
2.85
2.77
2.70
2.64
2.85
2.80
2.71
2.65
2.59
2.82
2.75
2.67
2.60
2.54
2.82
2.72
2.63
2.57
2.51
2.79
2.69
2.60
2.53
2.48
2.76
2.66
2.58
2.51
2.45
2.74
2.64
2.55
2.48
2.42
2.72
2.62
2.53
2.46
2.40
2.70
2.60
2.51
2.44
2.38
2.69
2.58
2.50
2.43
2.37
2.67
2.57
2.48
2.41
2.35
16
17
18
19
20
4.49
4.45
4.41
4.38
4.35
3.63
3.59
3.55
3.52
3.49
3.24
3.20
3.16
3.13
3.10
3.01
2.96
2.93
2.90
2.87
2.85
2.81
2.77
2.74
2.71
2.74
2.70
2.66
2.63
2.60
2.66
2.61
2.58
2.54
2.51
2.59
2.55
2.51
2.48
2.45
2.54
2.49
2.46
2.42
2.39
2.49
2.45
2.41
2.38
2.35
2.46
2.41
2.37
2.34
2.31
2.42
2.38
2.34
2.31
2.28
2.40
2.36
2.31
2.28
2.25
2.37
2.33
2.29
2.26
2.22
2.35
2.31
2.27
2.23
2.20
2.33
2.29
2.25
2.21
2.18
2.32
2.27
2.23
2.20
2.17
2.30
2.26
2.22
2.18
2.15
21
22
23
24
25
4.32
4.30
4.28
4.26
4.24
3.47
3.44
3.42
3.40
3.39
3.07
3.05
3.03
3.01
2.99
2.82
2.84
2.80
2.78
2.76
2.68
2.66
2.64
2.62
2.60
2.57
2.55
2.53
2.51
2.49
2.49
2.46
2.44
2.42
2.40
2.42
2.40
2.37
2.36
2.34
2.37
2.34
2.32
2.30
2.28
2.32
2.30
2.27
2.25
2.24
2.28
2.26
2.23
2.21
2.20
2.25
2.23
2.20
2.18
2.16
2.22
2.20
2.18
2.15
2.14
2.20
2.17
2.15
2.13
2.11
2.18
2.15
2.13
2.11
2.09
2.16
2.13
2.11
2.09
2.07
2.14
2.11
2.09
2.07
2.05
2.12
2.10
2.07
2.05
2.04
26
27
28
29
30
4.23
4.21
4.20
4.18
4.17
3.37
3.35
3.34
3.33
3.32
2.98
2.96
2.95
2.93
2.92
2.74
2.73
2.71
2.70
2.69
2.59
2.57
2.56
2.55
2.53
2.47
2.46
2.45
2.43
2.42
2.39
2.37
2.36
2.35
2.33
2.32
2.31
2.29
2.28
2.27
2.27
2.25
2.24
2.22
2.21
2.22
2.20
2.19
2.18
2.16
2.18
2.17
2.15
2.14
2.13
2.15
2.13
2.12
2.10
2.09
2.12
2.10
2.09
2.08
2.06
2.09
2.08
2.06
2.05
2.04
2.07
2.06
2.04
2.03
2.01
2.05
2.04
2.02
2.01
1.99
2.03
2.02
2.00
1.99
1.98
2.02
2.00
1.99
1.97
1.96
32
34
36
38
40
4.15
4.13
4.11
4.10
4.08
3.29
3.28
3.26
3.24
3.23
2.90
2.88
2.87
2.85
2.84
2.67
2.65
2.63
2.62
2.61
2.51
2.49
2.48
2.46
2.45
2.40
2.38
2.36
2.35
2.34
2.31
2.29
2.28
2.26
2.25
2.24
2.23
2.21
2.19
2.18
2.19
2.17
2.15
2.14
2.12
2.14
2.12
2.11
2.09
2.08
2.10
2.08
2.07
2.05
2.04
2.07
2.05
2.03
2.02
2.00
2.04
2.02
2.00
1.99
1.97
2.01
1.99
1.98
1.96
1.95
1.99
1.97
1.95
1.94
1.92
1.97
1.95
1.93
1.92
1.90
1.95
1.93
1.92
1.90
1.89
1.94
1.92
1.90
1.88
1.87
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 523
Appendix: Tables
523
= 0.05
20
22
24
26
28
30
35
40
45
50
60
80
100
200
500
1
2
3
4
5
248
19.4
8.67
5.81
4.57
248
19.4
8.66
5.80
4.56
249
19.5
8.65
5.79
4.54
249
19.5
8.64
5.77
4.53
249
19.5
8.63
5.76
4.52
250
19.5
8.62
5.75
4.50
250
19.5
8.62
5.75
4.50
251
19.5
8.60
5.73
4.48
251
19.5
8.59
5.72
4.46
251
19.5
8.59
5.71
4.45
252
19.5
8.58
5.70
4.44
252
19.5
8.57
5.69
4.43
252
19.5
8.56
5.67
4.41
253
19.5
8.55
5.66
4.41
254
19.5
8.54
5.65
4.30
254
19.5
8.53
5.64
4.37
254
19.5
8.53
5.63
4.37
6
7
8
9
10
3.88
3.46
3.16
2.95
2.78
3.87
3.44
3.15
2.94
2.77
3.88
3.43
3.13
2.92
2.75
3.84
3.41
3.12
2.90
2.74
3.83
3.40
3.10
2.89
2.72
3.82
3.39
3.09
2.87
2.71
3.81
3.38
3.08
2.86
2.70
3.79
3.36
3.06
2.84
2.68
3.77
3.34
3.04
2.83
2.66
3.76
3.33
3.03
2.81
2.65
3.75
3.32
3.02
2.80
2.64
3.74
3.30
3.01
2.79
2.62
3.72
3.20
2.99
2.77
2.60
3.71
3.29
2.97
2.76
2.59
3.09
2.25
2.95
2.23
2.55
3.63
3.24
2.94
2.72
2.55
3.67
3.23
2.93
2.71
2.54
11
12
13
14
15
2.66
2.56
2.47
2.40
2.34
2.65
2.54
2.46
2.39
2.33
2.63
2.52
2.44
2.37
2.31
2.61
2.51
2.42
2.35
2.29
2.59
2.49
2.41
2.33
2.27
2.58
2.48
2.39
2.32
2.26
2.57
2.47
2.38
2.31
2.25
2.55
2.44
2.36
2.28
2.22
2.53
2.43
2.34
2.27
2.20
2.52
2.41
2.33
2.26
2.19
2.51
2.40
2.31
2.24
2.18
2.49
2.38
2.30
2.22
2.16
2.47
2.36
2.27
2.20
2.14
2.46
2.35
2.26
2.19
2.12
2.43
2.32
2.23
2.16
2.10
2.42
2.31
2.22
2.14
2.08
2.40
2.30
2.21
2.13
2.07
16
17
18
19
20
2.29
2.24
2.20
2.17
2.14
2.28
2.23
2.19
2.16
2.12
2.25
2.21
2.17
2.13
2.10
2.24
2.19
2.15
2.11
2.08
2.22
2.17
2.13
2.10
2.07
2.21
2.16
2.12
2.08
2.05
2.19
2.15
2.11
2.07
2.04
2.17
2.12
2.08
2.05
2.01
2.15
2.10
2.06
2.03
1.99
2.14
2.09
2.05
2.01
1.98
2.12
2.08
2.04
2.00
1.97
2.11
2.06
2.02
1.98
1.95
2.08
2.03
1.99
1.96
1.92
2.07
2.02
1.98
1.94
1.91
2.04
1.09
1.95
1.91
1.88
2.02
1.97
1.93
1.89
1.86
2.01
1.96
1.92
1.88
1.84
21
22
23
24
25
2.11
2.08
2.06
2.04
2.02
2.10
2.07
2.05
2.03
2.01
2.07
2.03
2.02
2.00
1.98
2.05
2.03
2.00
1.98
1.96
2.04
2.01
1.99
1.97
1.95
2.02
2.00
1.97
1.95
1.93
2.01
1.98
1.96
1.94
1.92
1.98
1.96
1.93
1.91
1.89
1.96
1.94
1.91
1.89
1.87
1.95
1.92
1.90
1.88
1.85
1.94
1.91
1.88
1.86
1.84
1.92
1.89
1.85
1.84
1.82
1.89
1.86
1.84
1.82
1.80
1.88
1.85
1.82
1.80
1.78
1.84
1.82
1.79
1.77
1.75
1.82
1.80
1.77
1.75
1.73
1.81
1.78
1.76
1.73
1.71
26
27
28
29
30
2.00
1.99
1.97
1.96
1.95
1.99
1.97
1.96
1.94
1.93
1.97
1.95
1.93
1.92
1.91
1.95
1.93
1.91
1.90
1.89
1.93
1.91
1.90
1.88
1.87
1.91
1.90
1.88
1.87
1.85
1.90
1.88
1.87
1.85
1.84
1.87
1.86
1.84
1.83
1.81
1.85
1.84
1.82
1.81
1.79
1.84
1.82
1.80
1.79
1.77
1.82
1.81
1.79
1.77
1.76
1.80
1.79
1.77
1.75
1.74
1.78
1.76
1.74
1.73
1.71
1.76
1.74
1.73
1.71
1.70
1.73
1.71
1.69
1.67
1.65
1.71
1.69
1.67
1.65
1.64
1.69
1.67
1.65
1.64
1.62
32
34
36
38
40
1.92
1.90
1.88
1.87
1.85
1.91
1.89
1.87
1.85
1.84
1.88
1.86
1.85
1.83
1.81
1.86
1.84
1.82
1.81
1.79
1.85
1.82
1.81
1.79
1.77
1.83
1.80
1.79
1.77
1.76
1.82
1.80
1.78
1.76
1.74
1.79
1.77
1.75
1.73
1.72
1.77
1.75
1.73
1.71
1.69
1.75
1.73
1.71
1.69
1.67
1.74
1.71
1.69
1.68
1.66
1.71
1.69
1.67
1.65
1.64
1.69
1.66
1.64
1.62
1.61
1.67
1.65
1.62
1.61
1.50
1.63
1.61
1.59
1.57
1.55
1.61
1.59
1.56
1.54
1.53
1.59
1.57
1.55
1.53
1.51
(continued )
H1317_CH49-AP.qxd
524
10/15/07
4:14 PM
Page 524
10
11
12
13
14
15
16
17
18
42
44
46
48
50
4.07
4.06
4.05
4.04
4.03
3.22
3.21
3.20
3.19
3.18
2.83
2.82
2.81
2.80
2.79
2.59
2.58
2.57
2.57
2.56
2.44
2.43
2.42
2.41
2.40
2.32
2.31
2.30
2.29
2.29
2.24
2.23
2.22
2.21
2.20
2.16
2.16
2.15
2.14
2.13
2.11
2.10
2.09
2.08
2.07
2.06
2.05
2.04
2.03
2.03
2.03
2.01
2.00
1.99
1.99
1.99
1.98
1.97
1.96
1.95
1.96
1.95
1.94
1.93
1.92
1.93
1.92
1.91
1.90
1.89
1.91
1.90
1.89
1.88
1.87
1.89
1.88
1.87
1.86
1.85
1.87
1.86
1.85
1.84
1.83
1.86
1.84
1.83
1.82
1.81
55
60
65
70
80
4.02
4.00
3.99
3.98
3.96
3.16
3.15
3.14
3.13
3.11
2.77
2.76
2.75
2.74
2.73
2.54
2.53
2.51
2.50
2.49
2.38
2.37
2.36
2.35
2.33
2.27
2.25
2.24
2.23
2.21
2.18
2.17
2.15
2.14
2.13
2.11
2.10
2.08
2.07
2.06
2.06
2.04
2.03
2.02
2.00
2.01
1.99
1.98
1.97
1.95
1.97
1.95
1.94
1.93
1.91
1.93
1.92
1.90
1.89
1.88
1.90
1.89
1.87
1.86
1.84
1.88
1.86
1.85
1.84
1.82
1.85
1.84
1.82
1.81
1.79
1.83
1.82
1.80
1.79
1.77
1.81
1.80
1.78
1.77
1.75
1.79
1.78
1.76
1.75
1.73
90
100
125
150
200
3.95
3.94
3.92
3.90
3.89
3.10
3.09
3.07
3.08
3.04
2.71
2.70
2.68
2.66
2.65
2.47
2.46
2.44
2.43
2.42
2.32
2.31
2.29
2.27
2.26
2.20
2.19
2.17
2.16
2.14
2.11
2.10
2.08
2.07
2.06
2.04
2.03
2.01
2.00
1.98
1.99
1.97
1.96
1.94
1.93
1.94
1.93
1.91
1.89
1.88
1.90
1.89
1.87
1.85
1.84
1.86
1.85
1.83
1.82
1.80
1.83
1.82
1.80
1.79
1.77
1.80
1.79
1.77
1.76
1.74
1.78
1.77
1.76
1.73
1.72
1.76
1.75
1.72
1.71
1.69
1.74
1.73
1.70
1.69
1.67
1.72
1.71
1.69
1.67
1.65
2.40
2.39
2.38
2.37
2.24
2.23
2.22
2.21
2.13
2.12
2.11
2.10
2.04
2.03
2.02
2.01
1.97
1.96
1.95
1.94
1.91
1.90
1.89
1.88
1.86
1.85
1.84
1.83
1.82
1.81
1.80
1.79
1.78
1.77
1.76
1.75
1.75
1.74
1.73
1.72
1.72
1.71
1.70
1.69
1.70
1.69
1.68
1.67
1.68
1.66
1.65
1.64
1.66
1.64
1.63
1.62
1.64
1.62
1.61
1.60
20
22
24
26
28
30
35
40
45
50
60
80
100
200
500
42
44
46
48
50
1.84
1.83
1.82
1.81
1.80
1.83
1.81
1.80
1.79
1.78
1.80
1.79
1.78
1.77
1.76
1.78
1.77
1.76
1.75
1.74
1.76
1.75
1.74
1.73
1.72
1.74
1.73
1.72
1.71
1.70
1.73
1.72
1.71
1.70
1.69
1.70
1.69
1.68
1.67
1.66
1.68
1.67
1.65
1.64
1.63
1.66
1.65
1.64
1.62
1.61
1.65
1.63
1.62
1.61
1.60
1.62
1.61
1.60
1.59
1.58
1.59
1.58
1.57
1.56
1.54
1.57
1.56
1.55
1.54
1.52
1.53
1.52
1.51
1.49
1.48
1.51
1.49
1.48
1.47
1.46
1.49
1.48
1.46
1.45
1.44
55
60
65
70
80
1.78
1.76
1.75
1.74
1.72
1.76
1.75
1.73
1.72
1.70
1.74
1.72
1.71
1.70
1.68
1.72
1.70
1.69
1.67
1.65
1.70
1.68
1.67
1.65
1.63
1.68
1.66
1.65
1.64
1.62
1.67
1.65
1.63
1.62
1.60
1.64
1.62
1.60
1.59
1.57
1.61
1.59
1.58
1.57
1.54
1.59
1.57
1.56
1.55
1.52
1.58
1.56
1.54
1.53
1.51
1.55
1.53
1.52
1.50
1.48
1.52
1.50
1.49
1.47
1.45
1.50
1.48
1.48
1.45
1.43
1.46
1.44
1.42
1.40
1.38
1.43
1.41
1.39
1.37
1.35
1.41
1.39
1.37
1.35
1.32
90
100
125
150
200
1.70
1.69
1.67
1.66
1.64
1.69
1.68
1.65
1.64
1.62
1.66
1.65
1.63
1.61
1.60
1.64
1.63
1.60
1.59
1.57
1.62
1.61
1.58
1.57
1.55
1.60
1.59
1.57
1.55
1.53
1.59
1.57
1.55
1.53
1.52
1.55
1.54
1.52
1.50
1.48
1.53
1.52
1.49
1.48
1.46
1.51
1.49
1.47
1.45
1.43
1.49
1.48
1.45
1.44
1.41
1.46
1.45
1.42
1.41
1.39
1.43
1.41
1.39
1.37
1.35
1.41
1.39
1.36
1.34
1.32
1.36
1.34
1.31
1.29
1.26
1.32
1.31
1.27
1.23
1.22
1.30
1.28
1.25
1.22
1.19
1.58
1.56
1.55
1.54
1.55
1.54
1.53
1.52
1.53
1.52
1.51
1.50
1.51
1.50
1.49
1.48
1.50
1.48
1.47
1.46
1.46
1.45
1.44
1.42
1.43
1.42
1.41
1.39
1.41
1.40
1.38
1.37
1.39
1.38
1.36
1.35
1.38
1.34
1.33
1.32
1.32
1.30
1.29
1.27
1.30
1.28
1.26
1.24
1.23
1.21
1.19
1.17
1.19
1.16
1.13
1.11
1.15
1.11
1.08
1.00
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 525
Appendix: Tables
525
= 0.025
10
11
12
13
14
15
16
17
18
1
2
3
4
5
648
38.5
17.4
12.2
10.0
800
39.0
16.0
10.6
8.43
864
39.2
15.4
9.98
7.76
900
29.2
15.1
9.60
7.39
922
39.3
14.9
9.36
7.15
937
39.3
14.7
9.20
6.98
948
39.4
14.6
9.07
6.85
957
39.4
14.5
8.98
6.76
963
39.4
14.5
8.90
8.68
969
39.4
14.4
8.84
6.62
975
39.4
14.4
8.79
6.57
977
39.4
14.3
8.75
6.52
980
39.4
14.3
8.79
6.49
983
39.4
14.3
8.69
6.46
985
39.4
14.3
8.66
6.43
987
39.4
14.2
8.64
6.41
989
39.4
14.2
8.62
6.39
990
39.4
14.2
8.60
6.37
6
7
8
9
10
8.81
8.07
7.57
7.21
6.94
7.28
6.54
6.06
5.71
5.46
6.00
5.89
5.42
5.08
4.83
6.23
5.52
5.05
4.72
4.47
5.90
5.29
4.62
4.48
4.24
5.82
5.12
4.65
4.32
4.07
5.70
4.99
4.53
4.20
3.95
5.60
4.90
4.43
4.10
3.85
5.52
4.82
4.66
4.03
3.78
5.48
4.76
4.30
3.96
3.72
5.41
4.71
4.24
3.91
3.86
5.37
4.67
4.20
3.87
3.67
5.33
4.63
4.16
3.83
3.59
5.30
4.60
4.13
3.80
3.55
5.27
4.57
4.10
3.77
3.52
5.25
4.54
4.06
3.74
3.50
5.23
4.52
4.05
3.72
3.47
5.21
4.50
4.03
3.70
3.45
11
12
13
14
15
6.72
6.55
6.41
6.30
6.20
5.26
5.10
4.97
4.86
4.76
4.63
4.47
4.35
4.24
4.15
4.28
4.12
4.00
3.89
3.80
4.04
3.89
3.77
3.66
3.58
3.88
3.73
3.70
3.50
3.41
3.76
3.61
3.48
3.38
3.29
3.66
3.51
3.39
3.29
3.20
3.59
3.44
3.31
3.21
3.12
3.53
3.37
3.24
3.15
3.06
3.47
3.32
3.20
3.09
3.01
3.43
3.28
3.14
3.05
2.96
3.39
3.24
3.12
3.01
2.92
3.36
3.21
3.08
2.98
2.89
3.33
3.18
3.05
2.95
2.86
3.30
3.15
3.03
2.92
2.84
3.28
3.13
3.00
2.90
2.81
3.26
3.11
2.98
2.88
2.79
16
17
18
19
20
6.12
6.04
5.98
5.92
5.87
4.69
4.62
4.56
4.51
4.48
4.08
4.01
3.95
3.90
3.86
3.73
3.66
3.61
3.56
3.51
3.50
3.44
3.38
3.33
3.29
3.34
3.28
3.22
3.17
3.13
3.22
3.16
3.10
3.05
3.01
3.12
3.06
3.01
2.96
2.91
3.05
2.98
2.93
2.88
2.84
2.99
2.92
2.87
2.82
2.77
2.93
2.87
2.81
2.76
2.72
2.89
2.82
2.77
2.72
2.68
2.85
2.79
2.73
2.68
2.64
2.82
2.75
2.70
2.65
2.60
2.79
2.72
2.67
2.62
2.57
2.76
2.70
2.64
2.59
2.55
2.74
2.67
2.62
2.57
2.52
2.72
2.65
2.60
2.55
2.50
21
22
23
24
25
5.83
5.79
5.75
5.72
5.69
4.42
4.38
4.35
4.32
4.29
3.82
3.78
3.75
3.72
3.69
3.48
3.44
3.41
3.38
3.35
3.25
3.22
3.18
3.15
3.13
3.09
3.05
3.02
2.99
2.97
2.97
2.93
2.90
2.87
2.85
2.87
2.84
2.81
2.78
2.75
2.80
2.76
2.73
2.70
2.68
2.73
2.70
2.67
2.64
2.61
2.68
2.65
2.62
2.59
2.56
2.64
2.60
2.57
2.54
2.51
2.60
2.56
2.53
2.50
2.48
2.56
2.53
2.50
2.47
2.44
2.53
2.50
2.47
2.44
2.41
2.51
2.47
2.46
2.41
2.38
2.48
2.45
2.42
2.39
2.36
2.46
2.43
2.39
2.36
2.34
26
27
28
29
30
5.66
5.63
5.61
5.59
5.57
4.27
4.24
4.22
4.20
4.18
3.67
3.65
3.63
3.61
3.59
3.33
3.31
3.29
3.27
3.25
3.10
3.08
3.06
3.04
3.03
2.94
2.92
2.90
2.88
2.87
2.82
2.80
2.78
2.76
2.75
2.73
2.71
2.69
2.67
2.65
2.65
2.63
2.61
2.59
2.57
2.59
2.57
2.55
2.53
2.51
2.54
2.51
2.49
2.48
2.46
2.49
2.47
2.45
2.43
2.41
2.45
2.43
2.41
2.39
2.37
2.42
2.39
2.37
2.36
2.34
2.39
2.36
2.34
2.32
2.31
2.36
2.34
2.32
2.30
2.28
2.34
2.31
2.29
2.27
2.26
2.31
2.29
2.27
2.25
2.23
32
34
36
38
40
5.53
5.50
5.47
5.45
5.42
4.15
4.12
4.09
4.07
4.05
3.56
3.53
3.51
3.48
3.46
3.22
3.19
3.17
3.15
3.13
3.00
2.97
2.94
2.92
2.90
2.84
2.81
2.79
2.76
2.74
2.72
2.69
2.56
2.64
2.62
2.62
2.59
2.57
2.55
2.53
2.54
2.52
2.49
2.47
2.45
2.48
2.45
2.43
2.41
2.39
2.43
2.40
2.37
2.35
2.33
2.38
2.35
2.33
2.31
2.29
2.34
2.31
2.29
2.27
2.25
2.31
2.28
2.25
2.23
2.21
2.28
2.25
2.22
2.20
2.18
2.25
2.22
2.20
2.17
2.15
2.22
2.19
2.17
2.15
2.13
2.20
2.17
2.15
2.13
2.11
(continued )
H1317_CH49-AP.qxd
526
10/15/07
4:14 PM
Page 526
20
22
24
26
28
30
35
40
45
50
60
80
100
200
500
1
2
3
4
5
992
23.4
14.2
8.53
6.35
923
30.4
14.2
8.56
6.33
995
29.5
14.1
8.53
6.30
997
39.5
14.1
8.51
6.28
999
39.5
14.1
8.49
6.26
1000
39.5
14.1
8.48
6.24
1001
39.5
14.1
8.48
6.23
1004
39.5
14.1
8.44
6.20
1005
39.5
10.0
8.41
6.18
1007
39.5
14.0
8.39
6.16
1008
39.5
14.0
8.38
6.14
1010
39.5
14.0
8.36
6.12
1012
39.5
14.0
8.33
6.10
1013
39.5
14.0
8.32
6.06
1016
39.5
13.9
8.29
6.05
1017
39.5
13.9
8.27
6.03
1018
39.5
13.9
8.26
6.02
6
7
8
9
10
5.19
4.48
4.02
3.68
3.44
5.17
4.47
4.00
3.67
3.42
5.14
4.44
3.97
3.64
3.39
5.12
4.42
3.95
3.61
3.37
5.10
4.92
3.93
3.59
3.34
5.08
4.38
3.91
3.58
3.33
5.07
4.38
3.90
3.56
3.31
5.04
4.33
3.86
3.53
3.28
5.01
4.31
3.84
3.51
3.26
4.99
4.29
3.82
3.49
3.24
4.96
4.28
3.81
3.47
3.23
4.96
4.25
3.78
3.45
3.20
4.93
4.23
3.76
3.42
3.17
4.92
4.21
3.74
3.40
3.15
4.88
4.18
3.70
3.37
3.12
4.86
4.16
3.68
3.35
3.09
4.85
4.14
3.67
3.33
3.08
11
12
13
14
15
3.24
3.09
2.96
2.86
2.77
3.23
3.07
2.95
2.84
2.76
3.20
3.04
2.92
2.81
2.73
3.17
3.02
2.89
2.79
2.70
3.15
3.00
2.87
2.77
2.68
3.13
2.98
2.85
2.75
2.66
3.12
2.96
2.84
2.73
2.64
3.09
2.93
2.80
2.70
2.61
3.06
2.91
2.78
2.67
2.55
3.04
2.89
2.76
2.65
2.56
3.03
2.87
2.74
2.64
2.55
3.00
2.85
2.72
2.61
2.52
2.97
2.82
2.69
2.58
2.49
2.96
2.80
2.67
2.56
2.47
2.92
2.76
2.63
2.53
2.44
2.90
2.74
2.61
2.50
2.41
2.88
2.72
2.60
2.49
2.40
16
17
18
19
20
2.70
2.63
2.58
2.53
2.43
2.68
2.62
2.56
2.51
2.46
2.65
2.59
2.53
2.48
2.43
2.63
2.56
2.50
2.45
2.41
2.60
2.54
2.48
2.43
2.39
2.58
2.52
2.46
2.41
2.37
2.57
2.50
2.44
2.39
2.35
2.53
2.47
2.41
2.36
2.31
2.51
2.44
2.38
2.33
2.29
2.49
2.42
2.36
2.31
2.27
2.47
2.41
2.35
2.30
2.25
2.45
2.38
2.32
2.27
2.22
2.42
2.35
2.29
2.24
2.19
2.40
2.33
2.27
2.22
2.17
2.36
2.29
2.23
2.18
2.13
2.33
2.28
2.20
2.15
2.10
2.32
2.25
2.19
2.13
2.09
21
22
23
24
25
2.44
2.41
2.37
2.35
2.32
2.42
2.39
2.36
2.33
2.30
2.39
2.36
2.33
2.30
2.27
2.37
2.33
2.30
2.27
2.24
2.34
2.31
2.28
2.25
2.22
2.33
2.29
2.26
2.23
2.20
2.31
2.27
2.24
2.21
2.18
2.27
2.24
2.20
2.17
2.13
2.25
2.21
2.18
2.15
2.12
2.23
2.19
2.15
2.12
2.10
2.21
2.17
2.14
2.11
2.08
2.18
2.14
2.11
2.08
2.05
2.15
2.11
2.08
2.05
2.02
2.13
2.09
2.06
2.02
2.00
2.09
2.05
2.01
1.98
1.95
2.06
2.02
1.99
1.95
1.92
2.04
2.00
1.97
1.94
1.91
26
27
28
29
30
2.29
2.27
2.25
2.23
2.21
2.28
2.25
2.23
2.21
2.20
2.24
2.22
2.20
2.18
2.16
2.22
2.19
2.17
2.15
2.14
2.19
2.17
2.15
2.13
2.11
2.17
2.15
2.13
2.11
2.08
2.16
2.13
2.11
2.09
2.07
2.12
2.10
2.08
2.06
2.04
2.09
2.07
2.05
2.03
2.01
2.07
2.05
2.03
2.01
1.99
2.05
2.03
2.01
1.99
1.97
2.03
2.00
1.98
1.96
1.94
1.99
1.97
1.94
1.92
1.90
1.97
1.94
1.92
1.90
1.88
1.92
1.90
1.88
1.86
1.84
1.90
1.87
1.85
1.83
1.81
1.88
1.85
1.83
1.81
1.79
32
34
36
38
40
2.18
2.15
2.12
2.11
2.08
2.16
2.13
2.11
2.09
2.07
2.13
2.10
2.08
2.05
2.03
2.10
2.07
2.05
2.03
2.01
2.08
2.05
2.03
2.00
1.98
2.06
2.03
2.00
1.98
1.96
2.04
2.01
1.99
1.96
1.94
2.00
1.97
1.95
1.93
1.90
1.98
1.95
1.92
1.90
1.88
1.95
1.92
1.90
1.87
1.85
1.93
1.90
1.88
1.85
1.83
1.91
1.88
1.85
1.82
1.80
1.87
1.84
1.81
1.79
1.76
1.85
1.82
1.79
1.76
1.74
1.80
1.77
1.74
1.71
1.69
1.77
1.74
1.71
1.68
1.66
1.75
1.72
1.69
1.66
1.64
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 527
Appendix: Tables
527
= 0.025
10
11
12
13
14
15
16
17
18
42
44
46
48
50
5.40
5.39
5.37
5.35
3.34
4.03
4.02
4.00..
3.99
3.98
3.45
3.43
3.42
3.40
3.39
3.11
3.09
3.08
3.07
3.06
2.89
2.87
2.88
2.84
2.83
2.73
2.71
2.70
2.69
2.67
2.61
2.59
2.58
2.57
2.55
2.51
2.50
2.48
2.47
2.46
2.44
2.42
2.41
2.39
2.38
2.37
2.36
2.34
2.33
2.32
2.32
2.30
2.29
2.27
2.26
2.27
2.26
2.24
2.23
2.22
2.23
2.21
2.20
2.19
2.18
2.20
2.18
2.17
2.15
2.14
2.16
2.15
2.13
2.12
2.11
2.14
2.12
2.11
2.09
2.06
2.11
2.10
2.08
2.07
2.06
2.09
2.07
2.06
2.05
2.03
55
60
65
70
80
5.31
5.29
5.27
5.25
5.22
3.95
3.93
3.91
3.89
3.86
3.36
3.34
3.32
3.31
3.28
3.03
3.01
2.99
2.98
2.95
2.81
2.79
2.77
2.75
2.73
2.65
2.63
2.61
2.60
2.57
2.53
2.51
2.49
2.48
2.45
2.43
2.41
2.39
2.38
2.36
2.36
2.33
2.32
2.30
2.28
2.29
2.27
2.25
2.24
2.21
2.24
2.22
2.20
2.18
2.16
2.19
2.17
2.15
2.14
2.11
2.15
2.13
2.11
2.10
2.07
2.11
2.09
2.07
2.06
2.03
2.08
2.06
2.04
2.03
2.00
2.05
2.03
2.01
2.00
1.97
2.03
2.01
1.99
1.97
1.95
2.01
1.98
1.97
1.95
1.93
90
100
125
150
200
5.20
5.18
5.15
5.13
5.10
3.84
3.83
3.80
3.78
3.76
3.27
3.25
3.22
3.20
3.18
2.93
2.92
2.89
2.87
2.85
2.71
2.70
2.67
2.65
2.63
2.55
2.54
2.51
2.49
2.47
2.43
2.42
2.39
2.37
2.35
2.34
2.32
2.30
2.28
2.26
2.26
2.24
2.22
2.20
2.18
2.19
2.18
2.14
2.13
2.11
2.14
2.12
2.10
2.08
2.06
2.09
2.08
2.05
2.03
2.01
2.05
2.04
2.01
1.99
1.97
2.02
2.00
1.97
1.95
1.93
1.98
1.97
1.94
1.92
1.90
1.95
1.94
1.91
1.89
1.87
1.93
1.91
1.89
1.87
1.84
1.91
1.89
1.86
1.84
1.82
2.83
2.81
2.80
2.79
2.61
2.59
2.58
2.57
2.45
2.43
2.42
2.41
2.33
2.31
2.30
2.29
2.23
2.22
2.20
2.19
2.16
2.14
2.13
2.11
2.09
2.07
2.06
2.05
2.04
2.02
2.01
1.99
1.99
1.97
1.96
1.94
1.95
1.93
1.92
1.90
1.91
1.89
1.88
1.87
1.88
1.86
1.85
1.83
1.85
1.83
1.82
1.80
1.82
1.80
1.79
1.78
1.80
1.78
1.77
1.75
20
22
24
25
28
30
35
40
45
50
60
80
100
200
500
42
44
46
48
50
2.07
2.05
2.04
2.03
2.01
2.05
2.03
2.02
2.01
1.99
2.02
2.00
1.99
1.97
1.96
1.99
1.97
1.96
1.94
1.93
1.96
1.95
1.93
1.92
1.91
1.94
1.93
1.91
1.90
1.88
1.92
1.91
1.89
1.88
1.87
1.89
1.87
1.85
1.84
1.83
1.86
1.84
1.82
1.81
1.80
1.83
1.82
1.80
1.79
1.77
1.81
1.80
1.78
1.77
1.75
1.78
1.77
1.75
1.73
1.72
1.74
1.73
1.71
1.69
1.68
1.72
1.70
1.69
1.67
1.66
1.67
1.65
1.63
1.62
1.59
1.64
1.62
1.60
1.58
1.57
1.62
1.60
1.58
1.56
1.55
55
60
65
70
80
1.99
1.96
1.95
1.93
1.90
1.97
1.94
1.93
1.91
1.88
1.93
1.91
1.89
1.88
1.85
1.90
1.88
1.86
1.85
1.82
1.88
1.86
1.84
1.82
1.79
1.86
1.83
1.82
1.80
1.77
1.84
1.82
1.80
1.78
1.75
1.80
1.78
1.76
1.74
1.71
1.77
1.74
1.72
1.71
1.68
1.74
1.72
1.70
1.68
1.65
1.72
1.70
1.68
1.66
1.63
1.69
1.67
1.65
1.63
1.60
1.65
1.62
1.60
1.58
1.55
1.62
1.60
1.58
1.56
1.53
1.57
1.54
1.52
1.50
1.47
1.54
1.51
1.48
1.46
1.43
1.51
1.48
1.46
1.44
1.40
90
100
125
150
200
1.88
1.87
1.84
1.82
1.80
1.86
1.85
1.82
1.80
1.78
1.83
1.81
1.79
1.77
1.74
1.80
1.78
1.75
1.74
1.71
1.77
1.76
1.73
1.71
1.68
1.75
1.74
1.71
1.69
1.66
1.73
1.71
1.68
1.67
1.64
1.69
1.67
1.64
1.62
1.60
1.66
1.64
1.61
1.58
1.56
1.63
1.61
1.58
1.56
1.53
1.61
1.59
1.56
1.54
1.51
1.58
1.56
1.52
1.50
1.47
1.53
1.51
1.48
1.45
1.42
1.50
1.48
1.45
1.42
1.39
1.44
1.42
1.38
1.35
1.32
1.40
1.38
1.34
1.31
1.27
1.37
1.35
1.30
1.27
1.23
1.72
1.70
1.69
1.67
1.69
1.67
1.65
1.64
1.66
1.64
1.63
1.61
1.64
1.62
1.60
1.59
1.62
1.60
1.58
1.57
1.57
1.55
1.54
1.52
1.54
1.51
1.50
1.48
1.51
1.49
1.47
1.45
1.48
1.46
1.44
1.43
1.45
1.42
1.41
1.39
1.39
1.37
1.35
1.33
1.36
1.34
1.32
1.30
1.28
1.25
1.23
1.21
1.23
1.19
1.16
1.13
1.18
1.14
1.09
1.00
(continued )
H1317_CH49-AP.qxd
528
10/15/07
4:14 PM
Page 528
10
11
12
13
14
15
16
17
18
405
93.5
34.1
21.2
16.8
500
99.0
30.8
18.0
1.32
540
99.3
30.5
16.7
12.1
563
99.3
28.7
16.0
11.4
576
99.3
28.2
15.5
11.0
596
99.3
27.9
15.2
10.7
598
99.4
27.7
15.0
10.5
598
99.4
27.5
14.8
10.3
602
99.4
27.3
14.7
10.2
606
99.4
27.2
14.5
10.1
608
99.4
27.1
14.4
9.06
611
99.4
27.1
14.4
9.89
613
99.4
27.0
14.3
9.82
614
99.4
26.9
14.2
9.77
616
99.4
26.9
14.2
9.72
617
99.4
25.8
14.2
9.68
618
99.4
26.8
14.1
9.64
619
99.4
26.8
14.1
9.61
6
7
8
9
10
13.7
1.22
11.3
10.6
10.0
10.9
9.55
8.65
8.02
7.56
9.78
8.45
7.89
6.99
6.55
9.15
7.85
7.01
6.42
5.99
8.75
7.46
6.63
6.06
5.64
8.47
7.19
6.37
5.80
5.39
8.28
6.99
6.18
5.61
5.20
8.10
6.94
6.03
5.47
5.06
7.98
6.72
5.91
5.35
4.94
7.87
6.62
5.81
5.26
4.85
7.79
6.54
5.73
5.13
4.77
7.72
6.47
5.67
5.11
4.71
7.66
6.41
5.61
5.05
4.65
7.60
6.36
5.56
5.00
4.60
7.56
5.31
5.52
4.96
4.56
7.52
6.27
5.48
4.92
4.52
7.48
6.24
5.44
4.89
4.49
7.45
6.21
5.41
4.86
4.46
11
12
13
14
15
9.65
9.33
9.07
8.86
8.68
7.21
6.93
6.70
6.51
6.26
6.22
5.95
5.74
5.58
5.42
5.67
5.41
5.21
5.04
4.89
5.32
5.06
4.86
4.70
4.56
5.07
4.82
4.62
4.46
4.32
4.89
4.64
4.44
4.28
4.14
4.74
4.50
4.30
4.14
4.00
4.63
4.30
4.19
4.03
3.89
4.54
4.30
4.10
3.94
3.80
4.46
4.22
4.02
3.88
3.72
4.40
4.16
3.96
3.80
3.67
4.34
4.10
3.91
3.75
3.61
4.39
4.05
3.86
3.70
3.56
4.25
4.01
3.82
3.66
3.52
4.21
3.97
3.78
3.62
3.49
4.18
3.94
3.75
3.50
3.45
4.15
3.91
3.72
3.56
3.42
16
17
18
19
20
8.53
8.60
8.20
8.18
8.10
6.22
6.11
6.01
5.93
5.85
5.29
5.18
5.09
5.01
4.94
4.77
4.67
4.58
4.50
4.43
4.44
4.34
4.25
4.17
4.10
4.20
4.10
4.01
3.94
3.87
4.03
3.93
3.84
3.77
3.70
3.89
3.79
3.71
3.68
3.56
3.78
3.68
3.60
3.52
3.46
3.69
3.59
3.51
3.43
3.37
3.62
3.52
3.43
3.36
3.29
3.55
3.46
3.37
3.30
3.23
3.50
3.40
3.32
3.24
3.18
3.45
3.35
3.27
3.19
3.13
3.41
3.31
3.22
3.15
3.09
3.37
3.27
3.19
3.12
3.05
3.34
3.24
3.16
3.08
3.02
3.31
3.21
3.13
3.05
2.99
21
22
23
24
25
8.02
7.95
7.86
7.82
7.77
5.78
5.72
5.66
5.61
5.57
4.87
4.82
4.76
4.72
4.68
4.37
4.31
4.26
4.22
4.18
4.04
3.99
3.94
3.90
3.86
3.81
3.76
3.71
3.67
3.63
3.64
3.59
3.54
3.50
3.46
3.51
3.45
3.41
3.36
3.32
3.40
3.35
3.30
3.26
3.22
3.31
3.26
3.21
3.17
3.13
3.24
3.18
3.14
3.09
3.06
3.17
3.12
3.07
3.03
2.99
3.12
3.07
3.02
2.98
2.94
3.07
3.02
2.97
2.93
2.89
3.03
2.98
2.93
2.89
2.85
2.99
2.94
2.89
2.85
2.81
2.96
2.91
2.86
2.83
2.78
2.93
2.88
2.83
2.79
2.75
26
27
28
29
30
7.72
7.66
7.64
7.60
7.56
5.53
5.49
5.45
5.42
5.39
4.64
4.60
4.57
4.54
4.51
4.14
4.11
4.07
4.04
4.03
3.82
3.78
3.75
3.73
3.70
3.59
3.56
3.53
3.50
3.47
3.42
3.39
3.36
3.33
3.30
3.29
3.26
3.23
3.20
3.17
3.18
3.15
3.12
3.09
3.07
3.09
3.06
3.03
3.00
2.98
3.02
2.99
2.96
2.93
2.91
2.96
2.93
2.90
2.87
2.84
2.90
2.87
2.84
2.81
2.79
2.86
2.82
2.79
2.77
2.74
2.82
2.78
2.75
2.73
2.70
2.78
2.75
2.72
2.69
2.66
2.74
2.71
2.68
2.66
2.63
2.72
2.68
2.65
2.63
2.60
32
34
36
38
40
7.50
7.44
7.40
7.35
7.31
5.34
5.29
5.25
5.21
5.18
4.46
4.42
4.36
4.34
4.31
3.97
3.93
3.89
3.86
3.83
3.65
3.61
3.57
3.54
3.51
3.43
3.39
3.35
3.32
3.29
3.26
3.23
3.18
3.15
3.12
3.13
3.09
3.05
3.02
2.99
3.02
2.96
2.95
2.92
2.90
2.93
2.89
2.86
2.83
2.80
2.86
2.82
2.79
2.75
2.73
2.80
2.76
2.72
2.69
2.66
2.74
2.70
2.67
2.64
2.61
2.70
2.66
2.62
2.59
2.56
2.66
2.62
2.58
2.55
2.52
2.62
2.58
2.54
2.51
2.48
2.58
2.55
2.51
2.48
2.45
2.55
2.51
2.48
2.45
2.42
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 529
Appendix: Tables
529
= 0.010
Degrees of freedom for the numerator (ml)
19
20
22
24
26
28
30
35
40
45
50
60
80
100
200
500
620
99.4
26.7
14.0
9.58
621
99.4
26.7
14.0
9.55
623
99.5
26.6
14.0
9.51
623
99.5
26.6
13.9
9.47
624
99.5
26.6
13.9
9.43
625
99.5
26.5
13.9
9.40
626
99.5
26.5
13.8
9.38
628
99.5
26.5
13.8
9.33
629
99.5
26.4
13.7
9.39
630
99.5
26.4
13.7
9.36
630
99.5
26.4
13.7
9.24
631
99.5
26.3
13.7
9.20
633
99.5
26.3
13.6
9.16
633
99.5
26.2
13.6
9.13
635
99.5
26.2
13.5
9.08
636
99.5
26.1
13.5
9.04
637
99.5
26.1
13.5
9.02
6
7
8
9
10
7.42
6.18
5.38
4.53
4.43
7.40
6.16
5.28
4.81
4.41
7.35
6.11
5.32
4.77
4.36
7.31
6.07
5.28
4.73
4.33
7.28
6.04
5.25
4.70
4.30
7.25
6.02
5.22
4.67
4.27
7.23
5.99
5.20
4.65
4.25
7.18
5.94
5.15
4.60
4.20
7.14
5.91
5.12
4.57
4.17
7.11
5.88
5.09
4.54
4.14
7.09
5.88
5.07
4.52
4.12
7.06
5.82
5.03
4.48
4.08
7.01
5.78
4.99
4.44
4.04
6.99
5.75
4.96
4.42
4.01
6.93
5.70
4.91
4.36
3.96
6.90
5.67
4.88
4.33
3.93
6.88
5.65
4.86
4.31
3.91
11
12
13
14
15
4.12
3.88
3.69
3.53
3.40
4.10
3.86
3.66
3.51
3.37
4.06
3.82
3.62
3.46
3.33
4.02
3.78
3.59
3.43
3.29
3.99
3.75
3.56
3.40
3.26
3.95
3.72
3.53
3.37
3.24
3.94
3.70
3.51
3.35
3.21
3.89
3.65
3.46
3.30
3.17
3.88
3.62
3.43
3.27
3.13
3.83
3.59
3.40
3.24
3.10
3.81
3.57
3.38
3.22
3.08
3.78
3.54
3.34
3.18
3.05
3.73
3.49
3.30
3.14
3.00
3.71
3.47
3.27
3.11
2.98
3.66
3.41
3.22
3.06
2.92
3.62
3.38
3.19
3.03
2.89
3.60
3.36
3.17
3.00
2.87
16
17
18
19
20
3.28
3.18
3.10
3.03
2.93
3.26
3.16
3.08
3.00
2.94
3.22
3.12
3.03
2.96
2.90
3.18
3.08
3.00
2.92
2.88
3.15
3.05
2.97
2.89
2.83
3.12
3.03
2.94
2.87
2.80
3.10
3.00
2.92
2.84
2.78
3.05
2.96
2.87
2.80
2.73
3.02
2.92
2.84
2.76
2.69
2.99
2.89
2.81
2.73
2.67
2.97
2.87
2.78
2.71
2.64
2.93
2.83
2.75
2.67
2.61
2.89
2.79
2.70
2.63
2.56
2.88
2.76
2.68
2.60
2.54
2.81
2.71
2.62
2.55
2.48
2.78
2.68
2.59
2.51
2.44
2.75
2.68
2.57
2.49
2.42
21
22
23
24
25
2.90
2.85
2.80
2.76
2.72
2.88
2.83
2.78
2.74
2.70
2.84
2.78
2.74
2.70
2.66
2.80
2.75
2.70
2.66
2.62
2.77
2.72
2.67
2.63
2.50
2.74
2.69
2.64
2.60
2.56
2.72
2.67
2.62
2.58
2.54
2.67
2.62
2.57
2.53
2.49
2.64
2.58
2.54
2.49
2.45
2.61
2.55
2.51
2.46
2.42
2.58
2.53
2.48
2.44
2.40
2.55
2.50
2.45
2.40
2.36
2.50
2.45
2.40
2.36
2.32
2.46
2.42
2.37
2.33
2.29
2.42
2.36
2.32
2.27
2.23
2.38
2.33
2.28
2.24
2.19
2.36
2.31
2.26
2.21
2.17
26
27
28
29
30
2.69
2.66
2.63
2.60
2.57
2.66
2.63
2.60
2.57
2.55
2.62
2.59
2.56
2.53
2.51
2.58
2.55
2.52
2.49
2.47
2.55
2.52
2.49
2.46
2.44
2.53
2.49
2.46
2.44
2.41
2.50
2.47
2.44
2.41
2.39
2.45
2.42
2.39
2.36
2.34
2.42
2.38
2.35
2.33
2.30
2.39
2.35
2.32
2.30
2.27
2.36
2.33
2.30
2.27
2.24
2.33
2.29
2.26
2.23
2.21
2.28
2.25
2.22
2.19
2.16
2.25
2.22
2.19
2.16
2.13
2.19
2.16
2.13
2.10
2.07
2.16
2.12
2.09
2.06
2.03
2.13
2.10
2.06
2.03
2.01
32
34
36
38
40
2.53
2.49
2.45
2.42
2.39
2.50
2.46
2.43
2.40
2.37
2.46
2.42
2.38
2.35
2.33
2.42
2.38
2.35
2.32
2.39
2.39
2.35
2.32
2.28
2.26
2.36
2.32
2.29
2.26
2.23
2.34
2.30
2.26
2.23
2.20
2.29
2.25
2.21
2.18
2.15
2.25
2.21
2.17
2.14
2.11
2.22
2.18
2.14
2.11
2.08
2.20
2.16
2.12
2.09
2.06
2.16
2.12
2.08
2.05
2.02
2.11
2.07
2.03
2.00
1.97
2.08
2.04
2.00
1.97
1.94
2.02
1.98
1.94
1.90
1.87
1.98
1.94
1.90
1.86
1.83
1.96
1.91
1.87
1.84
1.80
(continued )
H1317_CH49-AP.qxd
530
10/15/07
4:14 PM
Page 530
10
11
12
13
14
15
16
17
18
7.28
7.25
7.22
7.19
7.17
5.15
5.12
5.10
5.08
5.06
4.29
4.26
4.24
4.22
4.20
3.80
3.78
3.76
3.74
3.72
3.49
3.47
3.44
3.43
3.41
3.27
3.24
3.22
3.20
3.19
3.10
3.08
3.06
3.04
3.02
2.97
2.95
2.93
2.91
2.89
2.86
2.84
2.82
2.80
2.79
2.78
2.75
2.73
2.72
2.70
2.70
2.68
2.66
2.64
2.63
2.64
2.62
2.60
2.58
2.56
2.59
2.56
2.54
2.53
2.51
2.54
2.52
2.50
2.48
2.46
2.50
2.47
2.45
2.44
2.42
2.46
2.44
2.42
2.40
2.38
2.43
2.40
2.38
2.37
2.35
2.40
2.37
2.35
2.33
2.32
55
60
65
70
80
7.12
7.08
7.04
7.01
6.98
5.01
4.98
4.95
4.92
4.88
4.16
4.12
4.10
4.08
4.04
3.68
3.65
3.62
3.60
3.56
3.37
3.34
3.31
3.29
3.26
3.15
3.12
3.09
3.07
3.04
2.98
2.95
2.93
2.91
2.87
2.85
2.82
2.80
2.78
2.74
2.75
2.72
2.69
2.67
2.64
2.66
2.63
2.61
2.59
2.55
2.59
2.56
2.53
2.51
2.48
2.53
2.50
2.47
2.45
2.42
2.47
2.44
2.42
2.40
2.36
2.42
2.39
2.37
2.24
2.21
2.38
2.35
2.33
2.31
2.27
2.34
2.31
2.29
2.27
2.23
2.31
2.28
2.26
2.23
2.20
2.28
2.24
2.23
2.20
2.17
90
100
125
150
200
6.93
6.90
6.84
6.81
6.76
4.85
4.83
4.78
4.75
4.71
4.01
3.96
3.94
3.92
3.88
3.54
3.51
3.47
3.45
3.41
3.23
3.21
3.17
3.14
3.11
3.01
2.99
2.95
2.92
2.89
2.84
2.82
2.79
2.76
2.73
2.72
2.69
2.66
2.63
2.60
2.61
2.59
2.55
2.53
2.50
2.52
2.50
2.47
2.44
2.41
2.45
2.43
2.39
2.37
2.34
2.39
2.37
2.33
2.31
2.27
2.33
2.31
2.28
2.25
2.22
2.29
2.26
2.23
2.20
2.17
2.24
2.22
2.19
2.16
2.13
2.21
2.19
2.15
2.12
2.09
2.17
2.15
2.11
2.09
2.06
2.14
2.12
2.08
2.06
2.02
3.38
3.36
3.34
3.32
3.08
3.05
3.04
3.02
2.86
2.84
2.82
2.80
2.70
2.68
2.66
2.64
2.57
2.55
2.53
2.51
2.47
2.44
2.43
2.41
2.36
2.36
2.34
2.32
2.31
2.28
2.27
2.25
2.24
2.22
2.20
2.18
2.19
2.17
2.15
2.13
2.14
2.12
2.10
2.08
2.10
2.07
2.06
2.04
2.06
2.04
2.02
2.00
2.03
2.00
1.98
1.97
1.99
1.97
1.95
1.00
80
100
200
500
20
22
24
26
28
30
35
40
45
50
60
2.37
2.35
2.33
2.31
2.29
2.34
2.32
2.30
2.28
2.27
2.30
2.28
2.26
2.24
2.22
2.26
2.24
2.22
2.20
2.18
2.23
2.21
2.19
2.17
2.15
2.20
2.18
2.16
2.14
2.12
2.18
2.15
2.13
2.12
2.10
2.13
2.10
2.08
2.06
2.05
2.09
2.06
2.04
2.02
2.01
2.08
2.03
2.01
1.99
1.97
2.03
2.01
1.99
1.97
1.95
1.99
1.97
1.95
1.93
1.91
1.94
1.92
1.90
1.88
1.86
1.91
1.89
1.86
1.84
1.82
1.85
1.82
1.80
1.78
1.76
1.80
1.78
1.75
1.73
1.71
1.78
1.75
1.73
1.70
1.68
55
60
65
70
80
2.25
2.22
2.20
2.18
2.14
2.23
2.20
2.17
2.15
2.12
2.18
2.15
2.13
2.11
2.07
2.15
2.12
2.09
2.07
2.03
2.11
2.08
2.06
2.03
2.00
2.08
2.05
2.03
2.01
1.97
2.06
2.03
2.00
1.98
1.94
2.01
1.98
1.95
1.93
1.89
1.97
1.94
1.91
1.89
1.85
1.92
1.90
1.88
1.85
1.81
1.91
1.88
1.85
1.83
1.79
1.87
1.84
1.81
1.78
1.75
1.81
1.78
1.75
1.73
1.69
1.78
1.75
1.72
1.70
1.66
1.71
1.68
1.65
1.62
1.58
1.67
1.63
1.60
1.57
1.53
1.64
1.60
1.57
1.54
1.49
90
100
125
150
200
2.11
2.09
2.05
2.03
2.00
2.09
2.07
2.03
2.00
1.97
2.04
2.02
1.98
1.96
1.93
2.00
1.98
1.94
1.92
1.89
1.97
1.94
1.91
1.88
1.84
1.94
1.92
1.88
1.85
1.82
1.92
1.89
1.85
1.83
1.79
1.86
1.84
1.80
1.77
1.74
1.82
1.80
1.79
1.73
1.69
1.79
1.76
1.72
1.69
1.66
1.76
1.73
1.69
1.66
1.63
1.72
1.69
1.65
1.63
1.58
1.66
1.63
1.59
1.56
1.52
1.62
1.60
1.55
1.52
1.48
1.54
1.52
1.47
1.43
1.39
1.49
1.47
1.41
1.35
1.33
1.46
1.43
1.37
1.33
1.28
1.89
1.87
1.85
1.83
1.85
1.83
1.81
1.79
1.82
1.79
1.77
1.76
1.79
1.76
1.74
1.72
1.76
1.74
1.72
1.70
1.71
1.68
1.66
1.64
1.68
1.63
1.61
1.58
1.62
1.60
1.57
1.55
1.58
1.55
1.54
1.52
1.55
1.53
1.50
1.47
1.48
1.45
1.43
1.40
1.44
1.41
1.38
1.36
1.35
1.31
1.28
1.25
1.28
1.23
1.19
1.15
1.22
1.16
1.11
1.00
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 531
Appendix: Tables
531
Cpk
desired 1.33
n
30
50
80
100
150
200
250
300
400
500
750
1000
1.61189
1.53779
1.48895
1.47040
1.44249
1.42636
1.41556
1.40769
1.39679
1.38944
1.37817
1.37153
0.7
0.8
0.9
1.0
1.1
1.2
1.3
1.4
1.5
1.6
1.7
1.8
1.9
2.0
0.86493
0.82225
0.79387
0.78303
0.76667
0.75719
0.75082
0.74618
0.73973
0.73538
0.72870
0.72475
0.98260
0.93513
0.90365
0.89165
0.87355
0.86306
0.85603
0.85090
0.84378
0.83898
0.83161
0.82726
1.10073
1.04838
1.01372
1.00052
0.98063
0.96911
0.96139
0.95576
0.94796
0.94270
0.93461
0.92985
1.21922
1.16190
1.12400
1.10958
1.08786
1.07530
1.06688
1.06074
1.05223
1.04650
1.03769
1.03250
1.33799
1.27563
1.23445
1.21880
1.19523
1.18159
1.17246
1.16581
1.15658
1.15036
1.14082
1.13520
1.45697
1.38953
1.34504
1.32813
1.30268
1.28797
1.27812
1.27094
1.26099
1.25428
1.24399
1.23793
1.57612
1.50356
1.45572
1.43755
1.41022
1.39442
1.38384
1.37613
1.36545
1.35825
1.34720
1.34069
1.69541
1.61770
1.56650
1.54705
1.51781
1.50092
1.48960
1.48136
1.46994
1.46224
1.45043
1.44348
1.81481
1.73192
1.67734
1.65662
1.62546
1.60746
1.59541
1.58663
1.57446
1.56627
1.55369
1.54629
1.93431
1.84622
1.78824
1.76624
1.73315
1.71404
1.70124
1.69192
1.67901
1.67032
1.65697
1.64911
2.05388
1.96058
1.89919
1.87590
1.84087
1.82065
1.80711
1.79725
1.78359
1.77438
1.76026
1.75195
2.17352
2.07500
2.01018
1.98559
1.94863
1.92728
1.91300
1.90259
1.88818
1.87847
1.86357
1.85480
2.29322
2.18945
2.12121
2.09532
2.05641
2.03394
2.01890
2.00795
1.99278
1.98256
1.96689
1.95766
2.41297
2.30395
2.23227
2.20507
2.16421
2.14062
2.12483
2.11333
2.09740
2.08667
2.07022
2.06053
Confidence 95%
Z = 1.645
Cpk
desired 1.33
n
30
50
80
100
150
200
250
300
400
500
750
1000
1.71339
1.60827
1.54075
1.51546
1.47777
1.45618
1.44180
1.43136
1.41695
1.40727
1.39248
1.38380
0.7
0.8
0.9
1.0
1.1
1.2
1.3
1.4
1.5
1.6
1.7
1.8
1.9
2.0
0.92311
0.86303
0.82405
0.80936
0.78737
0.77473
0.76628
0.76014
0.75165
0.74594
0.73718
0.73204
1.04740
0.98043
0.93711
0.92081
0.89645
0.88245
0.87311
0.86632
0.85694
0.85063
0.84097
0.83529
1.17228
1.09830
1.05053
1.03258
1.00578
0.99040
0.98014
0.97269
0.96239
0.95547
0.94487
0.93865
1.29763
1.21651
1.16423
1.14461
1.11533
1.09853
1.08733
1.07920
1.06797
1.06042
1.04887
1.04209
1.42333
1.33500
1.27815
1.25683
1.22503
1.20680
1.19464
1.18582
1.17364
1.16545
1.15293
1.14559
1.54930
1.45370
1.39224
1.36920
1.33485
1.31517
1.30205
1.29253
1.27939
1.27056
1.25706
1.24913
1.67550
1.57258
1.50646
1.48169
1.44477
1.42362
1.40954
1.39931
1.38520
1.37572
1.36122
1.35272
1.80187
1.69159
1.62079
1.59428
1.55477
1.53215
1.51708
1.50615
1.49105
1.48092
1.46542
1.45633
1.92839
1.81072
1.73522
1.70695
1.66484
1.64074
1.62468
1.61303
1.59695
1.58616
1.56965
1.55997
2.05503
1.92994
1.84971
1.81968
1.77497
1.74937
1.73232
1.71995
1.70288
1.69143
1.67391
1.66363
2.18178
2.04924
1.96427
1.93248
1.88513
1.85804
1.84000
1.82691
1.80884
1.79672
1.77818
1.76131
2.30861
2.16861
2.07888
2.04531
1.99534
1.96674
1.94770
1.93389
1.91483
1.90203
1.88248
1.81101
2.43551
2.28803
2.19354
2.15819
2.10558
2.07547
2.05543
2.04089
2.02083
2.00736
1.98678
1.97471
2.56247
2.40750
2.30823
2.27111
2.21585
2.18423
2.16319
2.14792
2.12685
2.11211
2.09110
2.07843
Confidence 99%
Z = 2.33
Cpk
desired 1.33
n
30
50
80
100
150
200
250
300
400
500
750
1000
1.94033
1.75813
1.64749
1.60724
1.54840
1.51530
1.49349
1.47777
1.45623
1.44187
1.42007
1.40736
0.7
0.8
0.9
1.0
1.1
1.2
1.3
1.4
1.5
1.6
1.7
1.8
1.9
2.0
1.05159
0.94888
0.88573
0.86258
0.82856
0.80932
0.79659
0.78740
0.77478
0.76634
0.75350
0.74599
1.19100
1.07606
1.00564
0.97989
0.94209
0.92075
0.90665
0.89647
0.88250
0.87317
0.85898
0.85069
1.33127
1.20390
1.12608
1.09765
1.05600
1.03251
1.01700
1.00581
0.99045
0.98021
0.96463
0.95553
1.47221
1.33225
1.24691
1.21578
1.17020
1.14451
1.12757
1.11534
1.09858
1.08740
1.07040
1.06049
1.61365
1.46098
1.36804
1.33418
1.28462
1.25671
1.23831
1.22504
1.20685
1.19471
1.17628
1.16553
1.75548
1.59002
1.48942
1.45279
1.39922
1.36906
1.34919
1.33486
1.31522
1.30213
1.28224
1.27064
1.89764
1.71931
1.61099
1.57157
1.51396
1.48154
1.46017
1.44478
1.42368
1.40961
1.38825
1.37580
2.04004
1.84879
1.73272
1.69050
1.62881
1.59411
1.57125
1.55478
1.53221
1.51717
1.49432
1.48101
2.18266
1.97843
1.85457
1.80954
1.74375
1.70676
1.68240
1.66484
1.64080
1.62477
1.60043
1.58626
2.32545
2.10821
1.97653
1.92867
1.85878
1.81948
1.79360
1.77496
1.74943
1.73241
1.70658
1.69153
2.46838
2.23810
2.09858
2.04788
1.97386
1.93226
1.90486
1.88513
1.85810
1.84009
1.81275
1.79683
2.61144
2.36809
2.22070
2.16716
2.08900
2.04508
2.01616
1.99533
1.96681
1.94780
1.91895
1.90215
2.75460
2.49815
2.34288
2.28650
2.20419
2.15795
2.12750
2.10557
2.07554
2.05553
2.02517
2.00749
2.89785
2.62828
2.46512
2.40588
2.31942
2.27084
2.23886
2.21584
2.18430
2.16329
2.13141
2.11284
H1317_CH49-AP.qxd
10/15/07
4:14 PM
Page 532
H1317_CH50_IDX.qxd
10/15/07
2:06 PM
Page 533
Index
B
basic simplex algorithm, sequential simplex
optimization, 411412
bathtub curve, 392393
bell curves, 319. See also normal
distributions
Benard median rank, 148149, 477478,
486487
bias factors, for attribute studies, 516
binomial distributions, 188190, 192
negative, 194
Box, George, 165, 197
brainstorming, 141, 142143
break-even point, 32
burn-in period, 393, 395
burnout period, 394
533
H1317_CH50_IDX.qxd
534
10/15/07
2:06 PM
Page 534
Index
C
Carter, C. W., 363
c charts, 327, 345
central limit theorem, 324325
centroid, defined, 412
centroid of the remaining hyperface, 412
chart interpretation, statistical process control,
459463
chart resolution, 6581. See also resolution,
of DOE
attribute chart: p charts, 7778
normal distribution, 7981
Poisson distribution, 7879
variables control charts: averages, 6977
variables control charts: individuals, 6569
chemical stress, 1
chi-square contingency tables, 8386
chi-square control charts, 9198
chi-square distribution table, 514
chi-square goodness-of-fit test, 479481
class intervals, determining, 217219
coefficient of correlation, testing, 387389
combined (gage) repeatability and
reproducibility error (GR&R), 264
combined repeatability and reproducibility
(R&R), confidence interval for,
269270
complex systems, 404405
confidence interval for the average, 99102,
99104
single-sided limit, 102104
confidence interval for proportions, 105113
sample size required given a level of
confidence C and acceptable error SE,
111113
single-sided where the sample size is large
relative to the population, 110111
single-sided where the sample size is small
relative to the population, 110
two-sided where the sample size is large
relative to the population, 108110
two-sided where the sample size is small
relative to the population, 105108
confidence intervals
for the average, 99104
for combined repeatability and
reproducibility, 269270
H1317_CH50_IDX.qxd
10/15/07
2:06 PM
Page 535
Index
Cpk
adjusted for Sk and Ku, using normal
probability plots, 310311
calculating, 299311
process capability indices, 369371
required sample, to achieve desired Cpk,
380381
Cpm indices, 376379
Cr indices, 374375
confidence interval for, 375376
critical decision factor, calculating, 37
critical run sum (CRS), 501
CSP-1 plans, 2429
values of i and f for, 518
CSP-2 plans, 2932
CSPs. See continuous sampling plans (CSPs)
cube plots, 150155
cumulative distribution tables, 195196
cumulative Poisson distribution (table),
519520
curvature of model, testing, 172178
D
data
attribute, 139, 327328
variables, 139
data transformation, 341342
DE control chart (Standard Euclidean
distance), 291297
defect, defined, 328
defects/unit control charts (u charts),
119126, 327
case study of, 121126
demerit/unit (Du) control charts, 127134
descriptive statistics, 135140
designed experiments. See designs of
experiments (DOE)
designs of experiments (DOE), 141185,
172
alternative method for evaluating effects
based on standard deviation, 178180
case study of, 158165, 181184
empirical predictions, 168178
fractional factorial experiments, 155158
normal probability plot, 148149
Plackett-Burman screening designs,
180181
535
E
Economic Control of Quality of
Manufactured Product (Shewhart), 91
effects, alternative method for evaluating,
178180
empirical predictions, 168178
errors
attribute inspection, 270271
type I, 224225
type II, 224225
variables measurement, 259270
visual inspection, 270271
evolutionary operation (EVOP), 197203
EVOP. See evolutionary operation (EVOP)
EWMA. See exponentially weighted moving
average (EWMA) control charts
experiments. See designs of experiments
(DOE)
exponential distribution, 395407
exponentially weighted moving average
(EWMA) control charts, 205211
exponential model, 338341
F
face, defined, 412
failure, defined, 392
H1317_CH50_IDX.qxd
536
10/15/07
2:06 PM
Page 536
Index
G
generator, 156
geometric distribution, 194
goodness-of-fit test, 86
for any specified distribution, 8990
to Poisson, 8688
for uniform distribution, 89
green zone, pre-control, 363
GR&R. See combined (gage) repeatability
and reproducibility error (GR&R)
H
Ha. See alternative hypotheses
Hazen median rank, 148149
hierarchy rule, 168
high-temperature operating life (HTOL), 1
histograms, 217222
historical data, 254, 255
collecting, 251252
Ho. See null hypotheses
Hoerl, Roger, 350
Hotellings T 2 statistic, 281. See also T 2
control chart
HTOL. See high-temperature operating life
(HTOL)
humidity stress, 1
hyperface, defined, 412
hypergeometric distributions, 187188
hypotheses
alternative, 223224
formulating, 223224
null, 223224
opposing, 223
hypothesis testing
conclusions and consequences for, 224225
formulation of hypothesis for, 223224
hypothesis about variance, 239241
selecting test statistic and rejection criteria
for, 225241
hypothesis tests
selecting rejection criteria for, 227238
selecting test statistic for, 225238
type I error, 224225
type II error, 224225
types of, 226227
I
identity, 156
income distribution, Paretos law of, 357
individual-median/range control charts
control limits for, 245248
plotting data for, 248
steps for constructing, 243250
individual/moving range control charts, 43,
251257, 449456
input factors, 141148
J
Juran, J. M., 218, 270
just-in-time operations, 437
K
Kelvin, Lord (William Thomson), 138
Kurskall-Wallis rank sum test, 316317
kurtosis (Ku), 299300, 481484
L
Laney, David B, 349
Laneys p control charts, 345355
Laneys u control charts, 345355
leaf-stem plot, 220222
H1317_CH50_IDX.qxd
10/15/07
2:06 PM
Page 537
Index
M
MAD. See maximum allowed defect
(MAD)
Manuale dEconomia Politica (Pareto), 357
manufacturers risk, 15
maximum allowed defect (MAD), 272273
mean, 137. See also averages (X bar)
mean time between failure (MTBF), 45, 6
reliability and, 392
measurement error assessment
attribute and visual inspection error,
270279
consequences of, 264265
variables measurement error, 259270
median, 244, 319. See also averages (X bar);
individual-median/range control charts
Melsheimer, C. A., 270
method of least squares, 383
MIL-STD-105E (ANSI/ASQ Z1.4) sampling
plan, 1932, 35
minimum life characteristic, calculating,
491492
mode, 319
modified control limit charts, 911
moving ranges. See also individual/moving
range control charts
average, 253
defined, 252
MTBF. See mean time between failure (MTBF)
multifactor strategy, 410411
multivariate control charts
DE control chart, 291297
T2 chart, 281291
537
N
negative binomial distribution, 194
Nelson, Lloyd S., 341
nonlinearity of model, testing, 172178
nonnormal distribution Cpk, 299311
nonparametric statistics
Kruskall-Wallis rank sum test, 316317
Wilcoxon-Mann-Whitney test, 313316
nonsymmetric specifications, Taguchi loss
function, 467469
NOPP. See normal probability plots
(NOPPs)
normal distributions, 319324. See also
testing for normal distribution
as approximation of the binomial,
192194
for p charts, 7981
table for extended, 510
table for standard, 509
testing for, 471484
normal probability plots (NOPPs), 148149,
471478
Cpk adjusted for Sk and Ku using,
310311
np charts, 327, 345
N (population), 135
null hypotheses, 223224
O
observed sample statistic, 226
OC curve. See operating characteristic (OC)
curve
operating characteristic (OC) curve, 1619,
7778
for CSP-1 plan, 2728
effect of acceptance criteria on, 3032
effect of lot size on, 3032
effect of sample size on, 3032
opposing hypotheses, 223
optimization
multifactor strategies, 410411
shotgun approach to, 410
strategies for, 410418
orthogonal settings, 169172
overdispersion, 345
H1317_CH50_IDX.qxd
538
10/15/07
2:06 PM
Page 538
Index
R
rainbow chart, pre-control, 363
range, 319
red zone, pre-control, 363
regression and correlation, 383389
confidence for the estimated expected mean
value, 384385
inferences about sample slope, 385389
linear least square fit, 383384
measuring linear relationship strength, 387
testing coefficient of correlation, 387389
rejectable quality level (RQL), 16
rejection criteria, selecting, 227238
reject limits, 9
reliability, 391395. See also reliability
modeling; unreliability
bathtub curve, 392393
burn-in period, 393395
burnout period, 394395
confidence intervals and, 398401
constant failure rate, 395
defined, 391
determining failure rates, 395398
exponential distribution, 395406
mean time between failure, 392
success run theorem, 401
useful life, 393395
reliability modeling. See also reliability
complex systems, 404405
parallel systems, 403404
series systems, 402403
shared-load parallel systems, 405406
standby parallel systems, 406
repeatability
confidence interval for, 265268
defined, 259
error determination, 261262
example of, 259261
repeatable quality level (RQL), 3539
reproducibility
confidence interval for, 268269
defined, 259
error determination, 262264
H1317_CH50_IDX.qxd
10/15/07
2:06 PM
Page 539
Index
S
sample/population relationship, 135136
sample size, 16, 327328
effect of, on OC curve, 3032
sample size code letters, 2023, 20t, 22t
sampling plans
based on AQL, 3739
based on risk, 3739
based on range R, 3941
based on RQL, 3739
based on risk, 3739
continuous, 2332
MIL-STD-105E (ANSI/ASQ Z1.4),
1932
Satterthwaite, F. E., 363
sequential simplex optimization. See also
simplexes (simplices)
basic simplex algorithm, 411412
shotgun approach, 410
simplex calculations, 412418
single factor at a time, 410411
theoretical optimum, 409
variable size, 419428
series systems, 402403
Shainin, Dorian, 363
shape, defining, 409
shared-load parallel systems, 405406
Shewhart, W. A., 91
Shewhart control charts, 43, 59, 499
short-run attribute control charts,
429430
case study, 430433
short-run average control charts, 437438
short-run control charts, case study,
440448
539
H1317_CH50_IDX.qxd
540
10/15/07
2:06 PM
Page 540
Index
descriptive, 138139
examples of, 136137
location, 138139
quality and, 138
variation, 138139
stochastic approach to optimization, 410
strength distribution, 1
stress
distribution, 1
modes for accelerating, 1
success run theorem, 401402
switching rules for ANSI Z1.4 system, 20,
21f
T
tabular Pareto analysis, 358359
Taguchi loss function, 465470
Taguchi philosophy, 213
defined, 465
t-distribution, 225
table, 513
temperature-humidity bias (THB) accelerated
testing, 56
tested value, 226
testing for normal distribution, 471484. See
also normal distributions
chi-square goodness-of-fit test, 479481
normal probability plots, 471478
skewness and kurtosis, 481484
test statistic
computing, 226
selecting, 225238
THB. See temperature-humidity bias (THB)
accelerated testing
theoretical optimum, sequential simplex
optimization, 409
thermal stress, 1
Torrey, M. N., 29
T 2 control chart, 281291
for two parameters, 284
two-factor interactions, visualizing, 150155
type I error, 224225
type II error, 224225
U
u charts. See defects/unit control charts (u
charts)
u control charts, Laneys, 345355
unreliability, 1, 2. See also reliability
useful life, 395
V
values of i and f for CSP-1 plans (table), 518
variable control charts, 6577
for averages, 6977
for individuals, 6569
variable data, 327
variables, defined, 139
variables data, 139
variable size simplexes, 419428
variables measurement error, 259270
variance, hypothesis about, 239241
variation reduction, 165168
variation statistics, 138139
for individual-median/range control charts,
245
for individual/moving range control charts,
252253
vertex, defined, 411
vibrational stress, 1
visual inspection error, 270271
visualization of two-factor interactions,
150155
W
Weibull, Waloddi, 485
Weibull analysis, 485492
plotting data using Weibull paper, 489492
probability plotting, 486489
Weibull plotting paper, 489
Wheeler, Donald J., 349
Wilcoxon-Mann-Whitney test, 313316
critical values of small rank sum for, 517
Wilcoxon sum of ranks, 493498
H1317_CH50_IDX.qxd
10/15/07
2:06 PM
Page 541
Index
Y
yellow zone, pre-control, 363
541
H1317_CH50_IDX.qxd
10/15/07
2:06 PM
Page 542