Professional Documents
Culture Documents
Day 1 Agenda
Welcome and Introductions
Course Structure
Group Expectations
Introduction to Six Sigma Applications
Red Bead Experiment
Introduction to Probability Distributions
Common Probability Distributions and Their Uses
Correlation Analysis
The National Graduate School of Quality Management v7 2
Day 2 Agenda
Team Report Outs on Day 1 Material
Central Limit Theorem
Process Capability
Multi-Vari Analysis
Sample Size Considerations
Day 3 Agenda
Team Report Outs on Day 2 Material
Confidence Intervals
Control Charts
Hypothesis Testing
ANOVA (Analysis of Variation)
Contingency Tables
Day 4 Agenda
Team Report Outs on Practicum Application
Design of Experiments
Wrap Up - Positives and Deltas
Class Guidelines
Q&A as we go
Breaks Hourly
Homework Readings
As assigned in Syllabus
Grading
Class Preparation
30%
Team Classroom Exercises 30%
Team Presentations
40%
10 Minute Daily Presentation (Day 2 and 3) on Application of previous days work
20 minute final Practicum application (Last day)
Copy on Floppy as well as hard copy
Powerpoint preferred
Rotate Presenters
Q&A from the class
INTRODUCTION TO SIX
SIGMA APPLICATIONS
Learning Objectives
Have a broad understanding of statistical concepts and
tools.
A Quality Level
Phased Project:
A Program
Sweet Fruit
Design for Manufacturability
Process Entitlement
Bulk of Fruit
Process Characterization
and Optimization
Ground Fruit
Logic and Intuition
PROCESSES WHICH
PROVIDE PRODUCT VALUE
IN THE CUSTOMERS EYES
FEATURES OR
CHARACTERISTICS THE
CUSTOMER WOULD PAY
FOR.
WASTE DUE TO
INCAPABLE
PROCESSES
EXCESS INVENTORY
REWORK
WAIT TIME
EXCESS HANDLING
EXCESS TRAVEL DISTANCES
TEST AND INSPECTION
Inventory Reduction
Product Development and Introduction
Labor Reduction
Increased Utilization of Resources
Product Sales Improvement
Capacity Improvements
Delivery Improvements
The National Graduate School of Quality Management v7 12
Y = f(x)
INSPECTION EXERCISE
The necessity of training farm hands for first class
farms in the fatherly handling of farm livestock is
foremost in the minds of farm owners. Since the
forefathers of the farm owners trained the farm hands
for first class farms in the fatherly handling of farm
livestock, the farm owners feel they should carry on
with the family tradition of training farm hands of first
class farms in the fatherly handling of farm livestock
because they believe it is the basis of good
fundamental farm management.
How many fs can you identify in 1 minute of inspection.
The National Graduate School of Quality Management v7 14
INSPECTION EXERCISE
The necessity of* training f*arm hands f*or f*irst class
f*arms in the f*atherly handling of* f*arm livestock is
f*oremost in the minds of* f*arm owners. Since the
f*oref*athers of* the f*arm owners trained the f*arm
hands f*or f*irst class f*arms in the f*atherly handling
of* f*arm livestock, the f*arm owners f*eel they should
carry on with the f*amily tradition of* training f*arm
hands of* f*irst class f*arms in the f*atherly handling
of* f*arm livestock because they believe it is the basis
of* good f*undamental f*arm management.
How many fs can you identify in 1 minute of inspection.36 total are available.
The National Graduate School of Quality Management v7 15
Six Sigma
Focus on Prevention
Focus on Firefighting
Stable/Predictable Processes
Proactive
Reactive
Efficient
Wasteful
IMPROVEMENT ROADMAP
Objective
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Phase 4:
Control
WHY STATISTICS?
THE ROLE OF STATISTICS IN SIX SIGMA..
USL
WHY STATISTICS?
THE ROLE OF STATISTICS IN SIX SIGMA..
LSL
USL
Learning Objectives
Have a broad understanding of statistical concepts and
tools.
Learning Objectives
Have an understanding of the difference between
random variation and a statistically significant event.
HIRING NEEDS
BEADS ARE OUR BUSINESS
PRODUCTION SUPERVISOR
4 PRODUCTION WORKERS
2 INSPECTORS
1 INSPECTION SUPERVISOR
1 TALLY KEEPER
The National Graduate School of Quality Management v7 25
STANDING ORDERS
Follow the process exactly.
Do not improvise or vary from the documented process.
Your performance will be based solely on your ability to produce
white beads.
No questions will be allowed after the initial training period.
Your defect quota is no more than 5 off color beads allowed per
paddle.
Insert the bead paddle at a 45 degree angle into the bead bowl.
Agitate the bead paddle gently in the bead bowl until all spaces are
filled.
Gently withdraw the bead paddle from the bowl at a 45 degree angle
and allow the free beads to run off.
Without touching the beads, show the paddle to inspector #1 and wait
until the off color beads are tallied.
Move to inspector #2 and wait until the off color beads are tallied.
Inspector #1 and #2 show their tallies to the inspection supervisor. If
they agree, the inspection supervisor announces the count and the
tally keeper will record the result. If they do not agree, the inspection
supervisor will direct the inspectors to recount the paddle.
When the count is complete, the operator will return all the beads to
the bowl and hand the paddle to the next operator.
The National Graduate School of Quality Management v7 27
INCENTIVE PROGRAM
Low bead counts will be rewarded with a bonus.
High bead counts will be punished with a reprimand.
PLANT RESTRUCTURE
Defect counts remain too high for the plant to be
profitable.
The two best performing production workers will be
retained and the two worst performing production
workers will be laid off.
Your performance will be based solely on your ability
to produce white beads.
Your defect quota is no more than 10 off color beads
allowed per paddle.
The National Graduate School of Quality Management v7 29
OBSERVATIONS.
Y = f(x)
Learning Objectives
Have an understanding of the difference between
random variation and a statistically significant event.
INTRODUCTION TO
PROBABILITY
DISTRIBUTIONS
Learning Objectives
Have a broad understanding of what probability
distributions are and why they are important.
Understand the role that probability distributions play in
determining whether an event is a random occurrence or
significantly different.
Understand the common measures used to characterize a
population central tendency and dispersion.
Understand the concept of Shift & Drift.
The National Graduate School of Quality Management v7 34
Why do we Care?
An understanding of
Probability Distributions is
necessary to:
Understand the concept and
use of statistical tools.
Understand the significance of
random variation in everyday
measures.
Understand the impact of
significance on the successful
resolution of a project.
IMPROVEMENT ROADMAP
Uses of Probability Distributions
Project Uses
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Phase 4:
Control
KEYS TO SUCCESS
Types of Measures
Measures where the metric is composed of a
classification in one of two (or more) categories is called
Attribute data. This data is usually presented as a
count or percent.
Good/Bad
Yes/No
Hit/Miss etc.
Time
Miles/Hr
600
Frequency
500
400
Cumulative Frequency
300
200
100
5000
0
80
90
100
110
120
130
"Head Count"
70
80
90
100
110
120
130
70
Cumulative Percent
C umulative Frequenc y
10000
50
0
70
80
90
100
110
"Head Count"
120
130
Cumulative Percent
100
50
0
70
80
90
100
110
120
130
600
Frequency
500
400
300
200
100
0
70
80
90
100
110
120
130
"Head Count"
Cumulative Percent
100
50
0
70
80
90
100
110
"Head Count"
120
130
SIGMA (s ) IS A MEASURE
OF SCATTER FROM THE
EXPECTED VALUE THAT
CAN BE USED TO
CALCULATE A PROBABILITY
OF OCCURRENCE
500
Frequency
If we know where
we are in the
population we can
equate that to a
probability value.
This is the purpose
of the sigma value
(normal data).
400
300
200
100
0
70
NUMBER OF HEADS
SIGMA VALUE (Z)
CUM % OF POPULATION
80
90
100
110
120
130
58
65
72
79
86
93
100
107
114
121
128
135
142
-6
-5
-4
-3
-2
-1
.003
.135
2.275
15.87
50.0
84.1
97.7
99.86
99.997
Common Occurrence
Rare Event
Since the sample does not contain all the possible values,
there is some uncertainty about the population. Hence any
statistics, such as mean and standard deviation, are just
estimates of the true population parameters.
Descriptive Statistics
Descriptive Statistics is the branch of statistics which
most people are familiar. It characterizes and summarizes
the most prominent features of a given set of data (means,
medians, standard deviations, percentiles, graphs, tables
and charts.
Descriptive Statistics describe the elements of
a population as a whole or to describe data that represent
just a sample of elements from the entire population
Inferential Statistics
The National Graduate School of Quality Management v7 48
Inferential Statistics
Inferential Statistics is the branch of statistics that deals with
drawing conclusions about a population based on information
obtained from a sample drawn from that population.
While descriptive statistics has been taught for centuries,
inferential statistics is a relatively new phenomenon having
its roots in the 20th century.
We infer something about a population when only information
from a sample is known.
Probability is the link between
Descriptive and Inferential Statistics
The National Graduate School of Quality Management v7 49
600
500
Frequency
400
300
200
100
0
70
NUMBER OF HEADS
SIGMA VALUE (Z)
CUM % OF POPULATION
80
90
100
110
120
130
58
65
72
79
86
93
100
107
114
121
128
135
142
-6
-5
-4
-3
-2
-1
.003
.135
2.275
15.87
50.0
84.1
97.7
99.86
99.997
Critical
Value
Rare
Occurrence
Common
Occurrence
Rare
Occurrence
Number of occurrences
80
90
100
110
1 20
130
xi
mean = x =
n
-5
-3
-1
-2
= -.17
12
-1
0
0
n=12
0
0
0
1
3
-6
-5
-4
-3
-2
-1
Mean
= -2
If we find the value half way (50%) through the data points, we
have another way of representing central tendency. This is
called the median value.
-3
-1
-1
50% of data
points
Median
0
0
0
0
1
3
-6
-5
-4
-3
-2
-1
Median
Value
The National Graduate School of Quality Management v7 55
We find that a single value occurs more often than any other.
Since we know that there is a higher chance of this
occurrence in the middle of the distribution, we can use this
feature as an indicator of central tendency. This is called the
mode.
-1
-1
0
Mode
Mode
0
0
0
0
1
3
-6
-5
-4
-3
-2
-1
-6 -5 -4 -3 -2 -1 0
n/2=6
Median
n=12
n/2=6
Mode = 0
Mode = 0
MEAN (
X)
X =
X
n
- 2
=
= .17
12
MEDIAN
(50 percentile data point)
Here the median value falls between two zero
values and therefore is zero. If the values were
say 2 and 3 instead, the median would be 2.5.
MODE
(Most common value in the data set)
The mode in this case is 0 with 5 occurrences
within this data.
The National Graduate School of Quality Management v7 57
Number of occurrences
80
90
100
110
120
130
What are some of the ways that we can easily indicate the dispersion
(spread) characteristic of the population?
Three measures have historically been used; the range, the standard
deviation and the variance.
The National Graduate School of Quality Management v7 60
-5
-3
-1
-1
0
0
0
Range
0
0
1
3
-6
-5
-4
-3
-2
-1
Range
Min
Max
The National Graduate School of Quality Management v7 61
X =
Xi
X
(
-2
12
= -.17
- X)
61.67
=
= 5.6
n -1
12 - 1
2
-6 -5 -4 -3 -2 -1 0
DATA SET
-5
-3
-1
-1
0
0
0
0
0
1
3
6
4
Xi - X
(X
- X)
-5-(-.17)=-4.83 (-4.83)2=23.32
-3-(-.17)=-2.83 (-2.83)2=8.01
-1-(-.17)=-.83
(-.83)2=.69
-1-(-.17)=-.83
(-.83)2=.69
0-(-.17)=.17
(.17)2=.03
0-(-.17)=.17
(.17)2=.03
0-(-.17)=.17
(.17)2=.03
0-(-.17)=.17
(.17)2=.03
0-(-.17)=.17
(.17)2=.03
1-(-.17)=1.17
(1.17)2=1.37
3-(-.17)=3.17
(3.17)2=10.05
4-(-.17)=4.17
(4.17)2=17.39
61.67
MEASURES OF DISPERSION
-6 -5 -4 -3 -2 -1 0
DATA SET X i - X
-5 -5-(-.17)=-4.83
Xi
-2
X =
=
= -.17
-3 -3-(-.17)=-2.83
n
12
-1 -1-(-.17)=-.83
-1
-1-(-.17)=-.83
0
0-(-.17)=.17
0
0 0-(-.17)=.17
0 0-(-.17)=.17
0 0-(-.17)=.17
1 0-(-.17)=.17
3 1-(-.17)=1.17
-6 -5 -4 -3 -2 -1 0 1 2 3 4 5 6
4 3-(-.17)=3.17
ORDERED DATA SET
-5 4-(-.17)=4.17
-3
-1
-1
0
0
0
0
0
1
3
-6 -5 -4 -3 -2 -1 0 1 2 3 4 5 6
4
RANGE (R)
Min=-5
R = X max - X min = 4 - ( - 6) = 10
Max=4
(X - X)
2
VARIANCE (s2)
(-4.83)2=23.32
(-2.83)2=8.01
(-.83)2=.69
(-.83)2=.69
(.17)2=.03
(.17)2=.03
(.17)2=.03
(.17)2=.03
(.17)2=.03
(1.17)2=1.37
X
(
- X)
61.67
=
= 5.6
n -1
12 - 1
2
(3.17)2=10.05
(4.17)2=17.39
61.67
s = s 2 = 5.6 = 2.37
The National Graduate School of Quality Management v7 63
$ = X =
Xi
s$ = s =
2
(X
-X
n -1
4
5
6
7
8
9
Xi
10
15
12
14
10
9
11
12
10
12
Xi - X
Xi - X
10
S Xi
X
s2
The National Graduate School of Quality Management v7 64
RANGE
Good for small samples (10 or
less).
LSL
-12
-10
-8
-6
-4
-2
10
12
SO WHAT HAPPENED?
All of our work was focused in a narrow time frame.
Over time, other long term influences come and go
which move the population and change some of its
characteristics. This is called shift and drift.
Original Study
VARIATION FAMILIES
Sources of
Variation
Within Individual
Sample
Variation is present
upon repeat
measurements within
the same sample.
Piece to
Piece
Time to Time
Variation is present
upon measurements of
different samples
collected within a short
time frame.
Variation is present
upon measurements
collected with a
significant amount of
time between samples.
PPM ST C pk
500,000
460,172
420,740
382,089
344,578
308,538
274,253
241,964
211,855
184,060
158,655
135,666
115,070
96,801
80,757
66,807
54,799
44,565
0.0
0.0
0.1
0.1
0.1
0.2
0.2
0.2
0.3
0.3
0.3
0.4
0.4
0.4
0.5
0.5
0.5
0.6
PPM LT (+1.5 s )
933,193
919,243
903,199
884,930
864,334
841,345
815,940
788,145
758,036
725,747
691,462
655,422
617,911
579,260
539,828
500,000
460,172
420,740
Learning Objectives
Have a broad understanding of what probability
distributions are and why they are important.
Understand the role that probability distributions play in
determining whether an event is a random occurrence or
significantly different.
Understand the common measures used to characterize a
population central tendency and dispersion.
Understand the concept of Shift & Drift.
The National Graduate School of Quality Management v7 73
COMMON PROBABILITY
DISTRIBUTIONS AND
THEIR USES
Learning Objectives
Have a broad understanding of how probability
distributions are used in improvement projects.
Why do we Care?
Probability distributions are
necessary to:
determine whether an event is
significant or due to random
chance.
predict the probability of specific
performance given historical
characteristics.
IMPROVEMENT ROADMAP
Uses of Probability Distributions
Common Uses
Phase 1:
Measurement
Baselining Processes
Characterization
Phase 2:
Analysis
Breakthrough
Strategy
Phase 3:
Improvement
Verifying Improvements
Optimization
Phase 4:
Control
KEYS TO SUCCESS
Focus on understanding the use of the distributions
Practice with examples wherever possible
Focus on the use and context of the tool
Data points vary, but as the data accumulates, it forms a distribution which occurs naturally.
Location
Spread
Shape
Continuous Distribution
3
2
1
0
0
-1
-2
-3
-4
4
Subgroup Average
3
2
Normal Distribution
1
0
0
-1
-2
-3
c2 Distribution
-4
4
3
2
1
0
0
Name
Alpha
c2
S
t
Statistic Meaning
Common Uses
Significance level
Sample
Size
Nu
Beta
Beta Risk
Delta
Sigma
Value
Difference between
population means
Number of Standard
Deviations a value Exists
from the Mean
Hypothesis Testing,
DOE
Confidence Intervals, Contingency
Tables, Hypothesis Testing
Variance Calculations
Hypothesis Testing, Confidence Interval
of the Mean
Nearly all Functions
Probability Distributions, Hypothesis
Testing, DOE
Sample Size Determination
Sample Size Determination
Probability Distributions, Process
Capability, Sample Size Determinations
Value
Population
Sample
Mean
Variance
s2
s2
Standard Deviation
Process Capability
Cp
Cp
Binomial Mean
t CALC
X -
= s
n
s12
Fcalc = 2
s2
ca ,df =
2
(f
- fa )
2
2
=
c
c
Significant
CALC
CRIT
fe
Note that in each case, a limit has been established to determine what is
random chance verses significant difference. This point is called the critical
value. If the calculated value exceeds this critical value, there is very low
probability (P<.05) that this is due to random chance.
The National Graduate School of Quality Management v7 83
Z TRANSFORM
-1s
+1s
+2s
-2s
+/- 1s = 68%
+/- 2s = 95%
2 tail = 32%
2 tail = 4.6%
1 tail = 16%
1 tail = 2.3%
68.26%
95.46%
+3s
-3s
+/- 3s = 99.7%
2 tail = 0.3%
1 tail = .15%
99.73%
Population
Average
Compare 2
Population
Averages
Population
Variance
F Stat (n>30)
Compare a
Population
Variance
Against Target
Value(s)
F Stat (n>30)
c2 Stat (n>5)
Z Stat (p)
Z Stat (n>30)
Compare a
Population
Average
Against a
Target Value
Z Stat (n>30)
t Stat (,n<30)
Z Stat (p)
Z Stat (p)
F Stat (n<10)
c2 Stat (s,n<10)
t Stat (n<30)
t Stat (n<30)
c2 Stat (n>5)
c2 Stat (Cp)
t Stat (n<10)
t Stat (n<10)
Z Stat (,n>30)
Compare 2
Population
Variances
Means Add
1 + 2 = new
new
snew
s1
s2
Variations Add
Means Subtract
Population means interact in a simple intuitive manner.
1 - 2 = new
new
snew
s1
s2
Variations Add
TRANSACTIONAL EXAMPLE
Orders are coming in with the following characteristics:
X = $53,000/week
s = $8,000
Shipments are going out with the following
characteristics:
X = $60,000/week
s = $5,000
TRANSACTIONAL EXAMPLE
Shipments
Orders
X = $53,000 in orders/week
s = $8,000
X = $60,000 shipped/week
s = $5,000
To solve this problem, we must create a new distribution to model the situation posed
in the problem. Since we are looking for shipments to exceed orders, the resulting
distribution is created as follows:
$7000
$0
Shipments > orders
The new distribution looks like this with a mean of $7000 and a
standard deviation of $9434. This distribution represents the
occurrences of shipments exceeding orders. To answer the original
question (shipments>orders) we look for $0 on this new distribution.
Any occurrence to the right of this point will represent shipments >
orders. So, we need to calculate the percent of the curve that exists
to the right of $0.
The National Graduate School of Quality Management v7 91
2
shipments
$7000
$0
+s
2
orders
0 - X
Shipments > orders
2
2
(
)
(
)
= 5000 + 8000 = $9434
$0 - $7000
=
=.74 s
$9434
Look up .74s in the normal table and you will find .77. Therefore, the answer to the original
question is that 77% of the time, shipments will exceed orders.
Now, as a classroom exercise, what percent of the time will
shipments exceed orders by $10,000?
MANUFACTURING EXAMPLE
2 Blocks are being assembled end to end and significant
variation has been found in the overall assembly length.
X1 = 4.00 inches
s1 = .03 inches
X2 = 3.00 inches
s2 = .04 inches
Learning Objectives
Have a broad understanding of how probability
distributions are used in improvement projects.
CORRELATION ANALYSIS
Learning Objectives
Understand how correlation can be used to demonstrate
a relationship between two factors.
Why do we Care?
Correlation Analysis is
necessary to:
show a relationship between
two variables. This also sets the
stage for potential cause and
effect.
IMPROVEMENT ROADMAP
Uses of Correlation Analysis
Common Uses
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Phase 4:
Control
KEYS TO SUCCESS
WHAT IS CORRELATION?
Output or y
variable
(dependent)
Correlation
Y= f(x)
As the input variable changes,
there is an influence or bias
on the output variable.
Input or x variable (independent)
WHAT IS CORRELATION?
A measurable relationship between two variable data characteristics.
Not necessarily Cause & Effect (Y=f(x))
TYPES OF CORRELATION
Y=f(x)
Strong
Y=f(x)
Weak
Y=f(x)
None
Positive
x
Negative
CALCULATING r
Coefficient of Linear Correlation
sxy
x
(
rCALC
sxy
=
sx s y
- x )(yi - y )
n -1
APPROXIMATING r
Coefficient of Linear Correlation
Plot the data on orthogonal axis
Draw an Oval around the data
Measure the length and width
of the Oval
Y=f(x)
W
r 1 -
L
6.7
r -1 = -.47
12.6
L
|
1
|
2
|
3
|
4
|
5
|
6
6.7
|
7
+ = positive slope
|
8
|
9
|
10
|
11
|
12
|
13
12.6
|
14
|
15
- = negative slope
r CRIT
.88
.81
.75
.71
.67
.63
.51
.44
.40
.36
.28
.22
.20
Learning Objectives
Understand how correlation can be used to demonstrate
a relationship between two factors.
Learning Objectives
Understand the concept of the Central Limit Theorem.
Understand the application of the Central Limit Theorem
to increase the accuracy of measurements.
Why do we Care?
The Central Limit Theorem is:
the key theoretical link between
the normal distribution and
sampling distributions.
IMPROVEMENT ROADMAP
Uses of the Central Limit Theorem
Common Uses
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Phase 4:
Control
KEYS TO SUCCESS
Normal
Why do we Care?
What this means is that no matter what kind
of distribution we sample, if the sample size
is big enough, the distribution for the mean
is approximately normal.
This is the key link that allows us to use
much of the inferential statistics we have
been working with so far.
This is the reason that only a few probability
distributions (Z, t and c2) have such broad
application.
If a random event happens a great many times,
the average results are likely to be predictable.
Jacob Bernoulli
The National Graduate School of Quality Management v7 113
Parent
Population
n=2
n=5
n=10
sx =
sx
n
The practical aspect of all this is that if you want to improve the precision of any
test, increase the sample size.
So, if you want to reduce measurement error (for example) to determine a better
estimate of a true value, increase the sample size. The resulting error will be
1
reduced by a factor of
. The same goes for any significance testing.
n
Increasing the sample
size will reduce the error in a similar manner.
DICE EXERCISE
Break into 3 teams
Team one will be using 2 dice
Team two will be using 4 dice
Team three will be using 6 dice
Each team will conduct 100 throws of their dice and
record the average of each throw.
Plot a histogram of the resulting data.
Each team presents the results in a 10 min report out.
Learning Objectives
Understand the concept of the Central Limit Theorem.
Understand the application of the Central Limit Theorem
to increase the accuracy of measurements.
PROCESS CAPABILITY
ANALYSIS
Learning Objectives
Understand the role that process capability analysis
plays in the successful completion of an improvement
project.
Know how to perform a process capability analysis.
Why do we Care?
Process Capability Analysis is
necessary to:
determine the area of focus
which will ensure successful
resolution of the project.
benchmark a process to enable
demonstrated levels of
improvement after successful
resolution of the project.
demonstrate improvement after
successful resolution of the
project.
IMPROVEMENT ROADMAP
Uses of Process Capability Analysis
Common Uses
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Phase 4:
Control
Baselining a process
primary metric (Y) prior to
starting a project.
Characterizing the
capability of causitive
factors (x).
Characterizing a process
primary metric after
changes have been
implemented to
demonstrate the level of
improvement.
KEYS TO SUCCESS
Spec
(Lower)
Out of Spec
Probability
Single Standard
(Specification)
In Spec
Out of Spec
Probability
Spec
In Spec
Out of Spec
Probability
High Cpk
Spec
(Upper)
Spec
(Lower)
Process Center
(|-spec|)
In Spec
Process Spread
In Spec
Out of
Spec
Out of
Spec
Out of
Spec
Spec
(Upper)
Spec
(Lower)
Spec
(Upper)
In Spec
In Spec
Out of
Spec
Spec
(Upper)
Spec
(Lower)
Out of
Spec
Spec
(Lower)
(s)
Poor Cpk
Out of
Spec
Out of
Spec
Out of
Spec
Spec
(LSL)
Spec
(USL)
C PK =
Note:
LSL = Lower Spec Limit
USL = Upper Spec Limit
Calculation Values:
Upper Spec value = $200,000 maximum
No Lower Spec
= historical average = $250,000
s = $20,000
Calculation: CPK = MIN ( - LSL,USL - ) = ($200,000 - $250,000) =-.83
3s
3 * $20,000
PPM ST C pk
500,000
460,172
420,740
382,089
344,578
308,538
274,253
241,964
211,855
184,060
158,655
135,666
115,070
96,801
80,757
66,807
54,799
44,565
35,930
28,716
22,750
17,864
13,903
10,724
8,198
6,210
4,661
3,467
2,555
1,866
1,350
968
687
483
337
233
159
108
72.4
48.1
31.7
0.0
0.0
0.1
0.1
0.1
0.2
0.2
0.2
0.3
0.3
0.3
0.4
0.4
0.4
0.5
0.5
0.5
0.6
0.6
0.6
0.7
0.7
0.7
0.8
0.8
0.8
0.9
0.9
0.9
1.0
1.0
1.0
1.1
1.1
1.1
1.2
1.2
1.2
1.3
1.3
1.3
PPM LT (+1.5 s )
933,193
919,243
903,199
884,930
864,334
841,345
815,940
788,145
758,036
725,747
691,462
655,422
617,911
579,260
539,828
500,000
460,172
420,740
382,089
344,578
308,538
274,253
241,964
211,855
184,060
158,655
135,666
115,070
96,801
80,757
66,807
54,799
44,565
35,930
28,716
22,750
17,864
13,903
10,724
8,198
6,210
C PK =
Learning Objectives
Understand the role that process capability analysis
plays in the successful completion of an improvement
project.
Know how to perform a process capability analysis.
MULTI-VARI ANALYSIS
Learning Objectives
Understand how to use multi-vari charts in completing an
improvment project.
Why do we Care?
Multi-Vari charts are a:
Simple, yet powerful way to
significantly reduce the number
of potential factors which could
be impacting your primary metric.
Quick and efficient method to
significantly reduce the time and
resources required to determine
the primary components of
variation.
IMPROVEMENT ROADMAP
Uses of Multi-Vari Charts
Common Uses
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Phase 4:
Control
KEYS TO SUCCESS
tooling
operator error
fatigue
tool wear
pressure
hardness
temperature
humidity
technique
wrong spec.
lubrication
handling damage
heat treat
Im thinking of a word
in this book? Can you
figure out what it is?
Is it a Noun?
217=2X2X2X2X2X2X2X2X2X2X2X2X2X2X2X2X2= 131,072
Most Unabridged dictionaries have 127,000 words.
Piece to
Piece
Time to Time
Variation is present
upon measurements of
different samples
collected within a short
time frame.
Variation is present
upon measurements
collected with a
significant amount of
time between samples.
Manufacturing
Time to Time
Measurement Accuracy
Machine fixturing
Material Changes
Out of Round
Setup Differences
Irregularities in Part
Tool Wear
Calibration Drift
(Order Rate)
Transactional
Operator Influence
Within Individual Sample
Piece to Piece
Time to Time
Measurement Accuracy
Customer Differences
Seasonal Variation
Order Editor
Management Changes
Sales Office
Economic Shifts
Sales Rep
Interest Rate
Sample 1
Average within a
single sample
Range within a
single sample
Step 2
Sample 1
Range between
two sample
averages
Step 3
Sample 3
Sample 2
Time 1
Time 2
Time 3
Within Piece
Piece to Piece
Characterized by large
variation in readings
taken of the same single
sample, often from
different positions within
the sample.
Characterized by large
variation in readings
taken between samples
taken within a short time
frame.
Time to Time
Characterized by large
variation in readings
taken between samples
taken in groups with a
significant amount of
time elapsed between
groups.
MULTI-VARI EXERCISE
We have a part dimension which is considered to be impossible to manufacture. A
capability study seems to confirm that the process is operating with a Cpk=0
(500,000 ppm). You and your team decide to use a Multi-Vari chart to localize the
potential sources of variation. You have gathered the following data:
Sample
1
2
3
4
5
6
7
8
9
Day/
Time
1/0900
1/0905
1/0910
2/1250
2/1255
2/1300
3/1600
3/1605
3/1610
Beginning Middle of
of Part
Part
.015
.017
.010
.012
.013
.015
.014
.015
.009
.012
.012
.014
.013
.014
.010
.013
.011
.014
End of
Part
.018
.015
.016
.018
.017
.016
.017
.015
.017
Learning Objectives
Understand how to use multi-vari charts in completing an
improvment project.
SAMPLE SIZE
CONSIDERATIONS
Learning Objectives
Understand the critical role having the right sample size
has on an analysis or study.
Why do we Care?
The correct sample size is
necessary to:
ensure any tests you design
have a high probability of
success.
properly utilize the type of data
you have chosen or are limited to
working with.
IMPROVEMENT ROADMAP
Uses of Sample Size Considerations
Common Uses
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Sample Size
considerations are used in
any situation where a
sample is being used to
infer a population
characteristic.
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Phase 4:
Control
KEYS TO SUCCESS
CONFIDENCE INTERVALS
The possibility of error exists in almost every system. This goes for point values as
well. While we report a specific value, that value only represents our best estimate
from the data at hand. The best way to think about this is to use the form:
Point
Value
Reality
Test Decision
Different (H1)
Different (H1)
Type I Error
a Risk
Producer Risk
Type II Error
risk
Consumer Risk
Correct Conclusion
Reality = Different
Test
Decision Point
Risk
a Risk
The National Graduate School of Quality Management v7 151
A person does not feel well and checks into a hospital for tests.
Reality
Test Decision
Different (H1)
Different (H1)
Type I Error
a Risk
Producer Risk
Type II Error
risk
Consumer Risk
Correct Conclusion
Error Impact
Type I Error = Treating a patient who is not sick
Type II Error = Not treating a sick patient
The National Graduate School of Quality Management v7 152
A change is made to the sales force to save costs. Did it adversely impact
the order receipt rate?
Reality
Test Decision
Different (H1)
Type I Error
a Risk
Producer Risk
Different (H1)
Type II Error
risk
Consumer Risk
Correct Conclusion
Error Impact
Type I Error = Unnecessary costs
Mean
X - t a / 2 ,n - 1
s
n
n-1
Standard Deviation
Process Capability
c12-a / 2,n-1
Percent Defective
Cp
p$ - Za / 2
c12-a / 2
n-1
X + t a / 2 ,n -1
s s
s
n
n-1
Cp Cp
ca2/ 2
ca2/ 2,n-1
p$ (1 - p$ )
p p$ + Za / 2
n
n-1
p$ (1 - p$ )
n
These individual formulas are not critical at this point, but notice that the only
opportunity for decreasing the error band (confidence interval) without decreasing the
confidence factor, is to increase the sample size.
The National Graduate School of Quality Management v7 154
Mean
Z a / 2s
n=
- X
Allowable error = - X
(also known as )
s 2
n = ca / 2 + 1
s
2
Standard Deviation
Percent Defective
Za / 2
n = p$ (1 - p$ )
Allowable error = E
Calculation Values:
Average tells you to use the mean formula
Significance: a = 5% (95% confident)
Za/2 = Z.025 = 1.96
s=10 pounds
-x = error allowed = 2 pounds
2
Za / 2s
1.96 *10
n=
=
= 97
2
- X
2
Calculation:
Answer: n=97 Samples
Calculation Values:
Percent defective tells you to use the percent defect formula
Significance: a = 5% (95% confident)
Za/2 = Z.025 = 1.96
p = 10% = .1
E = 1% = .01
Calculation:
Za / 2
1.96
n = p$ (1 - p$ )
= 3458
=.1(1-.1)
E
.01
Answer: n=3458 Samples
The National Graduate School of Quality Management v7 157
Learning Objectives
Understand the critical role having the right sample size
has on an analysis or study.
CONFIDENCE INTERVALS
Learning Objectives
Understand the concept of the confidence interval and
how it impacts an analysis or study.
Why do we Care?
Understanding the confidence
interval is key to:
understanding the limitations of
quotes in point estimate data.
being able to quickly and
efficiently screen a series of point
estimate data for significance.
IMPROVEMENT ROADMAP
Uses of Confidence Intervals
Common Uses
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Phase 4:
Control
KEYS TO SUCCESS
Point
Value
Original Study
TIME
2
Significant Change?
95% Confidence
Interval
Mean
X - t a / 2 ,n - 1
s
n
n-1
Standard Deviation
Process Capability
c12-a / 2,n-1
Percent Defective
Cp
p$ - Za / 2
c12-a / 2
n-1
X + t a / 2 ,n -1
s s
s
n
n-1
Cp Cp
ca2/ 2
ca2/ 2,n-1
p$ (1 - p$ )
p p$ + Za / 2
n
n-1
p$ (1 - p$ )
n
These individual formulas enable us to calculate the confidence interval for many of
the most common metrics.
The National Graduate School of Quality Management v7 168
Calculation:
p$ - Za / 2
.014 - 1.96
p$ (1 - p$ )
p p$ + Za / 2
n
.014 (1 - .014 )
p .014 + 1.96
10 ,000
p$ (1 - p$ )
n
.014 (1 - .014 )
10 , 000
.012 p .016
Answer: Yes, .023 is significantly outside of the expected 95% confidence interval of
.012 to .016.
The National Graduate School of Quality Management v7 169
Learning Objectives
Understand the concept of the confidence interval and
how it impacts an analysis or study.
CONTROL CHARTS
Learning Objectives
Understand how to select the correct control chart for an
application.
Why do we Care?
Control charts are useful to:
determine the occurrence of
special cause situations.
Utilize the opportunities
presented by special cause
situations to identify and correct
the occurrence of the special
causes .
IMPROVEMENT ROADMAP
Uses of Control Charts
Common Uses
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Phase 4:
Control
KEYS TO SUCCESS
Point
Value
Special cause
occurrence.
+/- 3s =
99% Confidence
Interval
Variable
What type of
data do I have?
X-s Chart
1 < n < 10
X-R Chart
Attribute
Counting defects or
defectives?
n=1
Defectives
Defects
Constant
Sample Size?
Constant
Opportunity?
IMR Chart
yes
np Chart
no
yes
no
p Chart
c Chart
u Chart
UCL = X +
UCL = D4 R
AR
LCL = X - A R
2
LCL = D3 R
n
2
3
4
5
6
7
8
9
D4
3.27
2.57
2.28
2.11
2.00
1.92
1.86
1.82
D3
0
0
0
0
0
0.08
0.14
0.18
A2
1.88
1.02
0.73
0.58
0.48
0.42
0.37
0.34
P(1 - P )
n
P(1 - P )
n
Note: P charts have an individually calculated control limit for each point plotted
U
UCLu = U + 3
n
UCLC = C + 3 C
U
LCLu = U - 3
n
LCLC = C - 3 C
Note: U charts have an individually calculated control limit for each point plotted
Zone B
Zone C
Centerline/Average
Zone C
Zone B
Zone A
Learning Objectives
Understand how to select the correct control chart for an
application.
HYPOTHESIS TESTING
Learning Objectives
Understand the role that hypothesis testing plays in an
improvement project.
Why do we Care?
Hypothesis testing is
necessary to:
determine when there is a
significant difference between
two sample populations.
determine whether there is a
significant difference between a
sample population and a target
value.
IMPROVEMENT ROADMAP
Uses of Hypothesis Testing
Common Uses
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Phase 4:
Control
Confirm sources of
variation to determine
causative factors (x).
Demonstrate a
statistically significant
difference between
baseline data and data
taken after improvements
were implemented.
KEYS TO SUCCESS
Ho = no difference
H1 = significant difference
(Null hypothesis)
(Alternate Hypothesis)
StatisticCALC< StatisticCRIT
StatisticCALC> StatisticCRIT
Statistic Value
We determine a critical value from a probability table for the statistic. This
value is compared with the calculated value we get from our data. If the
calculated value exceeds the critical value, the probability of this occurrence
happening due to random variation is less than our test a.
The National Graduate School of Quality Management v7 193
How you
say it
What it
means
Hypothesis
Symbol
Null
Ho
Fail to Reject
the Null
Hypothesis
Alternative
H1
Data supports
conclusion that
there is a
significant
difference
Population
Variance
Compare 2
Population
Averages
Compare a
Population
Average
Against a
Target Value
Compare 2
Population
Variances
Compare a
Population
Variance
Against a
Target Value
Test Used
Test Used
Test Used
Test Used
Z Stat (n>30)
Z Stat (n>30)
F Stat (n>30)
F Stat (n>30)
Z Stat (p)
Z Stat (p)
F Stat (n<10)
c2 Stat (n>5)
t Stat (n<30)
t Stat (n<30)
c2 Stat (n>5)
t Stat (n<10)
t Stat (n<10)
The National Graduate School of Quality Management v7 195
Examples
Ho: 1>2
Ho: s1<s2
Common
Occurrence
1-a Probability
Critical
Value
Critical
Value
Rare
Occurrence
a Probability
Rare
Occurrence
a/2 Probability
Examples
Common
Occurrence
1-a Probability
Rare
Occurrence
a/2 Probability
Ho: 1=2
Ho: s1=s2
Confidence Int
n1 = n2 20
n1 + n2 30
n1 + n2 > 30
2 Sample Tau
2 Sample t
(DF: n1+n2-2)
2 Sample Z
t dCALC
2 X1 - X 2
=
R1 + R 2
t=
X1 - X 2
1
1
+
n1 n2
[( n - 1) s12 + ( n - 1) s22]
n1 + n2 - 2
ZCALC =
X1 - X 2
s12
n1
s 212
n2
Statistic Summary:
n1 = n2 = 5
Significance: a/2 = .025 (2 tail)
taucrit = .613 (From the table for a = .025 & n=5)
Calculation:
t dCALC
R1=337, R2= 577
X1=2868, X2=2896
tauCALC=2(2868-2896)/(337+577)=|.06|
X1 - X 2
R1 + R2
Receipts 1 Receipts 2
3067
3200
2730
2777
2840
2623
2913
3044
2789
2834
Test:
Ho: tauCALC< tauCRIT
Ho: .06<.613 = true? (yes, therefore we will fail to reject the null hypothesis).
Conclusion:
Fail to reject the null hypothesis (ie. The data does not support the conclusion that there is a
significant difference)
The National Graduate School of Quality Management v7 199
n1 = n2 = 5
DF=n1 + n2 - 2 = 8
Significance: a/2 = .025 (2 tail)
tcrit = 2.306 (From the table for a=.025 and 8 DF)
Calculation:
s1=130, s2= 227
X1=2868, X2=2896
tCALC=(2868-2896)/.63*185=|.24|
Test:
t=
Receipts 1 Receipts 2
3067
3200
2730
2777
2840
2623
2913
3044
2789
2834
X1 - X 2
1
1
+
n1 n2
[( n - 1) s12 + ( n - 1) s22]
n1 + n2 - 2
Conclusion:
Fail to reject the null hypothesis (ie. The data does not support the conclusion that there is a significant
difference
The National Graduate School of Quality Management v7 200
n 20
1 Sample Tau
t1CALC =
X - 0
R
n 30
1 Sample t
(DF: n-1)
t CALC =
X - 0
s2
n
n > 30
1 Sample Z
ZCALC =
X - 0
s2
n
Range Test
F Test
(DF1: n1-1, DF2: n2-1)
FCALC
R MAX ,n1
=
R MIN ,n2
MAX
s
Fcalc = 2 MIN
s
Calculation:
R1=337, R2= 577
FCALC=577/337=1.7
Test:
FCALC
R MAX ,n1
=
R MIN ,n2
Receipts 1 Receipts 2
3067
3200
2730
2777
2840
2623
2913
3044
2789
2834
Conclusion:
Fail to reject the null hypothesis (ie. cant say there is a significant difference)
The National Graduate School of Quality Management v7 203
ZCALC =
p1 - p0
p0 (1 - p0 )
n
ZCALC =
p1 - p2
n1 p1 _ n2 p2 n1 p1 _ n2 p2 1 1
1 +
n1 + n2 n1 n2
n1 + n2
2.749
2.332
2.227
1.858
1.974
2.480
1.503
1.891
2.316
2.475
3.119
2.808
1.468
2.124
2.470
Learning Objectives
Understand the role that hypothesis testing plays in an
improvement project.
ANalysis Of VAriance
ANOVA
Learning Objectives
Understand the role that ANOVA plays in problem
solving tools & methodology.
Why do we Care?
Anova is a powerful method for
analyzing process variation:
Used when comparing two or
more process means.
IMPROVEMENT ROADMAP
Uses of Analysis of Variance Methodology---ANOVA
Common Uses
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Hypothesis Testing
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Design of Experiments
(DOE)
Phase 4:
Control
KEYS TO SUCCESS
L ite r s
P e r H r B y
F o r m u la t i o n
G r o u p M e a n s a r e In d i c a t e d b y L i n e s
L i te r s P e r H r
2 5
3 Groups = 3 Treatments
2 0
Ove ra ll M e a n
ti
1 5
2
r m u la t i o n
Lets say we run a study where weF ohave
three groups which we are evaluating for a
significant difference in the means. Each one of these groups is called a treatment
and represents one unique set of experimental conditions. Within each treatments, we
have seven values which are called repetitions.
The National Graduate School of Quality Management v7 212
L ite r s
P e r H r B y
F o r m u la t i o n
G r o u p M e a n s a r e In d i c a t e d b y L i n e s
L i te r s P e r H r
2 5
3 Groups = 3 Treatments
2 0
Ove ra ll M e a n
ti
1 5
2
m u la t i o n
The basic concept of ANOVA is Ftoo rcompare
the variation between the
treatments with the variation within each of the treatments. If the variation
between the treatments is statistically significant when compared with the
noise of the variation within the treatments, we can reject the null hypothesis.
The National Graduate School of Quality Management v7 213
( y
a
i =1 j =1
ij
- y
= n ( yi - y ) +
a
i =1
(y
a
i =1 j =1
SSTotal = SSTreatments
(Between Treatments)
ij
- yi
+ SSError
(Within Treatments)
a = # of treatments (i)
n = # of repetitions within each treatment (j)
The National Graduate School of Quality Management v7 214
FCalculated =
SS
Degrees of Freedom
DETERMINING SIGNIFICANCE
F Critical
Value
Ho : All t s = 0
All s equal
(Null hypothesis)
StatisticCALC< StatisticCRIT
Statistic Value
We determine a critical value from a probability table. This value is compared with
the calculated value. If the calculated value exceeds the critical value, we will reject
the null hypothesis (concluding that a significant difference exists between 2 or more
of the treatment means.
The National Graduate School of Quality Management v7 216
Treatment 1
19
18
21
16
18
20
14
Treatment 2
20
15
21
19
17
22
19
Treatment 3
23
20
25
22
18
24
22
Here we have an example of three different fuel formulations that are expected to
show a significant difference in average fuel consumption. Is there a significant
difference? Lets use ANOVA to test our hypothesis..
126
18.0
19.7
2.8
Treatment 2
20
15
21
19
17
22
19
133
19.0
19.7
0.4
Treatment 3
23
20
25
22
18
24
22
154
22.0
19.7
5.4
Treatment 4
0
19.7
0.0
60.7
2
30.3
The first step is to come up with the Mean Squares Between. To accomplish this
we:
find the average of each treatment and the overall average
find the difference between the two values for each treatment and square it
sum the squares and multiply by the number of replicates in each treatment
divide this resulting value by the degrees of freedom
The National Graduate School of Quality Management v7 218
17
22
19
0
0
0
23
20
25
22
18
24
22
0
0
0
19.0
19.0
19.0
19.0
19.0
19.0
22.0
22.0
22.0
22.0
22.0
22.0
22.0
22.0
22.0
22.0
-2.0
3.0
0.0
0.0
0.0
0.0
1.0
-2.0
3.0
0.0
-4.0
2.0
0.0
0.0
0.0
0.0
Sum of Squares Within (SSw)
Degrees of Freedom Within (DFw = (# of treatments)(samples -1)
Mean Squares Within (MSw=SSw/DFw)
4.0
9.0
0.0
0.0
0.0
0.0
1.0
4.0
9.0
0.0
16.0
4.0
0.0
0.0
0.0
0.0
102.0
18
5.7
5.35
4.56
TRUE
Learning Objectives
Understand the role that ANOVA plays in problem
solving tools & methodology.
CONTINGENCY TABLES
(CHI SQUARE)
Learning Objectives
Understand how to use a contingency table to support an
improvement project.
Why do we Care?
Contingency tables are helpful
to:
Perform statistical significance
testing on count or attribute data.
IMPROVEMENT ROADMAP
Uses of Contingency Tables
Common Uses
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Phase 4:
Control
Confirm sources of
variation to determine
causative factors (x).
Demonstrate a
statistically significant
difference between
baseline data and data
taken after improvements
were implemented.
KEYS TO SUCCESS
Conduct ad hoc training for your team prior to using the tool
Gather data, use the tool to test and then move on.
Use historical data where ever possible
Keep it simple
c2CRIT
No significant
differences
Ho
LSL
USL
c2CALC
The National Graduate School of Quality Management v7 227
We ensure that we include both good and bad situations so that the sum of
each column is the total opportunities which have been grouped into good and
bad. We then gather enough data to ensure that each cell will have a
count of at least 5.
Fill in the data.
Orders On Time
Orders Late
Vendor A
25
7
Vendor B
58
9
Vendor C
12
5
Orders On Time
Orders Late
Total
Vendor A
25
7
32
Vendor B
58
9
67
Vendor C
12
5
17
Total
95
21
116
Bar #1
5
7
Bar #2
7
9
Bar #3
2
4
Bar #1
5
7
Bar #2&3
9
13
Vendor A
25
7
32
Vendor B
58
9
67
Vendor C
12
5
17
Total Portion
95
0.82
21
0.18
116
1.00
.The row percentage times the column total gives us the expected occurrence
for each cell based on the percentages. For example, for the Orders on time
row, .82 x32=26 for the first cell and .82 x 67= 55 for the second cell.
Orders On Time
Orders Late
Total
Orders On Time
Orders Late
Actual Occurrences
Vendor A Vendor B Vendor C
25
58
12
7
9
5
32
67
17
Total Portion
95
0.82
21
0.18
116
1.00
Expected Occurrences
Vendor A Vendor B Vendor C
26
55
Orders On Time
Orders Late
Column Total
Actual Occurrences
Vendor A Vendor B Vendor C
25
58
12
7
9
5
32
67
17
Total Portion
95
0.82
21
0.18
116
1.00
Orders On Time
Orders Late
Expected Occurrences
Vendor A Vendor B Vendor C
26
55
14
6
12
3
Orders On Time
Orders Late
Calculations (Expected)
Vendor A Vendor B Vendor C
.82x32
.82x67
.82x17
.18x32
.18x67
.18x17
Now we need to calculate the c2 value for the data. This is done using the
formula (a-e)2/e (where a is the actual count and e is the expected count) for
each cell. So, the c2 value for the first column would be (25-26)2/26=.04.
Filling in the remaining c2 values we get:
Orders On Time
Orders Late
Column Total
Actual Occurrences
Vendor A Vendor B Vendor C
25
58
12
7
9
5
32
67
17
2
Total Portion
95
0.82
21
0.18
116
1.00
Orders On Time
Orders Late
Expected Occurrences
Vendor A Vendor B Vendor C
26
55
14
6
12
3
Orders On Time
Orders Late
Calculated c Values
Vendor A Vendor B Vendor C
0.04
0.18
0.27
0.25
0.81
1.20
(25-26) /26
2
(7-6) /6
(58-55) /55
2
(9-12) /12
(12-14) /14
2
(5-3) /3
Good Parts
Scrap
Good Parts
Scrap
Actual Occurrences
Machine 1 Machine 2 Machine 3
100
350
900
15
18
23
Actual Occurrences
Shift 1
Shift 2
Shift 3
500
400
450
20
19
17
The National Graduate School of Quality Management v7 233
Learning Objectives
Understand how to use a contingency table to support an
improvement project.
DESIGN OF
EXPERIMENTS (DOE)
FUNDAMENTALS
Learning Objectives
Have a broad understanding of the role that design of
experiments (DOE) plays in the successful completion of
an improvement project.
Understand how to construct a design of experiments.
Understand how to analyze a design of experiments.
Understand how to interpret the results of a design of
experiments.
Why do we Care?
Design of Experiments is
particularly useful to:
evaluate interactions between 2
or more KPIVs and their impact
on one or more KPOVs.
optimize values for KPIVs to
determine the optimum output
from a process.
IMPROVEMENT ROADMAP
Uses of Design of Experiments
Phase 1:
Measurement
Characterization
Phase 2:
Analysis
Breakthrough
Strategy
Phase 3:
Improvement
Optimization
Phase 4:
Control
KEYS TO SUCCESS
Even the most clever analysis will not rescue a poorly planned experiment
Dont be afraid to ask for help until you are comfortable with the tool
Ensure a detailed test plan is written and followed
Y=f(x)
f(x)
Key Process
Input Variables
Process
Noise Variables
A combination of inputs
which generate
corresponding outputs.
Variables
Input, Controllable (KPIV)
Input, Uncontrollable (Noise)
Key Process
Output Variables
4
2 IV
2 levels for
each KPIV
4 factors
evaluated
Resolution
IV
H. Steinhaus
Mumble, Mumble,
blackbelt, Mumble,
statistics stuff...
Half Fraction
2 level designs
3 level designs
screening designs
A
+
+
+
+
-
B
+
+
+
+
-
23=8
22=4
21=2
C
+
+
+
+
-
20=1
Yates algorithm is a quick and easy way (honest, trust me) to ensure that we get a
balanced design whenever we are building a full factorial DOE. Notice that the
number of treatments (unique test mixes of KPIVs) is equal to 23 or 8.
Notice that in the A factor column, we have 4 + in a row and then 4 - in a row. This
is equal to a group of 22 or 4. Also notice that the grouping in the next column is 21 or
2 + values and 2 - values repeated until all 8 treatments are accounted for.
Repeat this pattern for the remaining factors.
The National Graduate School of Quality Management v7 244
Factorial
Treatment
AB
AC
BC
ABC
+
+
+
+
-
+
+
+
+
-
+
+
+
+
-
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
-
2
3
4
5
6
7
8
You can see from the example above we have added additional columns for each of
the ways that we can mix the 3 factors which are under study. These are our
interactions. The sign that goes into the various treatment boxes for these
interactions is the algebraic product of the main effects treatments. For example,
treatment 7 for interaction AB is (- x - = +), so we put a plus in the box. So, in these
calculations, the following apply:
minus (-) times minus (-) = plus (+)
minus (-) times plus (+) = minus (-)
Determine the 2 levels for each factor. Ensure that the levels are
as widely spread apart as the process and circumstance allow.
Draw up the design using the Yates algorithm.
Treatment
AB
AC
BC
ABC
+
+
+
+
-
+
+
+
+
-
+
+
+
+
-
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
-
2
3
4
5
6
7
8
AB
AC
BC
ABC
AVG
+
+
+
+
-
+
+
+
+
-
+
+
+
+
-
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
-
18
18
12
12
2
3
4
5
6
7
8
Analysis of a DOE
Calculate the average output for each treatment.
Place the average for each treatment after the sign (+ or -)
in each cell.
Treatment
AB
AC
BC
ABC
AVG
+18
+12
+6
+9
-3
-3
-4
-8
+
+
+
+
-
+
+
+
+
-
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
-
18
18
12
12
3
4
5
6
7
8
Analysis of a DOE
Add up the values in each column and put the result under
the appropriate column. This is the total estimated effect of
the factor or combination of factors.
AB
AC
BC
ABC
AVG
+18
+12
+6
+9
-3
-3
-4
-8
+18
+12
-6
-9
+3
+3
-4
-8
+18
-12
+6
-9
+3
-3
+4
-8
+18
+12
-6
-9
-3
-3
+4
+8
+18
-12
+6
-9
-3
+3
-4
+8
+18
-12
-6
+9
+3
-3
-4
+8
+18
-12
-6
+9
-3
+3
+4
-8
18
27
6.75
9
2.25
-1
-0.25
21
5.25
7
1.75
13
3.25
5
1.25
2
3
4
5
6
7
8
SUM
AVG
12
6
9
3
3
4
8
63
Analysis of a DOE
These averages represent the average difference between the
factor levels represented by the column. So, in the case of factor
A, the average difference in the result output between the + level
and the - level is 6.75.
We can now determine the factors (or combination of factors)
which have the greatest impact on the output by looking for the
magnitude of the respective averages (i.e., ignore the sign).
Treatment
AB
AC
BC
ABC
AVG
+18
+12
-6
-9
+3
+3
-4
-8
+18
-12
+6
-9
+3
-3
+4
-8
+18
+12
-6
-9
-3
-3
+4
+8
+18
-12
+6
-9
-3
+3
-4
+8
+18
-12
-6
+9
+3
-3
-4
+8
+18
-12
-6
+9
-3
+3
+4
-8
18
+18
+12
+6
+9
-3
-3
-4
-8
SUM
AVG
27
6.75
9
2.25
-1
-0.25
21
5.25
7
1.75
13
3.25
5
1.25
63
2
3
4
5
6
7
12
6
9
3
3
4
8
Analysis of a DOE
We can see the impact,
but how do we know if
these results are
significant or just
random variation?
Ranked Degree of
Impact
A
AB
BC
B
AC
ABC
C
(6.75)
(5.25)
(3.25)
(2.25)
(1.75)
(1.25)
(-0.25)
(6.75)
(5.25)
(3.25)
(2.25)
(1.75)
(1.25)
(-0.25)
Confidence Interval
= Effect +/- Error
Some of these factors
do not seem to have
much impact. We can
use them to estimate
our error.
We can be relatively
safe using the ABC and
the C factors since they
offer the greatest
chance of being
insignificant.
The National Graduate School of Quality Management v7 253
A
AB
BC
B
AC
ABC
C
(6.75)
(5.25)
(3.25)
(2.25)
(1.75)
(1.25)
(-0.25)
Confidence = ta / 2 , DF
2
2
ABC
C
+
(
)
DF
DF=# of groups used
In this case we are using 2
groups (ABC and C) so
our DF=2
Confidence
Since only 2 groups meet or exceed
our 95% confidence interval of +/3.97. We conclude that they are the
only significant treatments.
Confidence = 4 .303
(1.25
+ ( - .25 )
2
6s
GUTTER!
IMPROVEMENT PHASE
Vital Few Variables
Establish Operating Tolerances
How do I know
what works for me......
Lane conditions?
Ball type?
Wristband?
Factor A
Factor B
Factor C
Wristband (+)
Wristband (+)
Wristband (+)
Wristband (+)
No Wristband(-)
No Wristband(-)
No Wristband(-)
No Wristband(-)
dry lane
hard ball
oily lane
oily lane
hard bowling ball
wearing a wristband
wristband
soft ball
ball
nohard
wristband
soft ball
dry lane
hard bowing ball
not wearing wrist band
188
183
174
191
Average of with
wristband scores =184
wristband
soft ball
hard ball
without wristband
Higher Scores !!
158
154
141
159
Average of without
wristband scores =153
soft ball
hard ball
188
183
Your best Scores are when:
wristband
soft ball
hard ball
no wristband
174
158
191
141
Dry Lane
OR
Oily Lane
154
Hard Ball
Soft Ball
159
soft ball
With Wristband
and
use:
Dry
Hard Ball
Oily
Soft Ball
Learning Objectives
Have a broad understanding of the role that design of
experiments (DOE) plays in the successful completion of
an improvment project.
Understand how to construct a design of experiments.
Understand how to analyze a design of experiments.
Understand how to interpret the results of a design of
experiments.