Quality Midsem
Quality Midsem
Quality Engineering
Prof. Sayak Roychowdhury
What this course is about
• Design of Experiments
• Control charts deal with online quality control
• Off-line control methods used in design phase of the
product/process/service
• Neverending process of quality improvement
Example
https://www.agry.purdue.edu/ext/corn/news/timeless/images/US_Corn_Yld_Trend.png
Nobel Prize and what not
• https://theconversation.com/how-randomised-trials-
became-big-in-development-economics-128398
Engineering Example
• Improvement of Injection Molding Process
(Park, K., & Ahn, J. H. (2004). Design of experiment considering two-way interactions and its application to
injection molding processes with numerical analysis. Journal of Materials Processing Technology, 146(2), 221-
227.)
Etch Rate
Power (W) 1 2 3 4 5
160 575 542 530 539 570
180 565 593 590 579 610
200 600 651 610 637 629
220 725 700 715 685 710
1 Factor: More than 2 Levels
• Concept of ANOVA
• ANOVA Table
• Formulas
• Conclusion
Central Limit Theorem (CLT)
• Definition: If 𝑥1 , … , 𝑥𝑛 are independent random variables with mean 𝜇𝑖 and variance 𝜎𝑖2 ,
𝑦−σ𝑛𝑖=1 𝜇𝑖
and if 𝑦 = 𝑥1 + ⋯ + 𝑥𝑛 , then the distribution of approaches the 𝑁 0,1
σ𝑛 2
𝑖=1 𝜎𝑖
distribution as 𝑛 approaches infinity.
(Montgomery D.C., Introduction to Statistical Quality Control)
• If 𝑥𝑖 are independent and identically distributed (IID), and distribution of each 𝑥𝑖 does not
depart radically from normal distribution, then CLT works quite well for 𝑛 ≥ 3 𝑜𝑟 4.
(common in SQC problems)
Important Sampling Distributions Derived
from Normal Distribution
1. 𝜒 2 distribution: If 𝑥1 , . . 𝑥𝑛 are standard normally and
independently distributed then 𝑦 = 𝑥12 + 𝑥22 … + 𝑥𝑛2
follow chi-squared distribution with 𝑛 degrees of
freedom.
2. 𝑡-distribution: If 𝑥 is standard normal variable and 𝑦 is chi-
squared random variable with 𝑘 degrees of freedom, and𝑥 if
𝑥 and 𝑦 are independent then the random variable 𝑡 = 𝑦 is
𝑘
distributed as 𝑡 with 𝑘 degrees of freedom.
3. If 𝑤 and 𝑦 are two independent random chi-sq distributed
variables with 𝑢 and 𝑣 degrees of freedom, then the ratio
𝑤
𝑢
𝐹= 𝑦 follows F distribution with (𝑢, 𝑣) degrees of freedom
𝑣
ANOVA
Minitab: Stat-> ANOVA-> One way ANOVA,
Graphs: Select Boxplot, Normal Probability Plot of Residuals
Effects Model
𝒚𝒊𝒋 = 𝝁 + 𝝉𝒊 + 𝝐𝒊𝒋
𝑖 = 1, . . 𝑎 levels
𝑗 = 1, . . 𝑛 observations in each level
Null hypothesis 𝐻0 : 𝜏1 = 𝜏2 … . = 𝜏𝑎 = 0
𝐻1 : 𝜏𝑖 ≠ 0 for at least one 𝑖
ANOVA
Grand
Mean Random Error
Component 𝑁(0, 𝜎 2 )
Linear Statistical Model
𝒚𝒊𝒋 = 𝝁 + 𝝉𝒊 + 𝝐𝒊𝒋
𝑖 = 1, . . 𝑎 levels
Treatment Effect
𝑗 = 1, . . 𝑛 observations in each level
(Effect due to level i)
ANOVA Calculations
𝑎 𝑛 𝑎 𝑎 𝑛
2 2 2
𝑦𝑖𝑗 − 𝑦.
ത . = 𝑛 𝑦ത𝑖 . −𝑦ത.. + 𝑦𝑖𝑗 − 𝑦ത𝑖 .
𝑖=1 𝑗=1 𝑖=1 𝑖=1 𝑗=1
𝜈 = 𝜈1 + ⋯ + 𝜈𝑠
Cochran’s Theorem
𝑆𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 𝑆𝑆𝐸
Cochran’s theorem implies that and are
𝜎2 𝜎2
independently distributed chi-square random variables.
𝑭𝟎.𝟎𝟓,𝟐,𝟏𝟐
Key Considerations for ANOVA
• Experiments have to be performed in random order so
that the environment in which the treatments are applied
is as uniform as possible. The experimental design should
be completely randomized design.
• When “𝑎”no. treatment levels are specifically chosen by
the experimenter, the conclusions cannot be extended for
similar treatments that were not considered. This is called
Fixed Effects Model
• When "𝑎“ no. of treatment levels are chosen randomly
out of a larger population of treatments, 𝜏𝑖 (treatment
effects) are random variables and we try to estimate the
variability in 𝜏𝑖 . This is called Random Effects Model.
Key Considerations for ANOVA
• In Fixed Effects Model we test hypotheses about the
treatment means.
• The conclusions for FEM are only applied to the factor
levels considered.
• The model parameters (𝜇, 𝜏𝑖 , 𝜎 2 ) can be estimated from
FEM.
• In Random Effects Model, conclusions can be exteneded
to all treatments.
• In REM, 𝜏𝑖 s are random variables, and the variability of
the 𝜏𝑖 s are estimated.
Which Factors are Significant
no
F-test finds
Terminate
Significance?
Perform Pair-wise
“LSD” Comparisons Terminate
Or Tukey’s test
Least Significant Difference
1 1
▪ LSD = 𝑡1−𝛼 , 𝑛𝑜. 𝑜𝑓 𝑑𝑜𝑓𝑠 𝑜𝑓 𝑟𝑒𝑠𝑖𝑑𝑢𝑎𝑙𝑠 * 𝑀𝑆𝐸 ( + )
2 𝑛1 𝑛2
2
3
levels are different
𝐻0 : 𝜇1 + 𝜇2 = 𝜇3 + 𝜇4 or 𝜇1 + 𝜇2 − 𝜇3 − 𝜇4 = 0
𝐻1 : 𝜇1 + 𝜇2 ≠ 𝜇3 + 𝜇4 or 𝜇1 + 𝜇2 − 𝜇3 − 𝜇4 ≠ 0
(2)
Contrast
• Contrast is the linear combination of the parameters of
the form Γ = σ𝑎𝑖=1 𝑐𝑖 𝜇𝑖 , where σ𝑎𝑖=1 𝑐𝑖 = 0
• The example hypotheses can be expressed in the form
𝐻0 : σ𝑎𝑖=1 𝑐𝑖 𝜇𝑖 = 0
𝐻1 : σ𝑎𝑖=1 𝑐𝑖 𝜇𝑖 ≠ 0
• 𝑐𝑖 s are called contrast constants
• For (1): the contrast constants are 𝑐1 = 𝑐2 = 0,
𝑐3 = 1, 𝑐4 = −1
• For (2): the contrast constants are 𝑐1 = 𝑐2 = +1,
𝑐3 = 𝑐4 = −1
Hypothesis Testing with Contrast
• Null hypothesis 𝐻0 : σ𝑎𝑖=1 𝑐𝑖 𝜇𝑖 = 0
• Treatment averages 𝐶 = σ𝑎𝑖=1 𝑐𝑖 𝑦𝑖∙
𝜎2 𝑎
• Variance of 𝐶: 𝑉 𝐶 = σ𝑖=1 𝑐𝑖2
𝑛
𝐶
• If 𝐻0 is true then follows 𝑁(0,1). Since 𝜎 2 is
𝑉 𝐶
estimated by 𝑀𝑆𝐸 , the new statistic would be:
σ𝑎 ത 𝑖∙
𝑖=1 𝑐𝑖 𝑦
𝑡0 =
𝑀𝑆𝐸
σ𝑎 𝑐
𝑖=1 𝑖
2
𝑛
• The null hypothesis is rejected if 𝑡0 > 𝑡1−𝛼,𝑁−𝑎
2
F-test
2
σ𝑎 ത 𝑖∙
𝑖=1 𝑐𝑖 𝑦
• 𝐹0 = 𝑡02 = 𝑀𝑆𝐸 is an F- statistic
σ𝑎 𝑐
𝑖=1 𝑖
2
𝑛
• Null hypothesis will be rejected if 𝐹0 > 𝐹1−𝛼,1,𝑁−𝑎
𝑆𝑆𝑐
𝑀𝑆𝑐
• 𝐹0 = = 1
,
𝑀𝑆𝐸 𝑀𝑆𝐸
2
σ𝑎 ത 𝑖∙
𝑖=1 𝑐𝑖 𝑦
where 𝑆𝑆𝑐 = 1 𝑎 is contrast sum of squares with 1
σ𝑖=1 𝑐𝑖2
𝑛
degree of freedom
Confidence Interval for Contrast
• 100 1 − 𝛼 % confidence interval is given by
𝑀𝑆𝐸 𝑎
σ𝑎𝑖=1 𝑐𝑖 𝑦ത𝑖∙ − 𝑡𝛼,𝑁−𝑎 σ𝑖=1 𝑐𝑖2 ≤ σ𝑎𝑖=1 𝑐𝑖 𝜇𝑖 ≤
2 𝑛
𝑀𝑆𝐸 𝑎
σ𝑎𝑖=1 𝑐𝑖 𝑦ത𝑖∙ + 𝑡𝛼,𝑁−𝑎 σ𝑖=1 𝑐𝑖2
2 𝑛
Example
Factor: RF power setting at 4 levels. 5 etch rates were
collected at each level. From the data
𝑦1∙ = 551.2 ; 𝑦2∙ = 587.4 ; 𝑦3∙ = 625.4; 𝑦4∙ = 707.0
𝑦. . = 12355; 𝑀𝑆𝐸 = 333.70
Test the hypothesis that
𝐻0 : 𝜇1 + 𝜇2 = 𝜇3 + 𝜇4
Scheffe’s Method for Comparison
• Scheffe (1953) proposed a method for any and all possible
contrasts
• Suppose there are 𝑚 contrasts given by
Γ𝑢 = 𝑐1𝑢 𝜇1 + ⋯ 𝑐𝑎𝑢 𝜇𝑎 𝑤ℎ𝑒𝑟𝑒 𝑢 = 1, . . 𝑚
• The corresponding contrast in treatment average is
𝑎
𝐶𝑢 = 𝑐𝑖𝑢 𝑦ത𝑖∙ 𝑤ℎ𝑒𝑟𝑒 𝑢 = 1, . . 𝑚
𝑖=1
• The standard error is given by
𝑆𝐶𝑢 = (𝑀𝑆𝐸 σ𝑎𝑖=1 𝑐𝑖2 /𝑛𝑖 )
• The critical value to be compared with 𝐶𝑢 is
S𝛼,𝑢 = 𝑆𝐶𝑢 𝑎 − 1 𝐹1−𝛼,𝑎−1,𝑁−𝑎
• If 𝐶𝑢 > 𝑆𝛼,𝑢 then the contrast is significant
Scheffe’s Method
• 𝐶𝑢 − 𝑆𝛼,𝑢 ≤ Γ𝑢 ≤ 𝐶𝑢 + 𝑆𝛼,𝑢 are simultaneous confidence
intervals.
• In Scheffe’s Method, Type-I error is atmost 𝛼 for any
possible comparison
Randomized Complete
Block Design and Latin
Square Design
Prof. Sayak Roychowdhury
Motivation
• Blocking is a technique to understand the extent of
influence of nuisance factors
• A nuisance factor is a factor that may have an effect on
the response variable, not of primary importance to the
experimenter. E.g Variability due to raw material,
operator, shift, etc.
Unknown and
Nuisance factors
Randomize
Uncontrollable
Randomized
Design (Assign b1 Level a2, 10 runs
and b2 randomly)
Level a3, 5
runs
Randomized
Block Design
Level a1, 5
runs
Level a3, 5
runs
Example 2
• We want to study whether the effect of different tips on
measurement of hardness
• 1 factor : tip type, number of levels: 4
• You may take 4 replications for each levels (total 4 X 4 = 16
runs) and apply ANOVA to determine the effect due to levels
• Nuisance factor : test coupons, as they maybe produced
from different batches of raw materials
• An example of RCBD in this case:
Test Coupon
1 2 3 4
Tip3 Tip3 Tip2 Tip1
Tip1 Tip4 Tip1 Tip4
Tip4 Tip2 Tip3 Tip2
Tip2 Tip1 Tip4 Tip3
Data Collection for RCBD
𝑆𝑆𝑇
𝑆𝑆𝐵𝑙𝑜𝑐𝑘𝑠
𝑆𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠
𝑆𝑆𝐸
Calculations
• For ease of calculations the following simplifications
maybe used:
Results and Conclusion
• If 𝐹0 > 𝐹 1−𝛼 , 𝑎−1 , 𝑎−1 𝑏−1 then conclude that null
hypothesis is rejected and the treatments have significant
effects
𝑀𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠
where 𝐹0 =
𝑀𝑆𝐸
𝐹 1−𝛼 , 𝑎−1 , 𝑎−1 𝑏−1 is found from the F-distribution table,
usually 𝛼 = 0.05
• Effect of blocking factor may also be estimated from the ratio
𝑀𝑆𝐵𝑙𝑜𝑐𝑘𝑠
(but not usually compared with F-distribution, as
𝑀𝑆𝐸
blocking puts a restriction on randomization)
• Usually if the ratio is much greater than 1, it suggests the
blocking factor has a significant contribution in the variability
of the response, and suggests that blocking was a good idea
Explanation
• Considering the error term 𝜖𝑖𝑗 ~𝑁(0, 𝜎 2 ), the expected
values of mean squares can be derived as
𝑀𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠
• Hence if 𝐹0 = ≫ 1 it suggests significance of
𝑀𝑆𝐸
treatment effects. Similar conclusion can be made for the
blocking factor as well
Estimation of Parameters
• The estimates of the parameters are given by:
𝜇ො = 𝑦.
ത.
𝜏ෝ𝑖 = 𝑦ത𝑖∙ − 𝑦.
ത . . 𝑖 = 1, . . , 𝑎
𝛽መ𝑗 = 𝑦ത∙𝑗 − 𝑦.
ത . . 𝑗 = 1, . . , 𝑏
𝑦ො𝑖𝑗 = 𝑦ത𝑖∙ + 𝑦ത∙𝑗 − 𝑦.ത..
Latin Square Design
• There are several other types of designs that utilize the blocking principle.
• Each formulation is mixed from a batch of raw material that is only large
enough for five formulations to be tested.
• Thus, it would seem that there are two nuisance factors to be “averaged
out” in the design: batches of raw material and operators.
Latin Square Design
𝑀𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡 − 𝑀𝑆𝐸𝑟𝑟𝑜𝑟
𝜎ො𝜏2 =
𝑛
Non-parametric Method
• Kruskal-Wallis Test:
When normality assumption is not justified, experimenter
may resort to non-parametric methods , that do not
depend on normality assumption
• Kruskal-Wallis Test, developed in 1952 is an alternative to
ANOVA
Kruskal-Wallis Test
• Steps:
• Rank observations 𝑦𝑖𝑗 in ascending order, and 𝑅𝑖𝑗 be the rank
• In case of ties, assign average rank to each of the tied obs
1 𝑎 𝑅 𝑖 .2 𝑁 𝑁+1 2
• Test statistic: 𝐻 = [σ𝑖=1 − ] where 𝑛𝑖 is the
𝑆2 𝑛𝑖 4
number of observations in 𝑖𝑡ℎ treatment
• 𝑅𝑖 . Is the sum of the observations in the 𝑖𝑡ℎ treatment row
1 𝑎 𝑛𝑖 2 𝑁 𝑁+1 2
• 𝑆2 = [σ𝑖=1 σ𝑗=1 𝑅𝑖𝑗 − ] is the variance of ranks
𝑁−1 4
2
• If 𝐻 > 𝜒1−𝛼,𝑎−1 , hypothesis is rejected.
Factorial Design
Prof. Sayak Roychowdhury
Applications
Product
DOE Develop
ment
Improve Process
ment of
Existing Improve
Product ment
Service
Troubles
Improvem
hooting
ent
Process
Optimize!!!
Input Model
Y=f(x) Output
𝑦
(taste)
Pg
5
Sequence of Steps
• Select factors and responses
• Select experimental design
• Run experiments
• Fit model form
• Determine significant factors
• Set unimportant factors to economical settings and plan
next experiment
• Optimize the model
Sequence of Steps
• Select factors and responses
• Select experimental design
• Run experiments
• Fit model form
• Determine significant factors
• Set unimportant factors to economical settings and plan
next experiment
• Optimize the model
Factors, Levels and Responses
▪ Problem Definition
▪ Select the factor levels that make the pizza delicious in taste and
consistent in shape.
▪ Factors, Levels, and Ranges of Interest
Factors Levels (low, high) Responses
B
Factors, Levels and Responses
▪ Problem Definition
▪ Select the factor levels that make the helicopter stay in the air as long as
possible. Maybe also, for the helicopter to spin many times before it
lands.
▪ Factors, Levels, and Ranges of Interest
Factors, Levels and Responses
http://www.itl.nist.gov/div898/handbook/pri/secti
on3/pri33.htm
Fractional
Factorial
Full Factorial
Box-Behnken
Latin
Plackett- Hypercube
Burman
Central
Composite
Factorial Design
• For experiments that involve the study of the effects of two or
more factors, factorial designs are most efficient for this type
of experiment.
• Factorial design, means that in each complete trial or replicate
of the experiment all possible combinations of the levels of
the factors are investigated.
• For example, if there are 𝑎 levels of factor A and 𝑏 levels of
factor B, each replicate contains all 𝑎𝑏 treatment
combinations.
• Factorial design are often said to be crossed design.
• The effect of a factor is defined to be the change in response
produced by a change in the level of the factor.
• This is frequently called a main effect because it refers to the
primary factors of interest in the experiment.
Main Effects
2 Factor 2 Level design without interaction
Run Factor A Factor B Response Y
1 + + 40
2 - + 20
3 + - 30
4 - - 10
Main Effects:
40 B 30 + 40 10 + 20
+B- 𝑒𝐴 = 𝑌ത𝐴+ − 𝑌ത𝐴− = − = 20
30 2 2
Y
B
20 +
10
B- 20 + 40 10 + 30
𝑒𝐵 = 𝑌ത𝐵+ − 𝑌ത𝐵− = −
A- A+ 2 2
= 10
Main Effects
𝐴=
B =
Effects
2 Factor 2 Level design with interaction (difference in response in
the levels of 1 factor is not the same for all the levels of other factors)
Run Factor A Factor B AB Response Y
1 + + + 0
2 - + - 20
3 + - - 30
4 - - + 10
Main Effects:
30+0 10+20
B+
B- 𝑒𝐴 = 𝑌ത𝐴+ − 𝑌ത𝐴− = 2 − 2 = 0*
20 + 0 10 + 30
ത ത
𝑒𝐵 = 𝑌𝐵+ − 𝑌𝐵− = − = 10
Y
B- 2 2
B+
10 + 0 20 + 30
𝑒𝐴𝐵 = 𝑌ത𝐴𝐵+ − 𝑌ത𝐴𝐵− = − = −20
A- A+ 2 2
Main Effects with Interaction
𝐴=
B =
Factorial Design to Fitting Regression Lines
• The linear regression model representation
𝑦 = 𝛽0 + 𝛽1 𝑥1 + 𝛽2 𝑥2 + 𝛽12 𝑥1 𝑥2 + 𝜖
Where 𝜖~𝑁 0, 𝜎 2
𝐸𝑓𝑓𝑒𝑐𝑡𝑗
• For 2 level factorial design 𝛽𝑗 =
2
• True for both main and interaction effects
• 𝛽0 = 𝐴𝑣𝑒𝑟𝑎𝑔𝑒 𝑜𝑓 𝑎𝑙𝑙 𝑟𝑒𝑠𝑝𝑜𝑛𝑠𝑒𝑠
• These estimates are equal to least square regression
estimates.
2 Factor Factorial Design
F(0.95,b-1, ab(n-1))
B SSB b-1 SSB/(b-1) MSB/MSE
F(0.95, (a-1)(b-1)
AB SSAB (a-1)(b-1) SSAB/((a-1)(b-1)) MSAB/MSE , ab(n-1))
Selected
fraction
Selected
fraction
• Estimation of interactions
A B C D E F G Time
1 -1 -1 -1 -1 1 1 0.25
1 1 -1 1 -1 -1 -1 0.6
-1 -1 -1 1 1 1 -1 1.74
-1 -1 1 1 -1 -1 1 1.17
-1 1 -1 -1 1 -1 1 1.3
1 -1 1 -1 1 -1 -1 0.93
1 1 1 1 1 1 1 1.32
-1 1 1 -1 -1 1 -1 0.7
Steps for NPP of Effects
Factor Avg. at - Avg. at + Effect Coef. Est. "i " (i -.5)/7 Effect Factor =NORMINV((i -0.5)/7,0,1)
A 1.2275 0.775 -0.4525 -0.22625 1 0.071428571 -0.4525 A -1.4652323
B 1.0225 0.98 -0.0425 -0.02125 2 0.214285714 -0.0425 B -0.7916378
C 0.9725 1.03 0.0575 0.02875 3 0.357142857 0.0025 F -0.3661057
D 0.795 1.2075 0.4125 0.20625 4 0.5 0.0175 G 0
E 0.68 1.3225 0.6425 0.32125 5 0.642857143 0.0575 C 0.36610572
F 1 1.0025 0.0025 0.00125 6 0.785714286 0.4125 D 0.79163783
G 0.9925 1.01 0.0175 0.00875 7 0.928571429 0.6425 E 1.46523234
NPP of Effects
2
1.5 E
1
D
0.5
C
0 G
-0.6 -0.4 -0.2 0 F 0.2 0.4 0.6 0.8
-0.5
B
-1
A -1.5
-2
Example
• A chemical product is produced in a pressure vessel. Four
factors are A: Temperature, B: Pressure, C: Concentration
of Formaldehyde, D: Stirring Rate.
• Response: Filtration Rate (to be maximized)
• Additional objective: Reduce formaldehyde concentration
Example 1
𝟐𝟒 Full Factorial Design
• Calculation of effects
• Normal probability
plot of effects
• ANOVA
𝐶𝑜𝑛𝑡𝑟𝑎𝑠𝑡 2
𝑆𝑆𝐹𝑎𝑐𝑡𝑜𝑟 =
𝑁
2 𝐶𝑜𝑛𝑠𝑡𝑟𝑎𝑠𝑡
𝐸𝑓𝑓𝑒𝑐𝑡 =
𝑁
Where N= number of runs
One-Quarter Fractional Factorial
• For moderately larger number of factors 𝑘 , 2𝑘−2 fractional
factorial is often useful
• This is full factorial with 𝑘 − 2 factors, 2 additional factors
added with appropriately chosen interactions with the first
𝑘 − 2 factors
• 2 generators 𝑃 and 𝑄, 𝐼 = 𝑃 and 𝐼 = 𝑄 are called generating
relations
• Complete defining relation 𝐼 = 𝑃 = 𝑄 = 𝑃𝑄 (𝑃𝑄 is
generalized interaction).
• Elements of defining relation 𝑃, 𝑄, 𝑃𝑄 are also called words
• Aliases of an effect are produced by multiplication of that
factor with each word
One-Quarter Fractional Factorial
• Aliases of an effect are produced by multiplication of that
factor with each word
• Example: In a 26−2 design 𝐼 = 𝐴𝐵𝐶𝐸 and 𝐼 = 𝐵𝐶𝐷𝐹 are
chosen as defining generator
• Complete defining relation 𝐼 = 𝐴𝐵𝐶𝐸 = 𝐵𝐶𝐷𝐹 = 𝐴𝐷𝐸𝐹
• Aliases :
𝐴 + 𝐵𝐶𝐸 + 𝐴𝐵𝐶𝐷𝐹 + 𝐷𝐸𝐹
𝐵 + 𝐴𝐶𝐸 + 𝐶𝐷𝐹 + 𝐴𝐵𝐷𝐸𝐹
etc.
• For 26−2 design, start with full factorial 24 design, then
put 𝐸 = 𝐴𝐵𝐶, 𝐹 = 𝐵𝐶𝐷 columns
𝑘−𝑝
General 2 Design
• 2𝑘 factorial design containing 2𝑘−𝑝 runs is called 1/2𝑝
fraction of the 2𝑘 design, or 2𝑘−𝑝 fractional factorial
design
• 𝑝 independent generators to be chosen
• 2𝑝 − 𝑝 − 1 generalized interactions
• Each effect has 2𝑝 − 1 aliases
• Care should be taken so that effects of interest should not
be aliased with each other
• A reasonable criterion is the resulting 2𝑘−𝑝 design should
be of highest possible resolution
Plackett – Burman Designs
• 2 level fractional factorial designs developed by Plackett
and Burman to study 𝑘 = 𝑁 − 1 variables in 𝑁 runs
• For 𝑁 = 12, 20, 24, 28, 36 runs. These designs cannot be
represented as cubes, so they are called non-geometric
designs.
• Main effects may be partially aliased with 2 factor
interactions
• PB designs are non-regular designs. In regular designs, all
effects can be estimated distinctly, or completely aliased.
• In non-regular designs, some information on aliased
effects maybe available
Plackett – Burman Designs
• Write the appropriate row in the table as a column
• A second column is then generated from the first by moving the
element of the column down 1 position, and the placing the last
element of column 1 in first position of column 2.
• 3rd column is generated from 2nd in the same way
• Generate 𝑘 columns in this way
• A row of minus signs are then added
• This method is only for 12, 20, 24 and 36 runs
Plackett – Burman Designs
• PB designs have complex alias structure
• In a 12 run design, every main effect is partially aliased
with every 2-factor interactions, not involving itself
• PB is a non-regular design. In a regular design, all effects
can be either estimated independently of other effects or
they are completely aliased.
Choosing Fractional Factorial Designs
1.5 E
1
D
0.5
z score
0
-0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8
-0.5
-1
A -1.5
-2
Effects
P
g
3
Definitions for DOE – continued
• k1 x k2 x k3 …-- The description of the basic design of a factorial
experiment. The number of “k’s” is the number of factors. The
value of the “k” is the number of levels of interest of that factor. For
example, a 2 x 3 x 3 experiment has three factors; one input has two
levels, and two have three levels. In this case, 2x3x3 = 18
combinations for this full factorial experiment.
• k-way Interaction – an interaction between k number of variables
• Level – Values of the factor being studied in the experiment.
• Main Effect – The change in the average response observed during
the change from one level to another for a single factor
• Test Run (Experimental Run) – A single combination of factor levels
that yields one or more observations of the responses
• Treatment Combination – An experimental run using a set of specific
levels in each input variable.
P
g
3
Advanced DOE Definitions
• Confounding – One or more effects that cannot be unambiguously
attributed to a single factor or interaction. Usually due to problems
in design.
• Fixed Factor – Factors whose levels are specifically chosen.
Conclusions about fixed factors generalize to only those levels.
Determination of the factor’s effect on the level of the output is
usually the goal of the experiment. (Knob settings, for example)
• Random Factor – Factors whose levels are selected randomly from a
larger population of possible levels. Determination of the factor’s
contribution to the overall variance of the system is the goal of this
experiment. (Selecting 3 machines out of 20, for example)
• Repetitions – consecutive experimental runs using the same
treatment combinations. With no change in setup between runs.
• Replications – experimental runs using the same treatment
combinations that were not run consecutively. Replications are
duplicate runs of the entire experiment. Sample size calculations
apply to replicates, not repetitions.
• http://www.itl.nist.gov/div898/handbook/pri/section7/pri7.htm
P
g
4
Factors, Levels and Responses
▪ Problem Definition
▪ Select the factor levels that make the pizza delicious in taste and
consistent in shape.
▪ Factors, Levels, and Ranges of Interest
Factors Levels (low, high) Responses
What if said the overall task is to minimize cost while satisfying your customers.
• Hickory Wood is not much more expensive than Unknown Wood
• Expensive Flour is significantly higher than cheap flour
• Also, the kneading machine can be set at a high or low level without significantly increasing the energy usage (and
hence, cost).
• Customer satisfaction ratings were consistently high for Taste levels of 8 or above and Crust Consistency levels of 6
or over.
Examine carefully all the model graphs and pick the best recipe for pizza that will minimize your cost while still
ensuring high customer satisfaction. (After Narrowing it Down to 4 Significant factors)
ANOVA- Taste
ANOVA-Consistency