Even if this "best" setting were included in the design, you should run it again as part of the confirmation runs to make sure nothing has changed and that the response values are close
Trang 2Normal or half-normal plots of effects (primarily for two-level full and fractional factorial experiments)
❍
Youden plots
❍
Other methods
❍
5.4.4 How to test and revise DOE models
Trang 35 Process Improvement
5.4 Analysis of DOE data
5.4.6 How to confirm DOE results
(confirmatory runs)
Definition of
confirmation
runs
When the analysis of the experiment is complete, one must verify that the predictions are good These are called confirmation runs.
The interpretation and conclusions from an experiment may include a
"best" setting to use to meet the goals of the experiment Even if this
"best" setting were included in the design, you should run it again as part of the confirmation runs to make sure nothing has changed and that the response values are close to their predicted values would get.
At least 3
confirmation
runs should
be planned
In an industrial setting, it is very desirable to have a stable process Therefore, one should run more than one test at the "best" settings A minimum of 3 runs should be conducted (allowing an estimate of variability at that setting).
If the time between actually running the experiment and conducting the confirmation runs is more than a few hours, the experimenter must be careful to ensure that nothing else has changed since the original data collection.
Carefully
duplicate the
original
environment
The confirmation runs should be conducted in an environment as similar as possible to the original experiment For example, if the experiment were conducted in the afternoon and the equipment has a warm-up effect, the confirmation runs should be conducted in the afternoon after the equipment has warmed up Other extraneous factors that may change or affect the results of the confirmation runs are:
person/operator on the equipment, temperature, humidity, machine parameters, raw materials, etc.
Trang 4Checks for
when
confirmation
runs give
surprises
What do you do if you don't obtain the results you expected? If the confirmation runs don't produce the results you expected:
check to see that nothing has changed since the original data collection
1
verify that you have the correct settings for the confirmation runs
2
revisit the model to verify the "best" settings from the analysis
3
verify that you had the correct predicted value for the confirmation runs.
4
If you don't find the answer after checking the above 4 items, the model may not predict very well in the region you decided was "best" You still learned from the experiment and you should use the
information gained from this experiment to design another follow-up experiment.
Even when
the
experimental
goals are not
met,
something
was learned
that can be
used in a
follow-up
experiment
Every well-designed experiment is a success in that you learn something from it However, every experiment will not necessarily meet the goals established before experimentation That is why it makes sense to plan to experiment sequentially in order to meet the goals.
5.4.6 How to confirm DOE results (confirmatory runs)
Trang 55 Process Improvement
5.4 Analysis of DOE data
5.4.7 Examples of DOE's
5.4.7.1 Full factorial example
Data Source
This example
uses data from
a NIST high
performance
ceramics
experiment
This data set was taken from an experiment that was performed a few years ago at NIST (by Said Jahanmir of the Ceramics Division in the Material Science and Engineering Laboratory) The original analysis was performed primarily by Lisa Gill of the Statistical Engineering Division The example shown here is an independent analysis of a modified portion of the original data set The original data set was part of a high performance ceramics experiment with the goal of
characterizing the effect of grinding parameters on sintered reaction-bonded silicon nitride, reaction bonded silicone nitride, and sintered silicon nitride
Only modified data from the first of the 3 ceramic types (sintered reaction-bonded silicon nitride) will be discussed in this illustrative example of a full factorial data analysis
The reader may want to download the data as a text file and try using other software packages to analyze the data
Description of Experiment: Response and Factors
Response and
factor
variables used
in the
experiment
Purpose: To determine the effect of machining factors on ceramic strength Response variable = mean (over 15 repetitions) of the ceramic strength Number of observations = 32 (a complete 25 factorial design)
Response Variable Y = Mean (over 15 reps) of Ceramic Strength Factor 1 = Table Speed (2 levels: slow (.025 m/s) and fast (.125 m/s)) Factor 2 = Down Feed Rate (2 levels: slow (.05 mm) and fast (.125 mm)) Factor 3 = Wheel Grit (2 levels: 140/170 and 80/100)
Factor 4 = Direction (2 levels: longitudinal and transverse) Factor 5 = Batch (2 levels: 1 and 2)
Since two factors were qualitative (direction and batch) and it was reasonable to expect monotone effects from the quantitative factors, no centerpoint runs were included
Trang 6spreadsheet of
the data
The design matrix, with measured ceramic strength responses, appears below The actual randomized run order is given in the last column (The interested reader may download the data
as a text file or as a JMP file.)
Analysis of the Experiment
Analysis
follows 5 basic
steps
The experimental data will be analyzed following the previously described 5 basic steps using SAS JMP 3.2.6 software
Step 1: Look at the data
5.4.7.1 Full factorial example
Trang 7Plot the
response
variable
We start by plotting the response data several ways to see if any trends or anomalies appear that would not be accounted for by the standard linear response models
First we look at the distribution of all the responses irrespective of factor levels
The following plots were generared:
The first plot is a normal probability plot of the response variable The straight red line is the fitted nornal distribution and the curved red lines form a simultaneous 95% confidence region for the plotted points, based on the assumption of normality
1
The second plot is a box plot of the response variable The "diamond" is called (in JMP) a
"means diamond" and is centered around the sample mean, with endpoints spanning a 95% normal confidence interval for the sample mean
2
The third plot is a histogram of the response variable
3
Clearly there is "structure" that we hope to account for when we fit a response model For example, note the separation of the response into two roughly equal-sized clumps in the histogram The first clump is centered approximately around the value 450 while the second clump is centered approximately around the value 650
Trang 8Plot of
response
versus run
order
Next we look at the responses plotted versus run order to check whether there might be a time sequence component affecting the response levels
Plot of Response Vs Run Order
As hoped for, this plot does not indicate that time order had much to do with the response levels
Box plots of
response by
factor
variables
Next, we look at plots of the responses sorted by factor columns
5.4.7.1 Full factorial example
Trang 9Several factors, most notably "Direction" followed by "Batch" and possibly "Wheel Grit", appear
to change the average response level
Step 2: Create the theoretical model
Theoretical
model: assume
all 4-factor and
higher
interaction
terms are not
significant
With a 25 full factorial experiment we can fit a model containing a mean term, all 5 main effect terms, all 10 2-factor interaction terms, all 10 3-factor interaction terms, all 5 4-factor interaction terms and the 5-factor interaction term (32 parameters) However, we start by assuming all three factor and higher interaction terms are non-existent (it's very rare for such high-order interactions
to be significant, and they are very difficult to interpret from an engineering viewpoint) That allows us to accumulate the sums of squares for these terms and use them to estimate an error term So we start out with a theoretical model with 26 unknown constants, hoping the data will clarify which of these are the significant main effects and interactions we need for a final model
Step 3: Create the actual model from the data
Output from
fitting up to
third-order
interaction
terms
After fitting the 26 parameter model, the following analysis table is displayed:
Output after Fitting Third Order Model to Response Data
Response: Y: Strength
Summary of Fit RSquare 0.995127 RSquare Adj 0.974821 Root Mean Square Error 17.81632 Mean of Response 546.8959 Observations 32
Effect Test Sum Source DF of Squares F Ratio Prob>F X1: Table Speed 1 894.33 2.8175 0.1442 X2: Feed Rate 1 3497.20 11.0175 0.0160 X1: Table Speed* 1 4872.57 15.3505 0.0078 X2: Feed Rate
X3: Wheel Grit 1 12663.96 39.8964 0.0007 X1: Table Speed* 1 1838.76 5.7928 0.0528 X3: Wheel Grit
Trang 10X2: Feed Rate* 1 307.46 0.9686 0.3630 X3: Wheel Grit
X1:Table Speed* 1 357.05 1.1248 0.3297 X2: Feed Rate*
X3: Wheel Grit X4: Direction 1 315132.65 992.7901 <.0001 X1: Table Speed* 1 1637.21 5.1578 0.0636 X4: Direction
X2: Feed Rate* 1 1972.71 6.2148 0.0470 X4: Direction
X1: Table Speed 1 5895.62 18.5735 0.0050 X2: Feed Rate*
X4: Direction X3: Wheel Grit* 1 3158.34 9.9500 0.0197 X4: Direction
X1: Table Speed* 1 2.12 0.0067 0.9376 X3: Wheel Grit*
X4: Direction X2: Feed Rate* 1 44.49 0.1401 0.7210 X3: Wheel Grit*
X4: Direction X5: Batch 1 33653.91 106.0229 <.0001 X1: Table Speed* 1 465.05 1.4651 0.2716 X5: Batch
X2: Feed Rate* 1 199.15 0.6274 0.4585 X5: Batch
X1: Table Speed* 1 144.71 0.4559 0.5247 X2: Feed Rate*
X5: Batch X3: Wheel Grit* 1 29.36 0.0925 0.7713 X5: Batch
X1: Table Speed* 1 30.36 0.0957 0.7676 X3: Wheel Grit*
X5: Batch X2: Feed Rate* 1 25.58 0.0806 0.7860 X3: Wheel Grit*
X5: Batch X4: Direction * 1 1328.83 4.1863 0.0867 X5: Batch
X1: Table Speed* 1 544.58 1.7156 0.2382 X4: Directio*
X5: Batch X2: Feed Rate* 1 167.31 0.5271 0.4952
5.4.7.1 Full factorial example
Trang 11JMP stepwise
regression
Starting with these 26 terms, we next use the JMP Stepwise Regression option to eliminate unnecessary terms By a combination of stepwise regression and the removal of remaining terms
with a p-value higher than 0.05, we quickly arrive at a model with an intercept and 12 significant
effect terms
Output from
fitting the
12-term model Output after Fitting the 12-Term Model to Response Data
Response: Y: Strength
Summary of Fit RSquare 0.989114 RSquare Adj 0.982239 Root Mean Square Error 14.96346 Mean of Response 546.8959
Observations (or Sum Wgts) 32
Effect Test Sum Source DF of Squares F Ratio Prob>F X1: Table Speed 1 894.33 3.9942 0.0602 X2: Feed Rate 1 3497.20 15.6191 0.0009 X1: Table Speed* 1 4872.57 21.7618 0.0002 X2: Feed Rate
X3: Wheel Grit 1 12663.96 56.5595 <.0001 X1: Table Speed* 1 1838.76 8.2122 0.0099 X3: Wheel Grit
X4: Direction 1 315132.65 1407.4390 <.0001 X1: Table Speed* 1 1637.21 7.3121 0.0141 X4: Direction
X2: Feed Rate* 1 1972.71 8.8105 0.0079 X4: Direction
X1: Table Speed* 1 5895.62 26.3309 <.0001 X2: Feed Rate*
X4:Direction X3: Wheel Grit* 1 3158.34 14.1057 0.0013 X4: Direction
X5: Batch 1 33653.91 150.3044 <.0001 X4: Direction* 1 1328.83 5.9348 0.0249 X5: Batch
Trang 12Normal plot of
the effects
Non-significant effects should effectively follow an approximately normal distribution with the same location and scale Significant effects will vary from this normal distribution Therefore, another method of determining significant effects is to generate a normal plot of all 31 effects Those effects that are substantially away from the straight line fitted to the normal plot are considered significant Although this is a somewhat subjective criteria, it tends to work well in practice It is helpful to use both the numerical output from the fit and graphical techniques such
as the normal plot in deciding which terms to keep in the model
The normal plot of the effects is shown below We have labeled those effects that we consider to
be significant In this case, we have arrived at the exact same 12 terms by looking at the normal plot as we did from the stepwise regression
5.4.7.1 Full factorial example
Trang 13Model appears
to account for
most of the
variability
At this stage, this model appears to account for most of the variability in the response, achieving
an adjusted R2 of 0.982 All the main effects are significant, as are 6 2-factor interactions and 1
3-factor interaction The only interaction that makes little physical sense is the " X4:
Direction*X5: Batch" interaction - why would the response using one batch of material react
differently when the batch is cut in a different direction as compared to another batch of the same formulation?
However, before accepting any model, residuals need to be examined
Step 4: Test the model assumptions using residual graphs (adjust and simplify as needed)
Plot of
residuals
versus
predicted
responses
First we look at the residuals plotted versus the predicted responses
The residuals appear to spread out more with larger values of predicted strength, which should not happen when there is a common variance
Next we examine the normality of the residuals with a normal quantile plot, a box plot and a histogram
Trang 14None of these plots appear to show typical normal residuals and 4 of the 32 data points appear as outliers in the box plot
Step 4 continued: Transform the data and fit the model again
Box-Cox
Transformation
We next look at whether we can model a transformation of the response variable and obtain residuals with the assumed properties JMP calculates an optimum Box-Cox transformation by finding the value of that minimizes the model SSE Note: the Box-Cox transformation used in JMP is different from the transformation used in Dataplot, but roughly equivalent
Box-Cox Transformation Graph
5.4.7.1 Full factorial example
Trang 15JMP data
transformation
menu
Data Transformation Column Properties
Fit model to
transformed
data
When the 12-effect model is fit to the transformed data, the "X4: Direction*X5: Batch"
interaction term is no longer significant The 11-effect model fit is shown below, with parameter
estimates and p-values.
JMP output for
fitted model
after applying
Box-Cox
transformation
Output after Fitting the 11-Effect Model to
Tranformed Response Data
Response: Y: Strength X
Summary of Fit RSquare 0.99041 RSquare Adj 0.985135 Root Mean Square Error 13.81065 Mean of Response 1917.115
Observations (or Sum Wgts) 32
Parameter Effect Estimate p-value Intercept 1917.115 <.0001 X1: Table Speed 5.777 0.0282 X2: Feed Rate 11.691 0.0001 X1: Table Speed* -14.467 <.0001 X2: Feed Rate
X3: Wheel Grit -21.649 <.0001 X1: Table Speed* 7.339 0.007 X3: Wheel Grit
X4: Direction -99.272 <.0001 X1: Table Speed* -7.188 0.0080 X4: Direction
X2: Feed Rate* -9.160 0.0013 X4: Direction