1. Trang chủ
  2. » Kỹ Thuật - Công Nghệ

Engineering Statistics Handbook Episode 7 Part 3 potx

18 451 0
Tài liệu đã được kiểm tra trùng lặp

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 18
Dung lượng 220,89 KB

Các công cụ chuyển đổi và chỉnh sửa cho tài liệu này

Nội dung

Even if this "best" setting were included in the design, you should run it again as part of the confirmation runs to make sure nothing has changed and that the response values are close

Trang 2

Normal or half-normal plots of effects (primarily for two-level full and fractional factorial experiments)

Youden plots

Other methods

5.4.4 How to test and revise DOE models

Trang 3

5 Process Improvement

5.4 Analysis of DOE data

5.4.6 How to confirm DOE results

(confirmatory runs)

Definition of

confirmation

runs

When the analysis of the experiment is complete, one must verify that the predictions are good These are called confirmation runs.

The interpretation and conclusions from an experiment may include a

"best" setting to use to meet the goals of the experiment Even if this

"best" setting were included in the design, you should run it again as part of the confirmation runs to make sure nothing has changed and that the response values are close to their predicted values would get.

At least 3

confirmation

runs should

be planned

In an industrial setting, it is very desirable to have a stable process Therefore, one should run more than one test at the "best" settings A minimum of 3 runs should be conducted (allowing an estimate of variability at that setting).

If the time between actually running the experiment and conducting the confirmation runs is more than a few hours, the experimenter must be careful to ensure that nothing else has changed since the original data collection.

Carefully

duplicate the

original

environment

The confirmation runs should be conducted in an environment as similar as possible to the original experiment For example, if the experiment were conducted in the afternoon and the equipment has a warm-up effect, the confirmation runs should be conducted in the afternoon after the equipment has warmed up Other extraneous factors that may change or affect the results of the confirmation runs are:

person/operator on the equipment, temperature, humidity, machine parameters, raw materials, etc.

Trang 4

Checks for

when

confirmation

runs give

surprises

What do you do if you don't obtain the results you expected? If the confirmation runs don't produce the results you expected:

check to see that nothing has changed since the original data collection

1

verify that you have the correct settings for the confirmation runs

2

revisit the model to verify the "best" settings from the analysis

3

verify that you had the correct predicted value for the confirmation runs.

4

If you don't find the answer after checking the above 4 items, the model may not predict very well in the region you decided was "best" You still learned from the experiment and you should use the

information gained from this experiment to design another follow-up experiment.

Even when

the

experimental

goals are not

met,

something

was learned

that can be

used in a

follow-up

experiment

Every well-designed experiment is a success in that you learn something from it However, every experiment will not necessarily meet the goals established before experimentation That is why it makes sense to plan to experiment sequentially in order to meet the goals.

5.4.6 How to confirm DOE results (confirmatory runs)

Trang 5

5 Process Improvement

5.4 Analysis of DOE data

5.4.7 Examples of DOE's

5.4.7.1 Full factorial example

Data Source

This example

uses data from

a NIST high

performance

ceramics

experiment

This data set was taken from an experiment that was performed a few years ago at NIST (by Said Jahanmir of the Ceramics Division in the Material Science and Engineering Laboratory) The original analysis was performed primarily by Lisa Gill of the Statistical Engineering Division The example shown here is an independent analysis of a modified portion of the original data set The original data set was part of a high performance ceramics experiment with the goal of

characterizing the effect of grinding parameters on sintered reaction-bonded silicon nitride, reaction bonded silicone nitride, and sintered silicon nitride

Only modified data from the first of the 3 ceramic types (sintered reaction-bonded silicon nitride) will be discussed in this illustrative example of a full factorial data analysis

The reader may want to download the data as a text file and try using other software packages to analyze the data

Description of Experiment: Response and Factors

Response and

factor

variables used

in the

experiment

Purpose: To determine the effect of machining factors on ceramic strength Response variable = mean (over 15 repetitions) of the ceramic strength Number of observations = 32 (a complete 25 factorial design)

Response Variable Y = Mean (over 15 reps) of Ceramic Strength Factor 1 = Table Speed (2 levels: slow (.025 m/s) and fast (.125 m/s)) Factor 2 = Down Feed Rate (2 levels: slow (.05 mm) and fast (.125 mm)) Factor 3 = Wheel Grit (2 levels: 140/170 and 80/100)

Factor 4 = Direction (2 levels: longitudinal and transverse) Factor 5 = Batch (2 levels: 1 and 2)

Since two factors were qualitative (direction and batch) and it was reasonable to expect monotone effects from the quantitative factors, no centerpoint runs were included

Trang 6

spreadsheet of

the data

The design matrix, with measured ceramic strength responses, appears below The actual randomized run order is given in the last column (The interested reader may download the data

as a text file or as a JMP file.)

Analysis of the Experiment

Analysis

follows 5 basic

steps

The experimental data will be analyzed following the previously described 5 basic steps using SAS JMP 3.2.6 software

Step 1: Look at the data

5.4.7.1 Full factorial example

Trang 7

Plot the

response

variable

We start by plotting the response data several ways to see if any trends or anomalies appear that would not be accounted for by the standard linear response models

First we look at the distribution of all the responses irrespective of factor levels

The following plots were generared:

The first plot is a normal probability plot of the response variable The straight red line is the fitted nornal distribution and the curved red lines form a simultaneous 95% confidence region for the plotted points, based on the assumption of normality

1

The second plot is a box plot of the response variable The "diamond" is called (in JMP) a

"means diamond" and is centered around the sample mean, with endpoints spanning a 95% normal confidence interval for the sample mean

2

The third plot is a histogram of the response variable

3

Clearly there is "structure" that we hope to account for when we fit a response model For example, note the separation of the response into two roughly equal-sized clumps in the histogram The first clump is centered approximately around the value 450 while the second clump is centered approximately around the value 650

Trang 8

Plot of

response

versus run

order

Next we look at the responses plotted versus run order to check whether there might be a time sequence component affecting the response levels

Plot of Response Vs Run Order

As hoped for, this plot does not indicate that time order had much to do with the response levels

Box plots of

response by

factor

variables

Next, we look at plots of the responses sorted by factor columns

5.4.7.1 Full factorial example

Trang 9

Several factors, most notably "Direction" followed by "Batch" and possibly "Wheel Grit", appear

to change the average response level

Step 2: Create the theoretical model

Theoretical

model: assume

all 4-factor and

higher

interaction

terms are not

significant

With a 25 full factorial experiment we can fit a model containing a mean term, all 5 main effect terms, all 10 2-factor interaction terms, all 10 3-factor interaction terms, all 5 4-factor interaction terms and the 5-factor interaction term (32 parameters) However, we start by assuming all three factor and higher interaction terms are non-existent (it's very rare for such high-order interactions

to be significant, and they are very difficult to interpret from an engineering viewpoint) That allows us to accumulate the sums of squares for these terms and use them to estimate an error term So we start out with a theoretical model with 26 unknown constants, hoping the data will clarify which of these are the significant main effects and interactions we need for a final model

Step 3: Create the actual model from the data

Output from

fitting up to

third-order

interaction

terms

After fitting the 26 parameter model, the following analysis table is displayed:

Output after Fitting Third Order Model to Response Data

Response: Y: Strength

Summary of Fit RSquare 0.995127 RSquare Adj 0.974821 Root Mean Square Error 17.81632 Mean of Response 546.8959 Observations 32

Effect Test Sum Source DF of Squares F Ratio Prob>F X1: Table Speed 1 894.33 2.8175 0.1442 X2: Feed Rate 1 3497.20 11.0175 0.0160 X1: Table Speed* 1 4872.57 15.3505 0.0078 X2: Feed Rate

X3: Wheel Grit 1 12663.96 39.8964 0.0007 X1: Table Speed* 1 1838.76 5.7928 0.0528 X3: Wheel Grit

Trang 10

X2: Feed Rate* 1 307.46 0.9686 0.3630 X3: Wheel Grit

X1:Table Speed* 1 357.05 1.1248 0.3297 X2: Feed Rate*

X3: Wheel Grit X4: Direction 1 315132.65 992.7901 <.0001 X1: Table Speed* 1 1637.21 5.1578 0.0636 X4: Direction

X2: Feed Rate* 1 1972.71 6.2148 0.0470 X4: Direction

X1: Table Speed 1 5895.62 18.5735 0.0050 X2: Feed Rate*

X4: Direction X3: Wheel Grit* 1 3158.34 9.9500 0.0197 X4: Direction

X1: Table Speed* 1 2.12 0.0067 0.9376 X3: Wheel Grit*

X4: Direction X2: Feed Rate* 1 44.49 0.1401 0.7210 X3: Wheel Grit*

X4: Direction X5: Batch 1 33653.91 106.0229 <.0001 X1: Table Speed* 1 465.05 1.4651 0.2716 X5: Batch

X2: Feed Rate* 1 199.15 0.6274 0.4585 X5: Batch

X1: Table Speed* 1 144.71 0.4559 0.5247 X2: Feed Rate*

X5: Batch X3: Wheel Grit* 1 29.36 0.0925 0.7713 X5: Batch

X1: Table Speed* 1 30.36 0.0957 0.7676 X3: Wheel Grit*

X5: Batch X2: Feed Rate* 1 25.58 0.0806 0.7860 X3: Wheel Grit*

X5: Batch X4: Direction * 1 1328.83 4.1863 0.0867 X5: Batch

X1: Table Speed* 1 544.58 1.7156 0.2382 X4: Directio*

X5: Batch X2: Feed Rate* 1 167.31 0.5271 0.4952

5.4.7.1 Full factorial example

Trang 11

JMP stepwise

regression

Starting with these 26 terms, we next use the JMP Stepwise Regression option to eliminate unnecessary terms By a combination of stepwise regression and the removal of remaining terms

with a p-value higher than 0.05, we quickly arrive at a model with an intercept and 12 significant

effect terms

Output from

fitting the

12-term model Output after Fitting the 12-Term Model to Response Data

Response: Y: Strength

Summary of Fit RSquare 0.989114 RSquare Adj 0.982239 Root Mean Square Error 14.96346 Mean of Response 546.8959

Observations (or Sum Wgts) 32

Effect Test Sum Source DF of Squares F Ratio Prob>F X1: Table Speed 1 894.33 3.9942 0.0602 X2: Feed Rate 1 3497.20 15.6191 0.0009 X1: Table Speed* 1 4872.57 21.7618 0.0002 X2: Feed Rate

X3: Wheel Grit 1 12663.96 56.5595 <.0001 X1: Table Speed* 1 1838.76 8.2122 0.0099 X3: Wheel Grit

X4: Direction 1 315132.65 1407.4390 <.0001 X1: Table Speed* 1 1637.21 7.3121 0.0141 X4: Direction

X2: Feed Rate* 1 1972.71 8.8105 0.0079 X4: Direction

X1: Table Speed* 1 5895.62 26.3309 <.0001 X2: Feed Rate*

X4:Direction X3: Wheel Grit* 1 3158.34 14.1057 0.0013 X4: Direction

X5: Batch 1 33653.91 150.3044 <.0001 X4: Direction* 1 1328.83 5.9348 0.0249 X5: Batch

Trang 12

Normal plot of

the effects

Non-significant effects should effectively follow an approximately normal distribution with the same location and scale Significant effects will vary from this normal distribution Therefore, another method of determining significant effects is to generate a normal plot of all 31 effects Those effects that are substantially away from the straight line fitted to the normal plot are considered significant Although this is a somewhat subjective criteria, it tends to work well in practice It is helpful to use both the numerical output from the fit and graphical techniques such

as the normal plot in deciding which terms to keep in the model

The normal plot of the effects is shown below We have labeled those effects that we consider to

be significant In this case, we have arrived at the exact same 12 terms by looking at the normal plot as we did from the stepwise regression

5.4.7.1 Full factorial example

Trang 13

Model appears

to account for

most of the

variability

At this stage, this model appears to account for most of the variability in the response, achieving

an adjusted R2 of 0.982 All the main effects are significant, as are 6 2-factor interactions and 1

3-factor interaction The only interaction that makes little physical sense is the " X4:

Direction*X5: Batch" interaction - why would the response using one batch of material react

differently when the batch is cut in a different direction as compared to another batch of the same formulation?

However, before accepting any model, residuals need to be examined

Step 4: Test the model assumptions using residual graphs (adjust and simplify as needed)

Plot of

residuals

versus

predicted

responses

First we look at the residuals plotted versus the predicted responses

The residuals appear to spread out more with larger values of predicted strength, which should not happen when there is a common variance

Next we examine the normality of the residuals with a normal quantile plot, a box plot and a histogram

Trang 14

None of these plots appear to show typical normal residuals and 4 of the 32 data points appear as outliers in the box plot

Step 4 continued: Transform the data and fit the model again

Box-Cox

Transformation

We next look at whether we can model a transformation of the response variable and obtain residuals with the assumed properties JMP calculates an optimum Box-Cox transformation by finding the value of that minimizes the model SSE Note: the Box-Cox transformation used in JMP is different from the transformation used in Dataplot, but roughly equivalent

Box-Cox Transformation Graph

5.4.7.1 Full factorial example

Trang 15

JMP data

transformation

menu

Data Transformation Column Properties

Fit model to

transformed

data

When the 12-effect model is fit to the transformed data, the "X4: Direction*X5: Batch"

interaction term is no longer significant The 11-effect model fit is shown below, with parameter

estimates and p-values.

JMP output for

fitted model

after applying

Box-Cox

transformation

Output after Fitting the 11-Effect Model to

Tranformed Response Data

Response: Y: Strength X

Summary of Fit RSquare 0.99041 RSquare Adj 0.985135 Root Mean Square Error 13.81065 Mean of Response 1917.115

Observations (or Sum Wgts) 32

Parameter Effect Estimate p-value Intercept 1917.115 <.0001 X1: Table Speed 5.777 0.0282 X2: Feed Rate 11.691 0.0001 X1: Table Speed* -14.467 <.0001 X2: Feed Rate

X3: Wheel Grit -21.649 <.0001 X1: Table Speed* 7.339 0.007 X3: Wheel Grit

X4: Direction -99.272 <.0001 X1: Table Speed* -7.188 0.0080 X4: Direction

X2: Feed Rate* -9.160 0.0013 X4: Direction

Ngày đăng: 06/08/2014, 11:20

TỪ KHÓA LIÊN QUAN