You are on page 1of 13

CHAPTER 3

DESIGN
OF
EXPERIMENTS

56

CHAPTER 3
DESIGN OF EXPERIMENTS
3.1 INTRODUCTION
Design of Experiment (DoE) is a useful method in identifying the significant parameters and
in studying the possible effect of the variables during the machining trials. This method also
can developed experiment between a ranges from uncontrollable factors, which will be
introduced randomly to carefully controlled parameters. The factors must be either
quantitative or qualitative. The range of values for quantitative factors must be decided on
how they are going to be measured and the level at which they will be controlled during the
trials. Meanwhile, the qualitative factors are parameters that will be determined discretely.
The advantages of design of experiments are as follows:

Numbers of trials is significantly reduced.

Important decision variables which control and improve the performance of the
product or the process can be identified.

Optimal setting of the parameters can be found out.

Qualitative estimation of parameters can be made.

Experimental error can be estimated.

Inference regarding the effect of parameters on the characteristics of the process


can be made.

Experiments are performed by investigators in virtually all fields of inquiry, usually to


discover something about a particular process or system. Generally experiments is test or
series of test in which purposeful changes are made to the input variable of the process or
system so that we may observe and identify the reason for changes that may be observed in
the output response and analyzing the result data so that valid and objective conclusions
are obtained.

57

The objective of the experiments may include the following:


I.

Determine which variable are most influential on the performance measures or


response.

II.

Determine where to set the influential controllable process variables so that response
is almost always near the desired nominal value.

III.

Determine where to set the controllable process variables so that variability in


response is small.

IV.

Determine where to set the influential process variables so that the effects of the
uncontrollable variables are minimized.

Statistical design of experiments refers to the process of planning the experiment so that
appropriate data that can be analyzed by statistical methods will be collected, resulting in
valid and objective conclusions.
The statistical approach to experimental design is necessary if we wish to draw a
meaningful conclusion from the data [62].

Figure3.1: General model of a process or system [63]

58

3.2 APPROCHE TO DESIGN OF EXPERIMENT


I.

Multiple Regression Analysis

II.

Mathematical Modeling

III.

Orthogonal Array

IV.

ANOVA

3.2.1 MULTIPLE REGRESSION ANALYSIS


Main purpose of multiple regression analysis (first used by Pearson in 1908) is to learn about
the several independent or predictor variables and dependable variable [64].

Multiple regression is a statistical technique that allows to predict someones score on one
variable on the basis of their scores on several other variables. Multiple linear regression
examines the linear relationships between one continuous response and two or more
predictors. The independent variable that is used to predict values of the dependent, or
response variable in a regression analysis, If the number of predictors is large, then before
fitting a regression model with all the predictors, use stepwise techniques to screen out
predictors not associated with the responses [65].
A current trend in statistics is to emphasize the similarity between multiple regression and
ANOVA, and between correlation and the t-test. All of these statistical techniques are
basically seeking to do the same thing explain the variance in the level of one variable on
the basis of the level of one or more other variables. These other variables might be
manipulated directly in the case of controlled experiments, or be observed in the case of
surveys or observational studies, but the underlying principle is the same. Thus, although we
have given separate chapters to each of these procedures they are fundamentally all the
same procedure. This underlying single approach is called the General Linear Model [64].

Multiple linear regression attempts to model the relationship between two or more
explanatory variables and a response variable by fitting a linear equation to observed data
[66].
Every value of the independent variable x is associated with a value of the dependent
variable y.
59

The regression line for p explanatory variables x1, x2,. ..., xp is defined to be
= + + + +
The mean response , described by the regression line changes with the explanatory
variables. The observed values for y vary about their means and are assumed to have
the

same

standard

deviation .

The

fitted

values b0, b1,..., bp estimate

the

parameters , , of the regression line. Since the observed values for y vary about
their means , the multiple regression model includes a term for this variation.
The regression model is expressed as Data = Fit + Residual [66], where the
"Fit" term represents the expression + + + + .
The "Residual" term represents the deviations of the observed values y from their
means , which is normally distributed with mean 0 and variance . The notation for the
model deviations is

Formally, the model for multiple linear regression, given n observations, is


= + + + + + = , ,
In the least-squares model, the best-fitting line for the observed data is calculated by
minimizing the sum of the squares of the vertical deviations from each data point to the
line. Since the deviations are first squared, then summed, there are no cancellations
between positive and negative values. The least-squares estimates b0, b1, ... bp are usually
computed by statistical software.

i , and the residuals ei are


The values fit by the equation b0 + b1xi1 + ... + bpxip are denoted y
equal to yi

y , the difference between the observed and fitted values.


i

The sum of the residuals is equal to zero.

60

The variance may be estimated by =

also known as the mean-squared error

(or MSE).
The estimate of the standard error s is the square root of the MSE [66].
3.2.2 MATHEMATICAL MODELING
Once experimental design becomes final, the next step is to fit the given data in
mathematical model using regression analysis.
A mathematical model is a description of a system using mathematical concepts and
language. The process of developing a mathematical model is termed mathematical
modeling. Mathematical models are used in engineers, statisticians, research analysts
and economists use mathematical models most extensively. In general, mathematical
models may include logical models, as far as logic is taken as a part of mathematics. In many
cases, the quality of a scientific field depends on how well the mathematical models
developed on the theoretical side agree with results of repeatable experiments. Lack of
agreement between theoretical mathematical models and experimental measurements
often leads to important advances as better theories are developed [67].
3.2.3 ORTHOGONAL ARRAY
Orthogonal array (OA) represents a simplified method of putting together an experiment.
Taguhis orthogonal arrays are selected on the basis of the condition that the total degree of
freedom of selected OA must be greater than or equal to the total degree of freedom
required for the experiment [68].
An orthogonal array provides a set of well balance (minimum experimental runs)
experiments and used to design experiments and describe trial condition. Experiments
design using orthogonal arrays yield result that are more reproducible.
Standard notation for orthogonal arrays [69].
LN (XY)
Where,
N = Number of experiments, X = Number of levels, Y = Number of factors
61

For example:
2- Level Arrays: L4 (27), L12 (211), L16 (215)
3- Level Arrays: L9 (34), L18 (2137), L27 (313)
4- Level Arrays: L16 (45), L32 (2148)
Example: L9 (34)
9 = Number of experiments, 3 = Number of levels, 4 = Number of factors
Taguchis orthogonal arrays are experimental designs that usually require only a fraction of
full factorial combination. The columns of arrays are balanced and orthogonal i.e, in each
pair of columns, all factors combinations occurs the same number of times. Orthogonal
designs allow estimating the effect of each factor on the response independently of all other
factors.
There are 18 basic types of standard Orthogonal array (OA) in the Taguchi parameter design
[]. Since four factors were studied in the present work, three levels of each were considered.
Therefore an L9 (34) Orthogonal array has been selected in the present study, for multiperformance optimisation shown in Table 3.1.

62

Table 3.1 Shows L9 (34) orthogonal array


Factors
A

Runs

3.2.4 ANOVA
The purpose of the ANOVA is to investigate which wire EDM process parameters
significantly affect the quality characteristics. This is accomplished by separating the total
variability of the S/N ratios, which is measured by the sum of the squared deviations from
the total mean of S/N ratio, into contributions by each Wire EDM process parameter and
error. The percentage contribution by each of the process parameter in the total sum of
squared deviations can be used to evaluate the importance of the process parameter
change on the quality characteristic. In addition the F-test method can also be used to
determine which Wire EDM process parameter has a significant effect on the quality
characteristic when the F value is large. The fundamental technique is a partitioning of the
total sum of squares S into components related to the effects used in the model. For
63

example, we show the model for a simplified ANOVA with one type of treatment at different
levels.
= +
So, the number of degrees of freedom f can be partitioned in a similar way and specifies
the chi- squared distribution which describes the associated sums of squares
= +
The F-test is used for comparisons of the components of the total deviation. For example, in
one-way or single-factor ANOVA, statistical significance is tested for by comparing the F test
statistic [70].
ANOVA for Multiple Linear Regression:
Multiple linear regression tries to fit a regression line for a response variable by using more
than one explanatory variable. ANOVA calculations for multiple regression are nearly
identical to the calculations for simple linear regression, except that the degrees of freedom
are adjusted to reflect the number of explanatory variables included in the model [71].
3.3 TEST FOR SIGNIFICANCE OF REGRESSION [72]
The test for significance is a test to determine if there is a linear relationship between the
response variable y and a subset of the regressor variables x1, x2,.., xk.
Once the co-efficient have been estimated and tested for their significance, the estimated
regression equation is then tested for the adequacy of fit.
The appropriate hypotheses are

: = = = =
:

Rejection of in equation indicates that at least one of the regressor variables x1, x2,..,
xk. Contribute significantly to the model. The test procedure involves an analysis of variance
partitioning of the total sum of squares due to the model (or regression) and a sum of
squares due to residual (or error).

64

= +
Now, if null hypothesis : = = = = is true, then / is distributed as
, where the number of degree of freedom for are equal to the number of regressor
variable in the model.
Computational formula for error sum of square

=(

y i )2

= , '
The regression sum of square
= '

And the total sum of square is

Table 3.2 Shows Analysis of Variance (ANOVA) for significance of Regression in Multiple
Regression

Source of

Degree of freedom

Sum of square

Mean square

variation

(df)

(SS)

(ms)

Due to regression

SSR

MSR = SSR/k

(error)

Nk-1

SSE

MSE = SSE/(n k 1)

Total

N-1

SST

Due to residual

65

FO

MSR/MSE

Apply F- test to test the adequacy of fit as below

The estimated regression equation fits the data adequately if P<0.05 at 95% confidence
level or if P< 0.99 at 99% confidence level.
The coefficient of multiple determinations R2
R squared: A measure of the amount of reduction in the variability of y obtained by using
regressor variables x1, x2,.., xk in the model.

2 =

=1

Adjusted R squared: A measure of the amount of variation around the mean explained by
the model, adjusted for the number of terms in the model. The adjusted R-squared
decreases as the number of forms in the model increase, if those additional terms do not
add value to the model.

2 =

/( )
1
= 1
(1 2 )
/( 1)

PRESS: The prediction sum of squares (PRESS) provides a useful residual scaling.

=
=1

Pred R- squared: A measure of the amount of variation in new data explained by the model.

= 1

66

The predicted R2 and adj- R2 should be within 0.20 of each other otherwise there may be
problem with either data or the model. In addition to the adequacy test mention above the
validity of developed models checked by drawing scatter diagram which shows the
relationship between the observed and predicted values of the weld bead dimension.

3.4 TESTING FOR LACK OF FIT [73]


A procedure for checking the adequacy of the model is called as lack of fit test of fitted
model. In general, to say that the fitted model is inadequate or it is lacking in fit is to imply
the proposed model does not contain sufficient number of terms. The inadequacy of model
can be due to
1. Factors (other than these in proposed model) that are omitted the proposed model
but which affect response.
2. The omission of high order terms involving the factor in proposed model which are
needed to adequately explain the behaviour of the response.

Suppose that we have n observation such that


, , , . ,

, , , . ,

We can see that there are m distinct levels of x.


The pure error sum of squares

yi

Degree of freedom associated with is

=nm
The sum of square for lack of fit
=

y i y i

=
=

67

Degree of freedom associated with is m p because there are m levels of x, and p


degree of freedom are lost because p parameters must be estimated for the model.
The test statistic for lack of fit is
=

/( )
=
/( )

The present study has utilized the Multiple linear regression analysis to predict model and
find the optimal parameter settings.

68

You might also like