1
1. TAGUCHI ANALYSIS
1.1 Background
The principles of Dr. Genichi Taguchi's Quality Engineering, or Taguchi Meth-
o ds, as they are often called in the United States, have had a dramatic impact on
U.S. industry, particularly since the early 1980's, as more and more companies have
started various programs and training metho ds with the hop e of achieving some of
the successes that various Japanese companies have exp erienced over the last thirty
or more years. Case studies detailing successes at the Ina Seito Tile Co., the Electri-
cal Communication Lab oratory Japan, the Kawasaki Oil Re nery,aswell as seven
other examples are given in Taguchi [130 ]. Many consultants travel throughout the
U.S. presenting seminars and training workshops concerned with the learning and
implementation of Taguchi metho ds, to ols and applications. Taguchi metho ds have
b een used in such situations as VLSI circuit design Welch et al [142], non-destructive
testing McEwan, Belavendram, and Ab ou-Ali [93], rob ot pro cess capability Black
and Jiang [14], thickness of aluminum dep osits on silicon wafers Aceves, Hernandez,
and Murphy [2], injection molded plastic comp onents Hanrahan and Baltus [63],
as well as cruise control vacuum valves, shrinkage of sp eedometer cable, computer
resp onse time, p owers supply circuits, noise reduction in hydraulic rollers, and in re-
ducing the rework rate of diesel injectors Wu [143 ]. More examples may b e found in
Dehned [37]. The American Supplier Institute holds an annual symp osium devoted
solely to Taguchi Metho ds, with the rst held in 1983.
In most quality or quality control situations, the goal is to pro duce output as
uniformly near a target value as p ossible. The reduction of variation is now regarded
2
as a fundamental concept, b oth for those concerned with manufacturing and pro duc-
tion and also for those fo cused on service or task pro cessing. The Taguchi metho d is
aimed at the manufacturing situation.
To understand the workings of Taguchi metho ds one must consider three p ossible
situations concerning a qualitycharacteristic of interest. These three are: larger val-
ues are b etter, smaller values are b etter, target value is b est. A qualitycharacteristic
is an imp ortant dimension, prop erty or attribute of a manufactured output. The
three cases will b e considered separately.
For the case where a target value is b est, the aim of Taguchi metho ds is twofold:
center some measured qualitycharacteristic around a target value, and minimize the
variation around this target value. The famous Sony-USA compared with Sony-Japan
example illustrates that b eing centered at the target value is not enough, that b eing
correctly centered at the target and b eing within sp eci cation limits is b etter, but the
b est situation is one of b eing centered at the target and having minimal variation
around that target. See Sullivan, [124 ].
The realization that both the average output and the variability of the output
are crucial is a tremendous improvementover the traditional engineering practice
of only considering whether or not a part falls within sp eci cation limits. With
regard to the implementation of not the concept of this idea, Taguchi has exerted
a broad in uence and deserves much praise. Any deviation from the optimal value is
considered bad, the severity of the deviation is usually measured by a quadratic loss
function. The loss to the manufacturer or so ciety is considered to b e prop ortional to
the square of the distance from the optimal value. Mathematically, this loss can b e
2
written: l Y = k Y T , where Y is the observed value and T is the optimal or
target value. The constant, k , can b e determined based on the loss at a xed value of
Y T . This xed value may represent the edge of the sp eci cation limits, or mightbe
apoint at which it is felt that, say, 50 of the customers would complain or require
warrantywork. The squared error loss is motivated bya Taylor series expansion. See
Taguchi, [129] and [130 ]. Minimizing this loss or exp ected loss is consistent with
3
the two goals of the Taguchi Metho d. The squared error loss function is a continuous
function, and is strictly increasing in jY T j, as opp osed to the discrete pass/fail
nature of sp eci cation limits. This loss function is also e ectiveinconvincing p eople
to worry ab out variation even when the pro cess is on target and mayeven b e within
the desired sp eci cation limits. In particular, if the loss is expressed in dollar units,
those at the managementlevel will now b e more concerned ab out reducing variation.
Consider next the situation where smaller values are b etter, but the resp onse or
qualitycharacteristic of interest is non-negative. In this situation we are concerned
2
with the same loss, k Y T , except nowwe set the target value to zero and hop e
2
to minimize E Y , which represents the mean squared error, or mse, around zero.
The nal situation is that in which the resp onse is again non-negative, but here
2
larger values are considered to b e b etter. The goal is to maximize E Y . Here the
2
Taguchi practitioner considers E 1=Y and again tries to minimize this quantity.
The quantities to b e minimized stem from the desire to measure and reduce the
variation relative to the target value. However, the Taguchi metho d do es not base its
analysis directly on the ab ove quantities; the exact quantity to b e optimized for each
of these three situations will b e given later. In using the recipro cal squared as the
quantityofinterest, the larger is b etter situation now b ecomes that of smaller values
b eing b etter, and we pro ceed as in the smaller is b etter situation.
The rst step in implementing the Taguchi Metho d is that of parameter design.
Parameters are various inputs susp ected or known to have an e ect on the quality
characteristic, and might include such things as rate of ow, temp erature, humidity
or reaction time. Parameter design is an investigation to determine whichvariables
or parameters have an e ect on the qualitycharacteristic. This attempts to improve
on quality while still at the design stage, as opp osed to intensive sampling plans that
insp ect their way to high qualityby `catching' p o or output. Designing high quality
and lowvariabilityinto the pro duct will not only b e muchcheap er than extensiveor
exhaustive insp ection, but also will allow a pro duct to go from blueprint to usable
pro duct in much less time. Variables susp ected to have some e ect on the quality
4
characteristic of interest are known as factors. At this p ointany exp ert knowledge of
the sub ject matter should b e applied in the selection of variables to include as factors.
Those variables susp ected to have the largest e ects on the qualitycharacteristic are
generally considered rst but it is very p ossible to unknowingly omit variables that
should have b een considered and also to include others that have no e ect on the
qualitycharacteristic. Once chosen, these variables are then classi ed into one of
two categories: control variables or noise variables. Control variables represent those
parameters or factors which can b e controlled by the designer or manufacturer. In
contrast, noise variables are those parameters whichmayhave an e ect on the quality
characteristic but generally cannot b e controlled by the designer. Variables that are
very exp ensiveorvery dicult to control may also b e considered noise variables.
In order to screen these variables for their e ect on the qualitycharacteristic, a
classical Taguchi designed exp erimenttakes place. Various settings are selected for
eachofthecontrol variables and various settings are temporarily xed for eachof
the noise variables. The controllable factors are placed into what is known as the
inner array. The inner array is somewhat similar to the classical `design matrix'.
Eachrow of this inner array represents one particular combination of settings for the
control factors, or one `design p oint'. Thus an inner array with eightrows represents
eight p ossible con gurations of the control factors. The noise factors are placed into
what is termed the outer array. The settings for the outer array should representa
range of p ossible noise conditions that will b e encountered in everyday pro duction,
although in practice it is very common to consider only two or three levels for each
of the control factors. Again, these settings are xed solely for the purp ose of the
exp eriment, in general they will b e random. If the noise is assumed to have a linear
e ect, the two settings would generally b e , when the noise factor is
noise noise
susp ected or known to have a quadratic e ect on the qualitycharacteristic, the
p p
3=2; ; + 3=2 .\Thesechoices three settings would generally b e
noise noise noise
are apparently based on the assumption that the noise factors have approximately
symmetric distributions" Kackar, [78]. The twoarrays are then crossed, which
5
means that every exp erimental condition called for by the inner array of control factors
takes place at every condition called for by the outer array of noise factors. In this way
all of the p oints in the inner array are exp osed to the same values of the noise factors
in the outer array; it is hop ed that the temp orarily xed settings of the noise factors at
the exp erimentation level will approximate the random conditions that will o ccur at
the pro duction level. Data measurements on the qualitycharacteristic of interest are
then collected. In some cases the qualitycharacteristic will b e obvious: nal weightof
an exp ensive pro duct sold byweight, diameter of a cylinder, length of a shaft, shipping
weightofabox of cereal, numb er of defectivewelds on an assembly, time to complete
a crucial task, purity of a nal pro duct, numb er of blemishes on a car do or, number of
cycles until failure of a key mechanical part, or the amount of stress b efore failure of a
cable or attaching mechanism. The question of what to measure is b est answered by
a p erson with exp ert knowledge in the area, not by an exp ert in Taguchi metho ds or
in statistical analysis. Particularly in the engineering sciences, anyknowledge of the
underlying physical structure should b e used as an aid in cho osing whichvariables to
include in the design, and also in determining what qualitycharacteristic should b e
measured. This sub ject matter knowledge may also servetopoint out what ranges
of these variables should b e considered, ranges that mayormay not coincide with
the ranges of the `knobs' that can b e turned. Moreover, sub ject matter knowledge
mayserve to p oint out that the control variables or `knobs' available representsome
transformation of the underlying variables that drive the pro cess or reaction. An
example of this would b e where the qualitycharacteristic dep ends critically on the
density of a mixture, but the `knob' available to the exp erimenter is blending sp eed.
In general, for an inner arraywithm settings of the control factors and n con g-
urations of the noise variables, the Taguchi metho d requires mn observations, even
more if there are replications. The two arrays are most often balanced, which means
that for any column of either array, an equal numb er of observations are taken for
each setting that is given in the array.Furthermore, for any pair of columns from the
same array, anycombinations of factors that do o ccur will o ccur the same number of
6
times. Thus if the setting where factor A is at level one and factor B is at level three
o ccurs four times in the inner array, then any combination of factors A and B that
do es o ccur must o ccur exactly four times in the inner array. The arrays are generally
0
is zero if i 6= j ,wherex x also orthogonal, which means that the vector pro duct x
i j
i
~ ~ ~
and x represent columns from the same array, suitably scaled, centered at zero. This
j
~
balance in the arrays allows the exp erimenter to determine the e ects of individual
factors byaveraging over the levels of the other factors. The data collected consist
of measurements of some qualitycharacteristic, collected at each setting dictated by
the two crossed arrays. This characteristic might b e, for example, p ercent purity,
yield, or length of some critical dimension. Controllable factors to b e placed into the
inner arraymight include engine sp eed, amount of a certain chemical, thickness of
a metal part, amount of adhesive applied, voltage through a resistor, gas ow, mix-
ing sp eed, spring tension, initial load etc. These representvariables that can easily
and e ectively b e controlled by the exp erimenterormanufacturer. Noise factors in
the outer array could include ambienthumidity in a large factory, a particular lot
or batchofa key ingredient, cleanliness of a part, the supplier of a key input, or
amount of time since last lubrication. The noise variables and control variables to b e
included in any exp eriment tend to b e particular to the situation b eing considered;
those most familiar with the output of interest will often have a general idea of what
forces, controllable or otherwise, have an in uence on the output of interest.
The inner array is sometimes referred to as the design space, while the outer
array is sometimes referred to as the environment space. The settings in eachof
the arrays generally represent a factorial design or a fractional factorial design. A
factorial design is one where every value or level of each factor o ccurs with all levels
of every other factor. Consider an inner array with three factors. If one factor has
twolevels, one has three levels and one has four, a factorial design for this array
would require 2 3 4 = 24 observations, more if replications are desired. This inner
arraywould then b e crossed with the outer array, requiring even more observations. A fractional factorial design is a design that considers only some carefully selected
7
Table 1.1 Taguchi Crossed Arrays, 32 observations
z =1 z =1 z =1 z =2 z =2 z =1 z =2 z =2
1 2 1 2 1 2 1 2
x 11112222 1 1 11 22 2 2 11112222 11112222
1
x 11221122 1 1 22 11 2 2 11221122 11221122
2
x 12121212 1 2 12 12 1 2 12121212 12121212
3
1; 2;::: 9; 10;::: 17; 18;::: 25; 26;:::
fraction or p ortion of a factorial design. Typically two or three settings are considered
for eachofthevariables, and a fractional factorial design would reduce the number
of observations required bypowers of 1/2 or p owers of 1/3, by running, say, only
one half or one ninth of the p ossible design p oints required for the factorial design.
The design p oints actually chosen for the fractional factorial design usually, under
certain conditions, maintain the orthogonality and balance of the full design. Such
designs can often provide sucient information while saving time and money.For
example, imagine ve design variables, eachwithtwo levels. These could b e placed
1
5 5 2
fraction of a 2 exp eriment. The into a 2 fractional factorial design, whichisa
4
full factorial design calls for 2 2 2 2 2 = 32 observations, while this quarter
3
fraction calls for only eight. The outer array mightthenbea2 design, whichisan
exp erimenthaving three noise factors, eachattwolevels. The two arrays, inner
and outer, would then still b e crossed, requiring 88=64 runs. If constraints dictated
that only 32 runs were p ossible, the exp erimenter mightchose to run only one half
3 3 1
of the 2 outer array,2 , giving a 50 savings in time and money.To illustrate
the crossing of the two arrays, see table 1.1, where the following layout is presented.
3 2
The inner arrayofthecontrol factors, x ;x;x,isa2 design, crossed with a 2
1 2 3
outer array of noise factors, z and z , and would require 32 = 84 observations. In
1 2
3 2
the Taguchi notation, this would b e describ ed as a L 2 crossed with a L 2 . This
8 4
mayormay not b e a layout recommended byTaguchi, it is included only to illustrate
the crossing of the inner and outer arrays.
8
The Taguchi metho d includes the noise factors in the exp eriment for the purp ose
of identifying control factors settings which are robust against noise, i.e. those settings
of the design factors which pro duce the smallest variation in the resp onse across the
di erentlevels of the noise factors. Some combinations of control factor settings may
yield output that is a ected by the noise factors, thus causing the resp onse to vary
around its mean, while for others combinations the output is insensitivetothechanges
in the noise factors. Phadke [111 ]. Similarly, the qualitycharacteristic might exhibit
more variability at, say, the low setting of a certain control factor than it do es at the
high setting. For other control factors, the variation in the qualitycharacteristic might
b e nearly constant across the levels of those control factors, while the average quality
characteristic mightormight not change across the levels of these control factors. For
any control factor, there are four p ossible situations { e ect or no e ect on the mean,
coupled with e ect or no e ect on the variation. If the exp erimenter is fortunate,
some factors may arise that impact the mean but not the variation. Factors with this
typ e of e ect have b een termed \p erformance measures indep endent of adjustment"
or PerMIA - L eon, Sho emaker, and Kackar [88]. Such `adjustment' factors maybe
used to move a pro cess onto or towards its target value, without a ecting the variation
around the target. In a two-step adjustment pro cess, the variation is minimized by
the appropriate settings of the factors that a ect the variation, and then the output
is centered at the target value by the appropriate settings of the factors that only
in uence the mean. Factors which app ear to have little or no impact on either the
mean or the variation are typically set to the level representing the lowest cost.
In order to achieve its goals, the Taguchi metho d analyses not the measured
resp onse but rather some transformation of that resp onse, dep ending on the situation.
As mentioned previously,theTaguchi metho d aims to minimize the exp ected loss,
2
E Y T , but it do es not base the analysis on this quantity directly. A signal to
noise ratio is calculated, the choice of the particular signal to noise ratio dep ends
on the desired outcome of the resp onse: target or nominal is b est, smaller is b etter,
larger is b etter. The last two are considered equivalentby taking recipro cals. For
9
all three situations, the analysis falls under Taguchi's idea of Signal to Noise ratio,
although only in the target is b est case are b oth signal and noise considered. Taguchi
is said to have coined the term `signal to noise' in order that it app eal to the electrical
engineers with whichhewas working during the 1950s - Barker [10].
For the target is b est situation, deviations from target in either direction must b e
2
considered. In order to minimize the E Y T the exp erimentermust b e concerned
with b oth variance and bias. In this situation Taguchi suggests that the following
2 2
y =s , which is to b e maximized. This signal to noise ratio b e considered: 10 log
10
ratio is calculated for eachlevel of eachcontrol factor, averaging over the various
noise levels, and over the other control factors. This ratio is also calculated for each
`design p oint', averaging over the settings of the noise factors in the inner array.It
y=s,which should b e noted that maximizing this ratio is equivalent to maximizing
is equivalent to minimizing the sample co ecientofvariation, s= y . In calculating
this ratio, the Taguchi practitioner is simply p erforming a data transformation. For
P
1
2
the larger is b etter situation the signal to noise ratio is 10 log 1=y , for the
i
10
n
P
1
2
smaller is b etter situation the signal to noise ratio is 10 log y : In all cases,
10
i
n
the goal is to maximize the signal to noise ratio.
By analyzing the signal to noise ratio that is recommended for a particular sit-
uation, the Taguchi practitioner hop es to nd which levels of design variables were
insensitivetothevariation in the noise variables, i.e. which setting in the inner array
gave the smallest variation in resp onse across the outer array. By constructing the
same signal to noise ratio for each level of eachcontrol factor, across all other factors,
the researcher may also discover that some factors a ect the mean while others have
an e ect on the variation. Such factors are known as lo cation e ects and disp ersion
e ects, resp ectively Box and Meyer [19].
As stated previously,theTaguchi metho d calls for this signal to noise ratio to b e
calculated for the data coming from each exp erimental setup of control factors. Thus
this ratio is calculated for eachrow of the inner array. Based on the calculated signal
to noise ratios, with higher b eing b etter, the `b est' levels of each of the control factors
10
are assigned. For `lo cation e ects', the signal to noise ratio allows the researcher to see
the magnitude of the e ect on the qualitycharacteristic, measured across the levels of
the noise factors. If the e ect is small the exp erimentermaycho ose the control factor
setting that represents the lowest cost. Mean plots may also b e calculated for each
of the levels of the control factors; the extenttowhich these agree with the signal
to noise ratio plots will dep end on the nature of the dep endence b etween and .
This graphical technique often accompanies the more classical technique of Analysis
of Variance ANOVA. Since the Taguchi design crosses the twoarrays, it is p ossible
to conduct an ANOVA on b oth the mean and the signal to noise ratio. Under the
Taguchi system it is quite typical to verify the `optimal' results with exp erimental
follow-up.
Another approach whichmay b e imp ossible in some situations is to rst mini-
mize the variation bypicking the b est levels of factors that a ect only the variation,
and the bring the pro cess `on target' by selecting the b est levels of other factors
that in uence only the mean. The hop e here is that the appropriate signal to noise
ratio can b e minimized in a two stage fashion. If the pro cess target changes, the
exp erimenter then only has to worry ab out one set of factors, since the other factors
have b een selected to minimize variation indep endently of the mean. In many ap-
plications of Taguchi metho ds, this partitioning of variables into disjoint categories
based on whether they a ect the mean or the variance is an assumption made b efore
the data are collected. For a multiplicative error mo del, whichsays that the error
increases in prop ortion to the mean, the maximization of the signal to noise ratio
2
and the minimization of E Y T are equivalent. The Taguchi idea is to use a
two-stage minimization pro cedure, rst maximizing the signal to noise ratio over the
factors that e ect the variation, then adjusting the pro cess to target by selecting the
appropriate levels of those factors that e ect the mean but not the variation. Under
2 2
the assumption of a multiplicative error, wehave E Y = ; V ar Y = ,thus
2 2 2 2
the mse is T + , and is minimized over at = T=1 + . The signal
2 2
to noise ratio is 10 log E Y =V ar Y , which can b e written 10 log 1= ,and
10 10
11
is to b e maximized. Thus the variance could rst b e minimizedby maximizing the
signal to noise ratio, and would haveaminimum value, say . The other factors
min
2
could b e used to set the resp onse at = T=1 + . Therefore, maximizing the
min
min
signal to noise ratio will b e equivalent to minimizing the mse in a two step fashion.
More information on the details of the two-stage minimization, the assumptions re-
quired and a comparison of signal to noise ratio against mse for various mo dels are
given in L eon, Sho emaker, and Kackar [88 ], as well as Box [21 ]. Note that in an
2 2 2
additivemodel,E Y = + ,andthus the signal to noise ratio for the smaller is
P
1
2
y , clearly mixes together the e ect on the mean and b etter situation, 10 log
10
i
n
the e ect on the variance.
A third approach Vining and Myers [139] is to set the mean at the appropri-
ate level while at the same time minimizing the variance within the framework of
a Resp onse Surface mo del. Details concerning resp onse surface metho ds rsm and
mo dels will b e given in the following chapters. Within this context, the exp erimenter
is concerned with b oth lo cation and spread, and the mathematical mo deling that
takes places amounts to constrained optimization within the framework of a general
linear mo del. For example, in the target is b est situation this metho d calls for the
2
exp erimenter to minimize sub ject to the constraint = T: For the other two cases
more extreme is b etter the exp erimenter places the constraints on and optimizes
sub ject to the various restrictions. Lagrange multipliers are then used to determine
the optimal levels of the input factors. Quite often, however, an improvementinmean
resp onse will b e accompanied by increased variation, and thus the nal choice of de-
sign settings will often represent a tradeo b etween minimum variance and minimum
bias. Contour graphs for b oth the mean and the standard deviation are very useful
in determining what settings represent the b est choice. The approachgiven by these
authors amounts to mo deling b oth the mean and the variance but do es not suggest
how noise variables could b e incorp orated into the analysis.
12
1.2 Potential Flaws
Within the statistical community the acceptance and implementation of so called
`Taguchi-Metho ds' has not matched the optimism of industry.Many statisticians see
a lot of go o d things within the Taguchi framework but p oint out many shortcomings
as well. For a discussion involving b oth scho ols of thought, see Barker et al[9], as
well as Pignatiello and Ramb erg [112 ].
Although fractional factorial designs can provide a great deal of information with
small to mo derate sample sizes, the crossing of the twoarrays in the Taguchi design
often requires that a large numb er of observations b e taken, with more observations
almost always meaning more time and greater cost. This may b e particularly bad if
the exp eriment is of the `screening' typ e, where a large numb er of factors, b oth noise
and control, are b eing considered. The large numb er of runs called for by the crossed
inner and outer arrays is one of the frequent criticisms of the Taguchi metho d.
Furthermore, there are those in the statistical community who view this crossing
of the arrays as inecient in terms of information loss of information. Because the
two arrays are crossed, the overall resolution of the design may not b e as high as it
could b e with a di erent design. In many cases, this crossing results in the inabilityto
handle or estimate signi cantinteractions. Some suggest combining the twotyp es of
factors into a single combined array, i.e. consider a single fractional factorial design
with sucient resolution to estimate the main e ects as well as the interactions of
interest, including control-control interactions as well as control-noise interactions. In
considering a single array, it will often b e p ossible to nd a higher resolution design
that requires the same numb er of observations as the Taguchi crossed arrays, or to
achieve the same design resolution with less exp erimental runs. If certain interac-
tions are susp ected to exist, this can b e considered at the design stage and a design
with an appropriate aliasing structure that allows for the estimation of the desired
interactions can b e found. The fact that Taguchi designs do not handle interactions
easily is another ma jor shortcoming. Taguchi designs do allowforinteractions if it
13
is known whichinteractions exist before the data are collected; for unknown inter-
actions b etween control factors the Taguchi metho dology falls short. The Taguchi
metho d do es advo cate that follow-up exp erimentation take place, whichseemstobe
a go o d pro cedure but it may somewhat awed not only due to the inadequacy of the
signal to noise ratio but also to due to the p o or prop erties of the original design. See
Lorenzen in Nair, [103], pages 138-139 where the follow-up design had the same
numb er of runs but higher resolution. In this case the `b est p oint' was con rmed
but an improvementof30was also present. Most factorial or fractional factorial
designs can b e augmented by adding center p oints and/or axial p oints. Such designs
are commonly used within the context of resp onse surface mo deling and exp erimenta-
tion. These and similar designs of this form are the Central Comp osite Design CCD,
the Box-Behnken three level designs, as well as Plackett-Burnam designs. Boxand
Drap er [20], and Myers, Khuri, and Carter [97]. Incorp orating these ideas within
the context of resp onse surface metho dology allows for sequential investigation, a fea-
ture that, as is often noted, is not generally available with the Taguchi metho dology,
except in the sense of follow-up con rmation.
2
It must b e rememb ered that the goal is to nd settings that minimize k Y T
while the tactic is to pick settings that minimize the signal to noise ratio. That
the two are equivalent relies directly on the multiplicative error assumption and the
further assumption that the parameters can b e partitioned into distinct and disjoint
subsets, one containing `adjustment' factors which are used to adjust to the optimal
value, one containing `disp ersion' factors whichhave an e ect on the the variation
but not the mean, and one containing factors that in uence neither the mean nor the
variation. However,inanadditive error mo del, where the variance is unrelated to
the mean, the minimizations of the signal to noise ratio and the mse lead to di erent
`b est' p oints. That the set of parameters can b e partitioned as required is based on
the a priori assumption that the signal to noise ratio is a function that dep ends only
on some subset of the parameters. The question of when and whether the two are
equivalentiscovered in L eon, Sho emaker, and Kackar [88]. Furthermore, many
14
di erent data p oints representing drastically di erent means and variances can give
rise to the same signal to noise ratio see Box [21].
Within the context of a statistical analysis, data transformations are usually p er-
formed in order to stabilize the variance of the resp onse. Therefore, in any particular
situation, the actual transformation should b e stem from a physical or data based un-
derstanding of the relation b etween the mean and the variance, see Box and Jones
[22 ]. The recipro cal transformation or recipro cal squared may b e highly non-linear
and is taken indep endent of the actual data values. The e ectiveness of any trans-
formation clearly is related to the actual data; the e ect of the recipro cal squared
transformation for y between .3 and 2 is very di erent than the same transformation
for y between 20 and 25. The e ect of a transformation is well known to b e related
to the magnitude of the sample quantity, y =y . Another criticism of the signal
max min
to noise ratio and the issue of data transformation is that b oth the presence and
magnitude of e ects are related to the metric chosen by the exp erimenter. For this
reason many p eople feel that if a transformation is to b e done, it should b e one sug-
gested by the data, not one chosen b eforehand. Lamb da plots Box [21 ] are helpful
in determining an appropriate transformation, the Taguchi signal to noise ratio is
just one transformation. It should b e noted that this transformation is known to b e
helpful when is prop ortional to Bartlett [12].
To summarize, the ma jor complaints against the Taguchi Metho d are large and
exp ensive designs, the p o or abilitytoallowforinteractions, the lack of sequential
investigation, the a priori assumption of a multiplicative error mo del, and the inade-
quacy of the signal to noise ratio. New ideas that fall mostly within the framework of
RSM but could apply some of the Taguchi metho ds include multiple resp onse analy-
sis, sequential design/analysis, the realization that system variabilitymay itself vary
within the design region, and non-parametric design Myers, Khuri, and Carter [97].
15
2. GENERAL ALTERNATIVES TO TAGUCHI
Many of those who most strongly voice criticismsoftheTaguchi metho d favor
what is known as Resp onse Surface Metho dology, or RSM. Resp onse surface metho ds
and metho dologies b egan in 1951, with the work of Box and Wilson Myers, Khuri,
and Carter [97 ]. In this review pap er, the authors prop ose one de nition for RSM as
\acol lection of tools in design or data analysis that enhance the exploration of a region
of design variables in one or moreresponses ". The interest in exploring the resp onse
throughout a region is very much di erent than the Taguchi metho d of `picking the
b est' from what was observed. Applications of RSM in biological sciences, fo o d
science, so cial sciences, physical sciences, engineering and other general industrial
situations are also o ered in this pap er. The underlying premise of RSM is that
some resp onse, typically denoted , is a function of some set of design variables, say
x ;x ; :::; x . Although the exact functional relationship b etween and x is generally
1 2 k
~
unknown, the assumption is made that this exact relation can b e approximated, at
0
least lo cally,by a p olynomial function of the x s. Actual data collected, typically
denoted y , is expressed as y = + ",Box and Drap er [20 ]. In most cases the
approximating function will b e a rst or second degree p olynomial. The two most
common mo dels are thus:
= + x + ::: + x 2.1
0 1 1 k k
and
k k
X X X
2
= + x + x + x x 2.2
0 i i ii i ij i j
i=1 i=1 i 0 0 0 B x, + x and = + x In vector notation these can b e written as = + x 0 0 ~ ~ ~ ~ ~ ~ ~ resp ectively. Here, x =x ;x ; :::; x isak -vector of input variables, while , , 1 2 k 0 ~ ~ 16 and B represent a scalar parameter, a k -dimensional parameter vector, and a k by k ~ symmetric matrix of parameters, resp ectively. Some assumptions are required on the 2 error asso ciated with the resp onse, assuming that this error followaN 0; distri- 2 bution for some allows the researcher to employmany of the to ols that accompany classical least squares theory. Inherent to the RSM framework is the idea of sequential investigation. This concept is not a part of the Taguchi metho dology other than in follow-up or con rmation of the `b est' design p oint. Sequential investigation lo oks for a p ossible improvement in the qualitycharacteristic. This improvementmay require settings of factors that were not part of the original exp eriment. By tting a mo del, it b ecomes p ossible to estimate the resp onse and p ossibly the variation throughout the entire design region, suggesting areas where the resp onse may b e closer to target, or other areas where the variation is much less, or is more stable. In contrast to sequential investigation, the Taguchi metho d lo oks at optimization, generally consid- ering only design p oints at which data were actually collected. Remember that the Taguchi inner array is often highly fractionated, which means that many `p otential' design p oints were not included in the inner array. In using resp onse surface metho ds the researcher conducts a designed exp eriment, estimates the necessary parameters to t a mo del, checks the adequacy of the mo del lack of t test, determines or guesses where improvement will or mighttake place, and then conducts another exp eriment with design p oints concentrated around the area of susp ected improvement. This pro- cess could b e rep eated as often as needed but careful planning and exp erimentation will keep the numb er of exp eriments and observations to a minimum. This parallels the idea of trying to understand the pro cess, a ma jor tenet of those in quality circles. In most resp onse surface situations, all of the `x'variables in the mo del can gener- ally b e placed at arbitrary levels and are not considered random. To compare directly with the Taguchi metho dology, the resp onse surface mo del needs to b e expanded to consider noise variables as well as p otential interactions b etween noise variables and control variables. In many cases it will b e relatively easy to observe and record the noise variables. In such situations it may also b e p ossible to place a non-sub jective 17 prior distribution on the noise variables. Based on an initial exp eriment, the qual- itycharacteristic could b e mo deled as a resp onse of both the noise variables and the control variables, typically a p olynomial of rst or second degree. Sp eci cally,ifwe let Y denote some qualitycharacteristic of interest, we can t a general mo del of the form: 0 0 0 0 0 Y = Y = + x + x B x + z + z C z + z D x + " 2.3 xz 0 ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ Box and Jones [22 ]. In this mo del, x represents a k -vector of design variables, z ~ ~ represents a p-dimensional random vector of noise or environmental variables, and " 2 represents error, usually one assumes " N 0; . The parameters , , B , , C , 0 ~ ~ ~ ~ 2 D , and are generally unknown. ~ Within the context of sequential investigation, the researcher nowwould liketo know which design p oints should b e considered for the next exp eriment, making use of the current tted mo del, realizing that the tted mo del is based on parameter estimates, not parameters, and after integrating over the p-dimensional prior on z , ~ the vector of noise variables. Many criterion are available for deciding what makes a new design p oint or a new design matrix optimal. These include D-optimality, A-optimality, minimum bias, minimum variance of prediction, etc. A very common goal is to minimize exp ected mean square relative to a desired or sp eci ed target value. When the concern is for what will happ en in the future, the mean square er- ror of prediction b ecomes a ma jor concern Allen [4],Dwivedi and Srivastava [41]. Noise factors may b e regarded as random regressors, thus changing the interpretation of the regression mo del slightly, see Narula [101 ], Reilman, Gunst, and Lakshmi- narayanan [115 ], Sha er [122], Walls and Weeks [140 ], and Kackar and Harville [77 ] for various viewp oints concerning random regressors. In some instances, the re- sp onse surface may dep end on b oth quantitative and qualitative factors, see Drap er and John [40]. Various approaches to the dual-resp onse problem are considered in Castillo and Montgomery [29 ], Vining and Myers [139 ] and Gan [51]. 18 3. RESPONSE SURFACE ALTERNATIVES Avery common statistical measure of p erformance is squared error loss, or ex- p ected squared error loss. This concept has b een used by statisticians for manyyears but the increased use of Taguchi metho ds has made many p eople, particularly en- gineers and others in industrial situations, realize that any deviation from target is detrimental, and that considering only whether or not a part is within sp eci cation limits can lead to a false sense of security ab out a pro cess. Insp ection plans concerned only ab out sp eci cation limits without considering variation around a target value can also lead to distributions that are almost uniformly spread across the sp eci ca- tion limits, as opp osed to b eing concentrated at the target value. The previously mentioned example of Sony-USA compared to Sony-Japan is a classic example of this phenomenon. Sullivan, [124 ]. Within the framework of a Taguchi design the quan- 2 titytominimize would b e E Y T , where Y is the characteristic of interest and T represents the target value. Inner and outer arrays would b e crossed and signal to noise ratios would b e calculated to determine which settings of the design variables would minimize squared error loss. Recall that although the goal is to minimize the mse,theTaguchi criterion is actually some signal to noise ratio. Here a di erent formulation will b e considered in an e ort to improve on standard industry practice. Let Y represent the qualitycharacteristic of interest and recall the general resp onse surface mo del 0 0 0 0 0 D x + " 3.1 C z + z + z B x + z + x Y = Y = + x xz 0 ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ Box and Jones [22]. In this mo del, x represents a k -dimensional vector of design ~ variables, z represents a p-dimensional random vector of noise variables, and " repre- ~ 2 sents error, usually it is assumed that " N 0; . The parameters , , B , , C , 0 ~ ~ ~ ~ 19 2 2 D , and are generally unknown. This mo del allows for pure quadratic e ects x or i ~ 2 z ,aswell as second order terms suchas x x ;xz ; and z z . i j i j i j j If we consider the noise variables as b eing part of an environment space, wecould rst integrate over this environment space and then attempt to minimize the mean squared error. Sp eci cally, the squared error loss would b e: Z Z Z 2 Y T f z f y jz dy jz dz 3.2 z Y jZ Y jZ Z Z ~ ~ ~ ~ p 1 ~ ~ ~ 2 ;A and, indep endentof z ;" N 0; , while all Here it assumed that z N e p ~ ~ ~ ~ ,is other terms are constants, p ossibly unknown. It is assumed that the mean, e ~ known and that the covariance matrix, A,isknown and is p ositive de nite. ~ In the most general case, namely 0 0 0 0 0 Y = + x + x B x + z + z C z + z D x + " xz 0 ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ 0 0 0 B x+z + z C z + z D x + " ~ ~ ~ ~ ~ ~ ~ ~ ~ wehave 0 0 0 0 C + D x+ + tr AC 3.3 E Y = + x + x B x + e 0 e e ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ 0 0 0 0 2 C z AD x + Varz D A + x VarY = + ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ 0 0 0 0 0 0 C z ] 3.4 ;z C z +cov z D x;z D x+cov z ;z + 2[cov z ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ where 0 0 C AC C z =2tr C AC A+4 Varz e e ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ 0 0 0 0 0 0 0 0 AC D C z =2x D x;z A ; cov z D D x=x ;z cov z e ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ 0 0 0 cov z ;z C z =2 AC e ~ ~ ~ ~ ~ ~ ~ ~ ~ thus 2 0 0 0 0 0 + + Varz onl y 3.5 VarY =x D AD x +2x D A +4x D AC e ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ 2 See Seb er, p4112, p40 4, [121 ] where is the variance in the resp onse and Varz onl y represents variation due to the noise variables but indep endentofthe ~ 20 value of x. These pieces may b e ignored when determining the values of x that ~ ~ minimize the mse: Therefore, ignoring some terms indep endentofx ~ 2 0 0 0 0 0 E Y T = x D AD x +2x D A +4x D AC e ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ 2 0 0 0 0 0 0 + + x + x B x + x D C + + + tr AC T 0 e e e e ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ 2 the rst term is VarY , the second is the squared bias, E Y T . Considered as function of x,thismay b e written as ~ 0 0 0 0 0 0 0 AD + + D + D +2B + + C + tr AC T ]x x [D e e 0 e e e ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ 0 0 0 0 0 + tr AC T ] C + + + + D +2x [D A +2C e e 0 e e e ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ 0 2 0 0 0 0 +x B x +2x B xx + D e ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ This is fourth degree function of x, but all of the third and fourth order terms will ~ vanish if the mo del contains no second order terms in x, i.e. when the B x term is ~ ~ 0 : strictly linear in x,when B x= + x 0 ~ ~ ~ ~ 3.1 Case 1 In the simplest case the mo del is given by: 0 0 0 + " B x + z + x y = + x xz 0 ~ ~ ~ ~ ~ ~ ~ This mo del is quadratic in x and linear or rst order in z ,withnointeraction b etween ~ ~ noise variables and design variables. For this mo del, it can b e shown that, after integrating over z , Y follows a normal distribution, with mean and variance ~ 2 0 0 0 0 ; + A + x + x B x + 0 e ~ ~ ~ ~ ~ ~ ~ ~ ~ ~ In this situation the mean squared error reduces to the following quadratic function of B x: ~ 2 2 0 0 2 2 0