net/publication/23884611
CITATIONS READS
140 1,409
4 authors, including:
Janet Allen
University of Oklahoma
355 PUBLICATIONS 6,092 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
The Goal-oriented Inverse Design (GoID) Method for Systems-based Design Under Uncertainty View project
Investigating the Effectiveness of Machine Learning Paradigms for Supporting Engineering Designers in Rapidly Evolving Digital Manufacturing View project
All content following this page was uploaded by Patrick N Koch on 05 June 2014.
DETC97DTM3881
ABSTRACT element codes can take from minutes to hours, if not longer.
Perhaps the most prevalent use of statistics in engineering What's more, this mode of query-and-response can often lead to
design is through Taguchi's parameter and robust design - using a trial and error approach to design, where a designer may never
orthogonal arrays to compute signal to noise ratios in a process uncover the functional relationship between x and y and
of design improvement. In our view, however, there is an therefore may never identify the best settings for input values.
equally exciting use of statistics in design that could become Statistical techniques are widely used in engineering design
just as prevalent: it is the concept of metamodeling, whereby to address these concerns. The basic approach is to construct
statistical models are built to approximate detailed computer approximations of the analysis codes that are much more effi-
analysis codes. Although computers continue to get faster, our cient to run and that yield insight into the functional
analysis codes always seem to keep pace, so that their computa- relationship between x and y. If the true nature of a computer
tional time remains non-trivial. Through metamodeling, analysis code is represented as
approximations of these codes are built that are orders of magni- y = f(x),
tude cheaper to run. These metamodels can then be linked to
optimization routines for fast analysis, or they can serve as a then a "model of the model" or metamodel (Kleijnen, 1987) of
bridge for integrating codes across different domains. the analysis code is taken to be
In this paper we first review metamodeling techniques that y = g(x), and so y = y + ε
encompass the Design of Experiments, Response Surface
Methodology, Taguchi methods, neural networks, inductive where ε represents both the error of approximation and
learning, and kriging. We discuss their existing applications in measurement or random errors. The most common meta-
engineering design and then address the dangers of applying modeling approach is to apply the Design of Experiments
traditional statistical techniques to approximate deterministic (DOE) to identify an efficient set of computer runs (x 1, x 2, ...,
computer analysis codes. We conclude with recommendations x n) and then employ regression analysis to create a polynomial
for the appropriate use of metamodeling techniques in given approximation of the computer analysis code. These approxi-
situations and how common pitfalls can be avoided. mations then can replace the existing code while providing:
KEYWORDS: Design of Experiments, Response • a better understanding of the relationship between x and y
Surface Methodology, Metamodeling, Kriging, Neural Net- (we typically lose this understanding when running computer
works, Rule Induction, D-Optimal, Taguchi Robust Design codes "blindly"),
• facilitated integration of domain dependent computer codes
(we no longer work with the codes themselves but rather
1 INTRODUCTION simple approximations of them), and
Much of today's engineering analysis work consists of run- • fast analysis tools for optimization and exploration of the
ning complex computer analysis codes -- supplying a vector of design space (we become more efficient because we are using
design variables (inputs) x and receiving a vector of responses the approximations rather than the computationally expensive
(outputs) y. Despite continual advances in computing power codes themselves).
and speed, the expense of running many analysis codes remains However, we have found that many published applications
non-trivial. Single evaluations of aerodynamic or finite- (including our own) utilizing these statistical methods for
Fit Using:
Optimize/Improve
Stepwise
OAs
Paper
inner and outer) and the use of the loss function or signal-to-
(Balabanov, et al., 1996) x x 2nd x x noise ratio. It has been argued and demonstrated that the use of
(Bauer and Krebs, 1996) x x x 2nd x x S/N
(Beard and Sutherland, 1993) x x x L
a single experiment combining control and noise factors, a
(Chen, et al., 1996) x x 2nd x R
more traditional RSM approach rather than inner "control" and
(Chi and Bloebaum, 1995) x x R outer "noise" arrays, is more efficient in that it requires fewer
(Engelund, et al., 1993) x x 2nd x experiments (Shoemaker, et al., 1991; Unal, et al., 1994a;
2nd
(Gadallah and ElMaraghy, 1994) x x x x x S/N Welch, et al., 1990). With respect to the loss function or
(Giunta, et al., 1996) x x 2nd x signal-to-noise ratio, the drawbacks of combining response
(Giunta, et al., 1994) x x x x 2nd x
(Healy, et al., 1975) x x 2nd x x
mean and variance into a single measure, a signal-to-noise ratio,
(Hong, et al., 1994) x x 2nd x x S/N
are well-documented in the literature and many authors advocate
(Koch, et al., 1996) x x 2nd x measuring a response directly rather than loss, and separately
(Korngold and Gabriele, 1995) x 2nd x tracking mean and variance (Chen, et al., 1995; Ramberg, et
2nd
(Li, et al., 1994) x x x S/N al., 1991; Welch, et al., 1990). Shoemaker et al. (1991),
(Mavris, et al., 1996) x x 2nd x R however, warn that "A potential drawback of the response-
(Mavris, et al., 1995) x x 2nd x R
(Roux, et al., 1996) x x Mix x
model approach that is depends more critically than the loss-
(Rowell, et al., 1996) x x x 2nd x x x model approach on how well the model fits."
(Stanley, et al., 1992) x x x R Given the general acceptance of the Taguchi principles and
(Sundaresan, et al., 1993) x x x R of robust design, with these criticisms of these techniques,
2nd
(Unal, et al., 1996) x x x many advocate a combined Taguchi-RSM approach, or simply
(Unal, et al., 1992) x x x R applying more traditional RSM techniques within the
(Unal, et al., 1994a) x x 2nd x
2nd
framework defined by Taguchi (Lucas, 1994; Myers, et al.,
(Unal, et al., 1994b) x x x x x R
(Venter, et al., 1996) x x 4th x 1989; Myers and Montgomery, 1995; Ramberg, et al., 1991).
(Yu and Ishii, 1993) x x 2nd x x R It is the belief of the authors of this paper that both orthogonal,
inner and outer arrays, and single, composite experiments each
have advantages and disadvantages and appropriate uses, and that
In addition to these specific engineering design applications
separate observation of mean and variance leads to useful
of DOE, RSM, and Taguchi, many methods and "tools" have
problem insight. Regardless, the core principles of both
been developed for more general design application. We
Taguchi and RSM have provided a foundation for the specific
examine several of these methods in the next section,
design methods discussed in Section 3.2.2.
highlighting various applications of each.
2
4 GUIDELINES AND RECOMMENDATIONS
L.S.E.=∑ (y i - y i)
2 ∑ (y i - y i) =0
How can a design engineer efficiently apply the
y = g(x) = second order
metamodeling tools of Section 2 while avoiding the pitfalls
least squares fit y = g(x) = spline fit described in Section 3.3? There are two ways to answer this
question: from the bottom up (tools applications) and from
x x
the top down (motives tools). The bottom-up approach is
(a) Non-Deterministic Case (b) Deterministic Case
presented in Section 4.1, the top-down in Section 4.2.
Figure 4. Deterministic and Non-Deterministic
4 . 1 Evaluation of Metamodeling Techniques
Curve Fitting
There are two components to this section. The first is an
evaluation of the four metamodeling techniques described in
So where can these methods go wrong? Unfortunately it is
Section 2.2. The second component is choosing an
very easy to unthinkingly classify the ε bias term from a
experimental design which has more direct applicability to
deterministic model fit as ε random and then proceed with standard
response surface methods but also applies to the remaining
statistical testing. Several authors have reported statistical
metamodeling techniques. The explorations of which designs
measures such as the F-statistics and root MSE for verification
are most appropriate are, as yet, open research areas.
of model adequacy (e.g., Healy, et al., 1975; Koch, et al., 1996;