Professional Documents
Culture Documents
Ubc 2021 May Rodriguez Gilberto
Ubc 2021 May Rodriguez Gilberto
Ubc 2021 May Rodriguez Gilberto
DOCTOR OF PHILOSOPHY
in
(Statistics)
Examining Committee:
ii
Abstract
A wide range of natural phenomena and engineering processes make phys-
ical experimentation hard to apply, or even impossible. To overcome these
issues, we can rely on mathematical models that simulate these systems
via computer experiments. Nonetheless, if the experimenter wants to ex-
plore many runs, complex computer codes can be excessively resource and
time-consuming.
Since the 1980s, Gaussian Stochastic Processes have been used in com-
puter experiments as surrogate models. Their objective can be predicting
outputs at untried input runs, given a model fitted with a training design
coming from the computer code. We can exploit different modelling strate-
gies to improve prediction accuracy, e.g., the regression component or the
correlation function. This thesis makes a comprehensive exploitation of two
additional strategies, which the existing literature has not fully addressed
in computer experiments.
One of these strategies is implementing non-standard correlation struc-
tures in model training and testing. Since the beginning of factorial designs
for physical experiments in the first half of the 20th century, there have been
basic guidelines for modelling from three effect principles: Sparsity, Hered-
ity, and Hierarchy. We explore these principles in a Gaussian Stochastic
Process by suggesting and evaluating novel correlation structures.
Our second strategy focuses on output and input transformations via
Dimensional Analysis. This methodology pays attention to fundamental
physical dimensions when modelling scientific and engineering systems. It
goes back at least a century but has recently caught statisticians’ attention,
particularly in the design of physical experiments. The core idea is to analyze
dimensionless quantities derived from the original variables.
While the non-standard correlation structures depict additive and low-
order interaction effects, applying the three principles above relies on a
proper selection of effects. Similarly, the implementation of Dimensional
Analysis is far from straightforward; choosing the derived quantities is par-
ticularly challenging. Hence, we rely on Functional Analysis of Variance as
a variable selection tool for both strategies. With the “right” variables, the
iii
Abstract
iv
Lay Summary
We use computer experiments to simulate natural phenomena and engi-
neering processes where physical experimentation is too costly or infeasi-
ble. One example is the prediction of hurricane flood hazards as a function
of diverse environmental factors. These experiments rely on mathematical
models via computer codes. Nonetheless, these codes can be excessively
resource-consuming. Thus we could rely on Gaussian Stochastic Processes,
which are computationally faster surrogate statistical models for these codes.
In this thesis, we pay attention to the prediction accuracy of these sur-
rogates using two strategies: novel data correlation structures and Dimen-
sional Analysis. Both have significant challenges in terms of variable selec-
tion, which we address with a statistical tool called Functional Analysis of
Variance. We apply our strategies in several case studies to improve pre-
diction accuracy. Thus we propose new modelling frameworks in computer
experiments to accomplish more accurate surrogate models.
v
Preface
This doctoral dissertation has been written under the supervision of Dr.
William J. Welch. Potential approaches were discussed in weekly research
meetings. I was responsible for conducting the coding for all the case studies,
along with necessary mathematical formulations. Furthermore, I was in
charge of the thesis composition. Subsequent revisions were discussed with
Professor Welch.
Chapters 3 and 4 are based on a paper manuscript already written
on non-standard correlation structures in a Gaussian Stochastic Process.
It details the basis of the principles coming from traditional factorial ex-
periments described in Chapter 1. The study cases are centred on the
Michalewicz, Friedman, and Franke functions. They are the main thread
in this work whose final objective is to improve prediction accuracy in a
Gaussian Stochastic Process, with the help of Functional Analysis of Vari-
ance. Another two additional case studies from Chapter 4 are likely to
be included in this manuscript since they are applications that go beyond
mathematical test functions.
Chapter 5 was initially projected as a single manuscript on the use of
Dimensional Analysis in a Gaussian Stochastic Process, along with Func-
tional Analysis of Variance based on three case studies. However, further
research allows us to split up this chapter in two paper manuscripts (already
in process) as follows:
• The first two case studies in this chapter are an appropriate example
of the use of Dimensional Analysis: Buckingham’s Π-theorem. They
set up a basis for the second manuscript.
vi
Table of Contents
Abstract . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iii
Lay Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . v
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi
List of Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
Glossary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvi
Acknowledgements . . . . . . . . . . . . . . . . . . . . . . . . . . . xix
Dedication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xx
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 The Use of a Factorial Effect Framework . . . . . . . . . . . 3
1.1.1 Historical Background of Experimental Guidelines . . 3
1.1.2 Formal Definitions of the Effect Principles . . . . . . 4
1.2 Dimensional Analysis . . . . . . . . . . . . . . . . . . . . . . 5
1.2.1 Why Dimensionality Should Be Important for Statis-
ticians? . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.2.2 Early Works in Dimensional Analysis . . . . . . . . . 7
1.3 Structure of the Thesis . . . . . . . . . . . . . . . . . . . . . 9
vii
Table of Contents
viii
Table of Contents
5 Dimensional Analysis . . . . . . . . . . . . . . . . . . . . . . . 90
5.1 Previous Applications of Dimensional Analysis . . . . . . . . 93
5.2 Buckingham’s Π-Theorem . . . . . . . . . . . . . . . . . . . . 96
5.3 Borehole Function . . . . . . . . . . . . . . . . . . . . . . . . 98
5.3.1 Dimensional Analysis . . . . . . . . . . . . . . . . . . 100
5.3.2 Simulation Settings . . . . . . . . . . . . . . . . . . . 104
5.3.3 Prediction Results . . . . . . . . . . . . . . . . . . . . 106
5.3.4 Extrapolation . . . . . . . . . . . . . . . . . . . . . . 109
5.4 Heat Transfer in a Solid Sphere . . . . . . . . . . . . . . . . 111
5.4.1 Dimensional Analysis . . . . . . . . . . . . . . . . . . 113
5.4.2 Simulation Settings . . . . . . . . . . . . . . . . . . . 118
5.4.3 Prediction Results . . . . . . . . . . . . . . . . . . . . 120
5.4.4 Extrapolation . . . . . . . . . . . . . . . . . . . . . . 121
5.5 Storm Surge Model . . . . . . . . . . . . . . . . . . . . . . . 123
5.5.1 The Joint Probability Method . . . . . . . . . . . . . 124
5.5.2 Advanced Circulation Model Settings . . . . . . . . . 124
5.5.3 Computation of Additional Inputs . . . . . . . . . . . 125
5.5.4 Distribution of Maximum Storm Surge . . . . . . . . 127
5.5.5 Dimensional Analysis . . . . . . . . . . . . . . . . . . 129
5.5.6 Simulation Settings . . . . . . . . . . . . . . . . . . . 136
5.5.7 Prediction Results . . . . . . . . . . . . . . . . . . . . 138
5.5.8 Extrapolation . . . . . . . . . . . . . . . . . . . . . . 142
5.6 Concluding Remarks . . . . . . . . . . . . . . . . . . . . . . 145
ix
Table of Contents
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 155
Appendices
x
List of Tables
xi
List of Figures
xii
List of Figures
xiii
List of Figures
xiv
List of Figures
xv
Glossary
ABCD Automatic Bayesian Covariance Discovery (p. 153)
E-JE2 Equal Weights on Joint Effects up to All 2-Input Interactions (p. 52)
E-JE3 Equal Weights on Joint Effects up to All 3-Input Interactions (p. 69)
xvi
Glossary
xvii
Glossary
xviii
Acknowledgements
I would like to express my gratitude and appreciation to my supervisor,
Dr. William J. Welch, who provided magnificent support, guidance, encour-
agement, and feedback during all these years of my doctoral research. All
our conversations and meetings were amazingly crucial and fruitful during
the development of this dissertation. I would also like to thank my super-
visory committee, Dr. Marie Auger-Méthé and Dr. James V. Zidek, for
their thoughtful and excellent comments and observations on this research.
Their recommendations were a vital part of this work. Moreover, I would
like to acknowledge Whitney Huang and Taylor Asher for their collaborative
support and guidance on this dissertation’s final case study.
Finally, I would also like to thank all the amazing people I had the
opportunity to work with and learn within the Department of Statistics
at UBC. I express my appreciation to all the faculty members who were
my professors during my program coursework. Also, I am grateful to the
Applied Statistics and Data Science (ASDa) group for all the consulting
projects I had the opportunity to collaborate on. A big thank you to Peggy
Ng, Ali Hauschildt, Andrea Sollberger, Mairead Roche, and Sima Jafari
in the department’s main office for their outstanding support and all-time
effort.
xix
Dedication
First of all, the most significant recognition for my family. I have had the
most amazing parents all over my life, who have struggled all their lives,
and they have raised the most amazing children. Marı́a del Carmen and
Gilberto, all this journey has been for you, and I am so happy to have you
in my heart. My brother and I always try to be the best professionals in
life: Airel as a physician, whereas I as a statistician. I am pretty sure we
have never disappointed you all over this time. And Airel, you also are a
vital part of my heart; never forget that.
Then, all my friends on this journey have been another important part
of it. I am grateful to have you all on my side still. Many of them have been
far away all these years, but we have managed to be close in at least one
text message away. On the other hand, all my local friends and colleagues
have been the best support ever, and I am grateful to have you here. Many
of you are already far away, though, while others are still here next to me
when I need them. It is impossible to list names. Although, you know who
you are, and that is what matters. You are also in my heart.
xx
Chapter 1
Introduction
Physical experiments can be hard to apply in a wide range of natural phe-
nomena and engineering processes. Difficulties arise in cases where physical
experimentation may be too resource-consuming or even impossible. One
example is the prediction of hurricane flood hazard as a function of diverse
environmental factors. To overcome these issues, we can rely on mathemat-
ical models that simulate a given physical system via computer codes.
Bastos and O’Hagan (2009) state that computer codes attempt to ex-
amine complex systems in different scientific fields. Nonetheless, the exper-
imenter can assume that the output is generated by a complex unknown
function subject to a specific input set. Hence, a statistical model can be
used as an emulator to approximate these expensive computer codes. The
specific surrogate model in computer experiments was adopted and adapted
from spatial statistics, namely kriging, where we view the response as a real-
ization of a stochastic process. This model is known as a Gaussian Stochastic
Process (GaSP) introduced by Sacks et al. (1989). It is also known merely
as a Gaussian Process (GP).
A computer experiment’s primary objective, and thus a GaSP emulator,
depends on the specific research context. Prediction can be an objective,
i.e., predicting outputs at untried inputs given a model fitted with a training
design coming from a computer code. We can also use a computer exper-
iment for optimization purposes, where a surface is fitted over a region of
interest to find a minimum or a maximum, as noted by Jones et al. (1998).
We explore the prediction accuracy paradigm in a GaSP model throughout
this thesis.
The assumption of independent random errors between runs (at different
input configurations) is common in traditional design and analysis of phys-
ical experiments. Given a set of inputs including any boundary conditions,
unlike physical experiments, a computer experiment is often deterministic
(McKay et al., 1979; Sacks et al., 1989). Therefore, replicate runs of a com-
puter code are wasteful as they would yield the same output. Nonetheless,
the emulator GaSP has a random component to consider uncertainty at
untried input vectors where the computer code is not run.
1
Chapter 1. Introduction
1. For given input and output variables, the choice of a suitable corre-
lation structure plays a key role in a GaSP. One of the critical ap-
plications of a computer code is the replacement of certain physical
experiments. Hence, we could modify the correlation structure by fol-
lowing the available guidelines for fractional factorial designs in physi-
cal experiments, namely three effect principles (Box and Meyer, 1985;
Hamada and Wu, 1992, 2000). Such alternative correlation structures
2
1.1. The Use of a Factorial Effect Framework
3
1.1. The Use of a Factorial Effect Framework
4
1.2. Dimensional Analysis
5
1.2. Dimensional Analysis
• For the case of statisticians, it is pointed out that “[...] Those statis-
ticians who recognize how deeply rooted their science is in data are
doubtless well aware of the relevance of dimensionality considerations.
I know of no statistical text that makes even the most elementary
points, yet these can be enormously useful in rapid checking of asser-
tions and formulae.”
6
1.2. Dimensional Analysis
factor, will impact the derived ones in the form of a product and/or quotient
of those fundamental dimensions’ corresponding scaling factors.
7
1.2. Dimensional Analysis
gt2
y = y0 + V 0 t − , (1.1)
2
where
8
1.3. Structure of the Thesis
could select two out of these three inputs (V0 and t) as base quantities so
equation (1.1) can become dimensionless as in
y y0 gt
= +1− ;
V0 t V0 t 2V0
which yields one dimensionless output q0 and two dimensionless inputs q1
and q2 :
y y0 gt
q0 = q1 = q2 = .
V0 t V0 t 2V0
Buckingham’s Π-theorem says that the relationship can be represented
using only these three variables. Thus, this new formulation reduces the sys-
tem’s dimensionality with composite output and inputs (i.e., dimensionless
variables based on products and quotients between the original ones). As a
side note, the input g could be considered non-constant in another environ-
ment, and it has to be taken into account in a DA. Note that y0 and t might
be an alternative pair of base quantities given their fundamental dimensions
L and T, respectively. Notwithstanding, DA’s performance in prediction ac-
curacy heavily relies on a given base quantity selection. Chapter 5 explores
this matter in further detail.
9
1.3. Structure of the Thesis
Table 1.1 illustrates the studies along with their respective sections in
each one of the subsequent chapters.
10
1.3. Structure of the Thesis
Friedman
(4.3)
function
Non-standard
GaSP correlation Franke
(4.6)
structures function
(4): Gaussian Processes
with Low-Order
Joint Effects
OTL circuit
(4.9)
function
Nilson-Kuusk
(4.11)
model
Borehole
(5.3)
function
Storm surge
(5.5)
model
11
Chapter 2
D = x(1) , . . . , x(n) ,
y = (y1 , . . . , yn )> ∈ Rn .
Z(x) ∼ N 0, σ 2 ),
12
2.1. The Regression Component
If the mean, variance, and covariances between the input vectors are
taken as constant all over the space; then we have a stationary GaSP. More-
over, let R be the n × n symmetric and positive semidefinite correlation
matrix where each element is R(x, x0 ) for all possible pairs x and x0 in the
training design D. We refer to R(x, x0 ) as the correlation structure of the
GaSP throughout this thesis, and its setup is the central focus in Chapters 3
and 4.
whose k functions fh (x) are known. Note that the regression component (2.3)
can be expressed in the matrix form
is unknown, and
>
f (x) = 1, f1 (x), . . . , fk (x) ∈ Rk+1
13
2.3. Correlation in Random Function
µ(x) = β0 .
On the other hand, besides a single constant term, the cases regarding
the Borehole function and heat transfer in solid sphere in Chapter 5 will
implement a linear regression term as follows:
Therefore, the most general case for this component (involving vector β and
matrix F) will be described throughout subsequent sections in this chapter.
which is function of the distance hj = xj − x0j for the jth input. A key
feature of the standard structure is that it is multiplicative across the d di-
mensions. It is important to mention that alternative correlation structures
will be introduced in Chapters 3 and 4. Another important characteristic is
that the correlation functions Rj (·) we will consider decay with increasing
hj . That implies that output values further apart in the xj direction will
be less correlated and hence differ more. In contrast, two output values at
nearby xj values are highly correlated and hence similar.
14
2.3. Correlation in Random Function
Jones et al. (1998) state that the correlation between a pair of points
cannot always be taken as a distance metric assigning the same weight to
all the d inputs. Thus, the distance is weighted using a hyperparameter
θj on the jth input. A large value of θj indicates a highly active input
variable. Furthermore, in terms of any correlation structure R(·, ·) in our d
dimensions, we have correlations in design D that are high when a pair of
input points is close and low when they are apart. A larger θj increases the
weighted distance between two inputs in dimension j larger, and hence the
correlation function decays faster as a function of distance. Again the lower
correlation between values of the output implies they are less related, i.e.,
the function is more active in dimension j.
We can find in the literature more than one choice available for Rj (hj )
as follows:
• Power Exponential (PowExp) Correlation Function. This func-
tion is expressed as
p
RjPowExp (hj ) = exp(−θj hj j ), (2.5)
where θj ≥ 0 and pj ∈ [1, 2] for j = 1, . . . , d. The parameter pj
is related to the smoothness of the function with respect to the jth
input. A value of pj = 2 corresponds to smooth functions, and less
smoothness is represented by values close or equal to 1 .
• Squared Exponential (SqExp) Correlation Function. It is a
special case of the previous correlation function when pj = 2:
RjSqExp (hj ) = exp(−θj h2j ); (2.6)
with θj ≥ 0 for j = 1, . . . , d. It is also known as the Gaussian correla-
tion function.
The literature offers other options for the correlation function Rj (hj ),
e.g., the Matérn class suggested by Stein (1999) which also controls for
smoothness as the PowExp class. The parametrization of the Matérn class
allows us to consider more than one possible correlation function. Never-
theless, the scope of this thesis is not on the correlation function Rj (hj ) but
on modifications of the multiplicative structure R(·, ·) (2.4).
We restrict our attention to the two cases mentioned above in our sub-
sequent simulation studies: the PowExp and SqExp correlation functions.
We use both correlation functions in Chapters 3 and 4 for the case stud-
ies to compare our non-standard correlation structures versus the standard
one. On the other hand, Chapter 5 is focused on the standard correlation
structure only with the PowExp case.
15
2.4. Estimation
2.4 Estimation
According to Sacks et al. (1989), from the training design D, we assume an
output random vector
Y = (Y1 , . . . , Yn )> ∈ Rn
Y ∼ Nn (Fβ, σ 2 R).
16
2.4. Estimation
and
1
σ̂ 2 (θ) = (y − Fβ̂)> R−1 (y − Fβ̂). (2.9)
n
If we plug in the MLEs (2.8) and (2.9) into (2.7), we obtain the following
profile likelihood function:
2 1 n
L(y | β̂, σ̂ , θ) = exp − . (2.10)
(2πσ̂ 2 )n/2 |R|1/2 2
n 1 n
= − log(2π) + log(σ̂ 2 ) − log |R| − . (2.11)
2 2 2
Both MLEs, β̂ and σ̂ 2 , depend on true vector θ and so does the pro-
file log-likelihood function (2.11); since we need to compute |R|, and σ̂ 2 is
present. Hence, we have to obtain the estimate θ̂ by maximizing Equa-
tion (2.11) numerically. As noted in Section 2.4.2, there is more than one
numerical method and software resource to achieve this goal.
17
2.4. Estimation
2. The second stage implements the L-BFGS method for the best solution
found by multiple simplex searches, described by Nocedal (1980) and
Liu and Nocedal (1989). It is a limited-memory quasi-Newton method
for large scale optimization where storage is critical. The optimization
procedure needs an available analytic expression for the gradient ∇l(θ)
of (2.11), which is also provided by mlegp.
18
2.5. Prediction
2.5 Prediction
As mentioned in Chapter 1, one of the objectives of a computer experiment
is the prediction of a given response. The first major step to achieve this
objective is done through the GaSP fitting that allows us to obtain the
respective MLE estimates, β̂ and σ̂ 2 , as well as the estimated correlation
hyperparameters in vector θ̂. Then, we can proceed with the corresponding
predictions on a testing set H of N further input vectors
(1) (N )
H = x t , . . . , xt ,
which has N unknown outputs
(1) (N )
y xt , . . . , y xt .
contains the n correlations between the test input vector xt , and each one
of the n input vectors in training design D. Here R(·, ·) and r(·) are also
conditional on θ̂, but we suppress the dependence for notational simplicity.
The BLUP (2.12) is the best linear predictor in the sense that it mini-
mizes the mean squared error (MSE) between a further output and a linear
combination of the n training data, subject to unbiased estimation of the
mean of the new output. Similarly to the GaSP model (2.1), predictor
ŷ xt | θ̂ has two parts:
where the correlations in r(xt ) have a certain structure; e.g., RStd (x, x0 ).
19
2.5. Prediction
where ȳ is the mean computed from vector y, the outputs of training set D,
n
1X
ȳ = yi .
n
i=1
The denominator in metric (2.14) is the test root mean squared error
(RMSE) of the trivial predictor ȳ and sets eN−RMSE on a range of 0 to
roughly 1, regardless of the scale on the original response. A value of 1 will
indicate no better performance than the trivial predictor.
The N-RMSE is meant to measure the overall prediction accuracy of a
GaSP in a testing set H. Alternatively to this metric, we could measure
the prediction uncertainty of a testing input vector xt for the BLUP (Sacks
et al., 1989; Cressie, 1993). The computation of its corresponding standard
error is
r
h i [1 − F> R−1 r(xt )]2
SE ŷ xt | θ̂ = σ̂ 1 − r> (xt )R−1 r(xt ) + , (2.15)
F> R−1 F
20
2.6. The Effect of the Nugget Term
The standard error (2.15) does not take into account the uncertainty asso-
ciated to plugging in the estimate θ̂, which is explored by Abt (1999). All
accuracy evaluations in this thesis are based on the point prediction (2.12)
and assessment metrics like (2.14). Comparing validity of uncertainty inter-
vals from various methods is further work.
21
2.7. Training and Testing Designs
22
Chapter 3
Main-Effect Gaussian
Processes
When modelling a deterministic function, one usually trains the GaSP model
in Chapter 2 using data from design D, with subsequent testing on design H.
We could apply certain transformations on the outputs or inputs. As we shall
see, selecting a non-standard correlation structure can have a key impact
on the GaSP fitting and prediction performance. Therefore, as claimed in
Chapter 1, GaSP prediction accuracy can be improved by some innovative
adjustments to the correlation structure.
Chen et al. (2016) tune different modelling settings, in a set of case
studies, to improve the GaSP’s prediction accuracy: the regression compo-
nent µ(x) (2.3) described in Section 2.1, different choices of the correlation
function Rj (hj ) introduced in Section 2.3, different training sizes n, and
more than one class of training design D besides the LHDs described in
Section 2.7. Based on the accuracy results obtained in those different case
studies, Chen et al. (2016) conclude that the common choices of a non-
constant regression component or the SqExp correlation function are often
sub-optimal.
For the regression setting, the authors point out a misleading use of a
non-constant component in Joseph et al. (2008), which is focused on im-
proving prediction accuracy for metamodels in product design optimization.
This work addresses a mainstream test function in computer experiments
literature: the Borehole function (Worley, 1987; Morris et al., 1993). Joseph
et al. (2008) claim that their experiment with a GaSP implementing selected
linear terms in the regression component improves prediction accuracy com-
pared to a GaSP with a constant term. Nevertheless, Chen et al. (2016)
obtain opposite results when replicating this experiment.
Such anecdotal examples, like the one found in Joseph et al. (2008),
make us think about the need for alternative modelling choices different
from the regression component or the correlation function. Hence, we can
explore non-standard correlation structures that go beyond the multiplica-
tive form RStd (x, x0 ) (2.4) to improve prediction accuracy. The literature
23
3.1. Basis Functions and Interpolation
= µ̂(xi ) + e>
i y − Fβ̂ , (3.1)
24
3.2. Implication of the Standard Correlation Structure
where ei is the unitary vector since r(xi ) is the ith column in the correlation
matrix R. This result holds as long as vector r(xi ) is a column of the
correlation matrix R.
If we plug in the regression component (2.13) and simplify the second
term in the BLUP (3.1), then we obtain
> > >
ŷ(xi | θ̂) = β̂ f (xi ) + e>
i y − Fβ̂ = β̂ f (xi ) + [yi − β̂ f (xi )] = yi .
Thus, the interpolation property of the BLUP in (3.1) holds regardless of the
class of correlation structure; as long as we provide an invertible correlation
matrix R.
Equation (2.12) is the same as (3.1) but for predicting a new test input
vector xt . Note that the corresponding BLUP, ŷ(xi | θ̂), for any test input
vector xt is made up of basis functions from the regression component µ̂(xt )
and the correlation functions in the vector r(xt ). Hence, properties of the
correlation function (and the regression) determine the properties of the
BLUP.
Note that optimization of the log-likelihood function remains unchanged
except for the redefinition of the correlation function. Similarly, the BLUP
in (2.12) for testing set H has to incorporate the same non-standard cor-
relation structure. The use of a nugget term (as described in Section 2.6)
will produce a non-interpolator BLUP, but it is a trade-off for having a
better-conditioned correlation matrix R.
hj = xj − x0j
25
3.3. Main-Effect Correlation Structures
for the jth input, a Taylor series expansion of the standard correlation struc-
ture (2.4) is:
d d
!
Y X
Std 0 2
R (x, x ) = Rj (hj ) = exp − θj hj
j=1 j=1
d
X d−1
X d
X
≈1− θj h2j + θj h2j θk h2k + . . .
j=1 j=1 k=j+1
d
Y
± θj h2j + terms involving h4j for at least one j,
j=1
where each subfunction yj (xj ) can be different and only depends on the
single input xj . Following the Effect Hierarchy principle, the main effects
would suffice to fit a predictive model for (3.2). However, the correlation
structure RStd (x, x0 ) is designed to handle input interactions up to order d
as shown in Section 3.2.
The first non-standard correlation structure to be considered has an
26
3.3. Main-Effect Correlation Structures
Note that structure (3.3) assigns equal weights (i.e., E-ME throughout) to
each one of the d main effects with 1/d. The weights ensure scaling of the
left-hand side to have correlations of 1 on the diagonal and [0, 1] elsewhere.
Furthermore, on the right, we have a linear combination of positive semidef-
inite matrices, and hence the matrix on the left is also positive semidefinite.
Similar comments apply to the additive forms later in this chapter and
Chapter 4.
There is a critical connection between the assumed output function
in (3.2) and the BLUP in (2.12). The BLUP depends on the new input
vector xt only through the correlation vector r(xt ); all other vectors and
matrices are fixed after training. Hence, the BLUP as a function of xt is
a linear combination of the basis functions in r(xt ). The correlation func-
tion (3.3) is additive in functions of the various inputs xj , and a linear com-
bination of these functions is also additive in functions of the xj , matching
the form of (3.2). Similar ideas will apply in Chapter 4.
In general, though, if we estimate the d weights one might obtain lower
eN−RMSE in (2.14). These weights need to be estimated along with the rest
of the GaSP parameters described in Chapter 2. Thus, in order to assess
whether unequal weights improve prediction accuracy, we can set up an
additional main-effect correlation structure. Thus, an additive correlation
structure with unequal weights on the main effects (U-ME) is defined as:
d
X
U−ME 0
R (x, x ) = ωj Rj (hj ) ∈ [0, 1], (3.4)
j=1
Since the weights in (3.4) are subject to constraint (3.5), the optimiza-
tion of the log-likelihood function for parameter estimation will also be con-
strained. Thus, we employ a multinomial logit transformation that allows
27
3.3. Main-Effect Correlation Structures
where
exp (τi )
ωi = Pd−1 for i = 1, . . . , d − 1;
1 + k=1 exp (τk )
and
1
ωd = Pd−1 .
1+ k=1 exp (τk )
Under this correlation structure, the d − 1 parameters in vector τ are
estimated and included in the vector of hyperparameters θ̂ (defined in Sec-
tion 2.4 for various correlation functions).
28
3.4. Previous Works
where σk2 is the variance assigned to all the interactions of order k. We can
see that the main-effect correlation structure RE−ME (x, x0 ) in (3.3) is part
of CovAdd1 [Z(x), Z(x0 )], but adjusted by the variance σ12 instead of 1/d.
Further drawbacks and details on the prediction accuracy performance of
this approach will be provided in Chapter 4.
29
3.5. Michalewicz Function
0.0
−0.5
ym
−1.0
−1.5
3
3
2 2
x2 1 1 x1
0 0
30
3.5. Michalewicz Function
0.0 0.00
−0.2 −0.25
−0.4 −0.50
y
y
−0.6 −0.75
−0.8 −1.00
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0
x1 x2
0.00 0.00
−0.25 −0.25
y
y
−0.50 −0.50
−0.75 −0.75
−1.00
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0
x3 x4
0.00 0.00
−0.25 −0.25
−0.50
y
−0.50
−0.75 −0.75
−1.00 −1.00
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0
x5 x6
0.00
−0.25
y
−0.50
−0.75
−1.00
0.0 0.5 1.0 1.5 2.0 2.5 3.0
x7
31
3.5. Michalewicz Function
32
3.5. Michalewicz Function
33
3.5. Michalewicz Function
Setting Values
Dimensionality d = 3, 4, 5, 6, 7
Inputs x ∈ [0, π]d
Output ym (x) as in (3.8)
GaSP Correlation
SqExp and PowExp
Functions
GaSP Correlation RStd (x, x0 ),
Structures RE−ME (x, x0 ), and RU−ME (x, x0 )
Additional
CGP
Approach
20 different mLHDs
D
for each d = 3, 4 of n = 100, 200, 400 runs and
Training Sets
each d = 5, 6, 7 of n = 100, 200, 400, 800 runs
1 rLHD corresponding
H
to d = 3, 4, 5, 6, 7
Testing Set
of N = 10, 000 runs
Table 3.1 summarizes the simulation settings. Since the test function be-
comes more complex as d is increased, we set up studies with d = 3, 4, 5, 6, 7.
Each d has 20 different training sets D for each of n = 100, 200, 400 runs for
d = 3, 4; n = 100, 200, 400, 800 runs for d = 5, 6, 7; and its own d-dimensional
testing set H of N = 10, 000 runs. Our non-standard correlation structures
aim to outperform the Std GaSP beginning with small training sizes. How-
ever, since this case allows us to obtain mLHDs with large training sizes
n via the package DoE.wrapper (Groemping, 2017), we expand our simula-
tion studies to n = 800 to show at what extent the main-effect correlation
structures outperform the Std GaSP.
Each one of the studies allows us to keep track of the evolution of the
prediction performance for the four GaSP approaches: Std, CGP, E-ME,
and U-ME. We use the SqExp (2.6) and PowExp (2.5) correlation functions
for the Std, E-ME, and U-ME GaSPs. The CGP model only uses the SqExp
correlation function as well as the correlation structure RStd (x, x0 ) for its two
independent GaSPs, which is implemented in the R package CGP (Ba and
Joseph, 2018). We do not conduct the comparison study with d = 1, since
34
3.5. Michalewicz Function
this case is trivial. The case d = 2 is also left out because of the risk of an
ill-conditioned correlation matrix R increases.
35
3.5. Michalewicz Function
ent results for the E-ME and U-ME GaSPs from n = 400 to 800 between the
correlation functions. Overall, the E-ME GaSP with the SqExp correlation
function clearly performs best.
Figure 3.4b shows the prediction accuracy for d = 6. Firstly, with
both correlation functions, the additive GaSPs outperform the other two
approaches. In the SqExp case, from n = 100 to 400, the U-ME GaSP
shows the best rate of convergence (1/n2.08 ). From n = 400 to 800, while
the spread in accuracy slightly persists, the E-ME GaSP has the best rate
(1/n12.5 ). Overall, the E-ME GaSP with the SqExp correlation function
clearly performs best.
Figure 3.5 shows the prediction accuracy for d = 7. In the case of the
SqExp correlation function, the CGP and Std GaSP are noticeably out-
performed by their main-effect counterparts even though the E-ME GaSP
shows a huge spread in its results at n = 800. In terms of the rate of con-
vergence, from n = 100 to 400, the U-ME GaSP has the best performance
(1/n0.83 ). This rate drastically increases to 1/n9.17 from n = 400 to 800. For
the PowExp function, both main-effect GaSPs show a similar performance
with a rate of 1/n1.67 from n = 100 to 400. Overall, both the U-ME GaSP
with the SqExp correlation function and the E-ME GaSP with the PowExp
correlation function clearly perform best.
36
3.5. Michalewicz Function
SqExp PowExp
100 0
10 −1
1 −2
log10(eN−RMSE)
eN−RMSE (%)
0.1 −3
0.01 −4
0.001 −5
0.0001 −6
0.00001 −7
100 200 400 100 200 400
n
(a) d = 3.
SqExp PowExp
100 0
10 −1
1 −2
log10(eN−RMSE)
eN−RMSE (%)
0.1 −3
0.01 −4
0.001 −5
0.0001 −6
0.00001 −7
100 200 400 100 200 400
n
(b) d = 4.
37
3.5. Michalewicz Function
SqExp PowExp
100 0
10 −1
1 −2
log10(eN−RMSE)
eN−RMSE (%)
0.1 −3
0.01 −4
0.001 −5
0.0001 −6
0.00001 −7
100 200 400 800 100 200 400 800
n
(a) d = 5.
SqExp PowExp
100 0
10 −1
1 −2
log10(eN−RMSE)
eN−RMSE (%)
0.1 −3
0.01 −4
0.001 −5
0.0001 −6
100 200 400 800 100 200 400 800
n
(b) d = 6.
38
3.5. Michalewicz Function
SqExp PowExp
100 0
31.63 −0.5
10 −1
log10(eN−RMSE)
eN−RMSE (%)
3.16 −1.5
1 −2
0.32 −2.5
0.1 −3
0.03 −3.5
39
3.6. Weighted Michalewicz Function
0.0
0.5
1.0
ywm
1.5
2.0
2.5
3
3
2 2
x2 1 1 x1
0 0
Figure 3.6: 2-dimensional weighted Michalewicz function (3.9).
40
3.6. Weighted Michalewicz Function
0 0
−2 −2
−4 −4
y
y
−6 −6
−8 −8
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0
x1 x2
0 0
−2 −2
−4 −4
y
−6 −6
−8 −8
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0
x3 x4
0 0
−2 −2
−4
y
−4
−6
−6
−8
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.5 1.0 1.5 2.0 2.5 3.0
x5 x6
−2
−4
y
−6
−8
0.0 0.5 1.0 1.5 2.0 2.5 3.0
x7
41
3.6. Weighted Michalewicz Function
42
3.6. Weighted Michalewicz Function
rate (1/n10 ). Overall, the E-ME GaSP with the SqExp correlation function
clearly performs best.
Figure 3.9b shows the prediction accuracy for d = 6. As with smaller d
for the SqExp case, the main-effect structures clearly outperform the CGP
and Std GaSP. We see the same behaviour in the main-effect GaSPs as
in d = 5: large accuracy spreads in the U-ME GaSP with increasing n,
unlike the E-ME GaSP. Note the equal performance of these approaches
from n = 100 to 200 of 1/n1.67 , which drastically changes to 1/n10 for the
E-ME GaSP. Overall, the E-ME GaSP with the SqExp correlation function
clearly performs best.
Figure 3.10 shows the prediction accuracy for d = 7. Regarding the
SqExp function, the CGP and Std GaSP are outperformed by the main-
effect approaches. For the PowExp case, the Std GaSP becomes competitive
with a rate of 1/n2.5 from n = 200 to 800. Nonetheless, this is not enough
to outperform the E-ME GaSP which shows a rate of 1/n5 from n = 200 to
800. Overall, the E-ME GaSP with the PowExp correlation function clearly
performs best.
43
3.6. Weighted Michalewicz Function
SqExp PowExp
100 0
10 −1
1 −2
log10(eN−RMSE)
eN−RMSE (%)
0.1 −3
0.01 −4
0.001 −5
0.0001 −6
0.00001 −7
100 200 400 100 200 400
n
(a) d = 3.
SqExp PowExp
100 0
10 −1
1 −2
log10(eN−RMSE)
eN−RMSE (%)
0.1 −3
0.01 −4
0.001 −5
0.0001 −6
0.00001 −7
100 200 400 100 200 400
n
(b) d = 4.
44
3.6. Weighted Michalewicz Function
SqExp PowExp
100 0
10 −1
1 −2
log10(eN−RMSE)
eN−RMSE (%)
0.1 −3
0.01 −4
0.001 −5
0.0001 −6
0.00001 −7
100 200 400 800 100 200 400 800
n
(a) d = 5.
SqExp PowExp
100 0
10 −1
1 −2
log10(eN−RMSE)
eN−RMSE (%)
0.1 −3
0.01 −4
0.001 −5
0.0001 −6
100 200 400 800 100 200 400 800
n
(b) d = 6.
45
3.7. Concluding Remarks
SqExp PowExp
100 0
31.63 −0.5
10 −1
log10(eN−RMSE)
eN−RMSE (%)
3.16 −1.5
1 −2
0.32 −2.5
0.1 −3
0.03 −3.5
46
3.7. Concluding Remarks
47
Chapter 4
48
4.1. Previous Works
X k
Y
0
CovAddk [Z(x), Z(x )] = σk2 Rum (hum ),
1≤u1 <u2 <···<uk ≤d m=1
whose term σk2 is the variance assigned to the kth order. Therefore, a full
covariance function is the following sum:
d
X
CovAdd [Z(x), Z(x0 )] = CovAddk [Z(x), Z(x0 )]. (4.1)
k=1
We can see that function (4.1) has 2d − 1 terms, which increases expo-
nentially with d and become intractable as recognized by Duvenaud et al.
(2011). This computation also increases processing times in the optimization
procedures of the profile log-likelihood function (2.11).
Furthermore, Duvenaud et al. (2011) do not specify additional details
about the variances σ12 , . . . , σd2 such as specific ranges or estimation proce-
dures. However, they provide some examples with normalized σk2 ∈ [0, 1]
where it is stated that these “order variance” hyperparameters control how
much of the process variance comes from a kth order interaction.
49
4.1. Previous Works
A GaSP using covariance function CovAdd [Z(x), Z(x0 )] (4.1) can pro-
vide an informative framework with each σk2 depicting what orders of in-
teraction are the most important. This covariance proves to be useful in
low-dimensional examples. However, any high-dimensional case leads to a
complex computation under a full order covariance function. Another draw-
back of this approach is the low reduction of the MSE compared to a Std
GaSP with the SqExp correlation function. This is shown in four case stud-
ies with d = 4, 8, 10 inputs.
y(x) = g(x) + ,
50
4.2. Joint-Effect Correlation Structures up to All 2-Input Interactions
Function (4.2) has an additive form, but the inputs are allowed to interact
in specific pairs. Thus, we explore the possibility of expanding the idea of
an additive correlation structure by fully applying the aforementioned effect
principles.
As already introduced in Section 3.3, a main-effect correlation structure
with d single-input main effects corresponds to
d
E−ME 0 1X
R (x, x ) = Rj (hj ) ∈ [0, 1]
d
j=1
or
d
X
RU−ME (x, x0 ) = ωj Rj (hj ) ∈ [0, 1].
j=1
We can see that the inputs do not interact under these previous struc-
tures. Hence, cases such as the Michalewicz function allow this class of
non-standard correlation structures to be competitive against the Std GaSP
in terms of prediction accuracy.
A joint-effect correlation structure goes beyond the previous additive
forms with single-input main effects. It incorporates specific interactions
between the input vectors x and x0 ∈ Rd , along with their respective cor-
relation hyperparameters. The notion of a joint-effect correlation structure
51
4.2. Joint-Effect Correlation Structures up to All 2-Input Interactions
is proposed by Duvenaud et al. (2011), but they clearly do not apply the
Effect Sparsity and Hierarchy principles.
Let x and x0 be two input vectors in Rd , and their respective values of
the random function are Z(x) and Z(x0 ). For the jth input, the correlation
function Rj (hj ) is defined as in Section 2.3. Firstly, we define the joint-
effect correlation structure with equal weights using all 2-input interactions
(E-JE2). This correlation structure assigns fixed weights to the main effects
and all 2-input interactions on a 50-50 basis:
d d d
1 X 1 X X
RE−JE2 (x, x0 ) = Rj (hj ) + d Rj (hj ) · Rk (hk ) ∈ [0, 1].
2d 2 2 j=1 k=j+1
j=1
(4.3)
For the jth input, the correlation function Rj (hj ) in the first addend of
structure (4.3) contains the difference hj = xj − x0j at the input vectors
x and x0 which is viewed as a main effect. If the jth and kth inputs have
a 2-way interaction in this structure, then their corresponding functions
Rj (hj ) and Rk (hk ) are supposed to pick up the correlation induced by that
interaction in the second addend of the structure. A similar idea is applied
to model 3-way interactions in the correlation structures of Sections 4.7 and
4.10.
Note that the two types of terms—main effects and 2-input interactions—
each receive weight 1/2, with the 1/2 distributed equally across all the cor-
relation functions in the type of term. We specify that the correlation func-
tions Rj (hj ) are shared between the main effects and 2-input interactions.
Hence, we use the same correlation hyperparameters in both types of effects.
It is worth mentioning that the correlation structure (4.3), as well as the
subsequent ones, all have different features compared to the CGP by Ba and
Joseph (2012).
As in the case of the U-ME GaSP with structure (3.4), we can incorporate
unequal weights in the aforementioned joint-effect structure. The joint-effect
correlation structure with unequal weights using all 2-input interactions (U-
JE2) is defined as:
d d d
λ1 X λ2 X X
RU−JE2 (x, x0 ) = Rj (hj ) + d Rj (hj ) · Rk (hk ) ∈ [0, 1],
d 2
j=1 j=1 k=j+1
(4.4)
52
4.3. Friedman Function
τ ∈ (−∞, ∞),
where λ1 is computed as
exp (τ )
λ1 = ;
1 + exp (τ )
and λ2 is defined as
1
λ2 = .
1 + exp (τ )
Furthermore, we follow the same optimization procedure for estimating
the parameter τ along with the vectors already defined in Section 2.4 for
various correlation functions. Unlike the U-ME GaSP, the number of pa-
rameters does not significantly increase since we are weighting by type of
effect only.
53
4.3. Friedman Function
Y (x) = f (x) + ,
where inputs are contained in the deterministic function f (·) that can be
non-linear, along with a random component whose mean is zero.
Note that MARS model is intended to tackle cases where we assume a
random part, as in any regular regression technique. It aims to construct
a predictor fˆ(x) on the domain X . This predictor is decomposed into M
basis functions Bi (x) in the following way:
M
X
fˆ(x) = ci Bi (x),
i=1
54
4.3. Friedman Function
Setting Values
Dimensionality d=5
Inputs x ∈ [0, 1]5
Output yfri (x) as in (4.5)
GaSP Correlation
SqExp and PowExp
Functions
GaSP Correlation RStd (x, x0 ),
Structures RE−JE2 (x, x0 ), and RU−JE2 (x, x0 )
Additional MARS
Approach up to 2-input interactions
D 20 different mLHDs with
Training Sets n = 50, 100, 200, 400 runs
H 1 rLHD with
Testing Set N = 10, 000 runs
55
4.4. Joint-Effect Correlation Structures up to Selected 2-Input Interactions
to 0.000003 at n = 400).
Hence, by following the Effect Hierarchy and Heredity principles with
the joint-effect GaSPs, we can obtain 1/3 to 1/20 of the N-RMSE of the
Std GaSP. It might be possible to achieve an even larger improvement if we
also apply the Effect Sparsity principle since function (4.5) has a single 2-
input interaction. Recall that structures (4.3) and (4.4) take into account all
possible 2-input interactions. We can make slight changes to our joint-effect
GaSP correlation structures, as shown in subsequent sections.
SqExp PowExp
100 0
10 −1
1 −2
log10(eN−RMSE)
eN−RMSE (%)
0.1 −3
0.01 −4
0.001 −5
0.0001 −6
50 100 200 400 50 100 200 400
n
56
4.5. Functional Analysis of Variance
in both structures. Recall that a function of the form (4.2) has specific pairs
of inputs interacting with each other, so it would be reasonable to make a
joint-effect correlation structure with selected 2-input interactions.
Under equal weights on the effect framework, a joint-effect correlation
structure with selected 2-input interactions (E-JE2S) is set up in the follow-
ing way:
d
E−JE2S 0 1 X 1 X
R (x, x ) = Rj (hj ) + Rj (hj ) · Rk (hk ) ∈ [0, 1],
2d 2b
j=1 {j,k}∈B
(4.6)
(4.7)
57
4.5. Functional Analysis of Variance
58
4.5. Functional Analysis of Variance
Note that a marginal main effect is obtained when there is only one input
in xe , or a marginal joint effect otherwise.
We use (4.8) to compute the marginal effects that decompose the output
y(x) in the following way:
d
X d−1 X
X d
y(x) = µ0 + µj (xj ) + µjj 0 (xj , xj 0 ) + · · · + µ1...d (x1 , . . . , xd ).
j=1 j=1 j 0 =j+1
(4.9)
Furthermore, the corrected joint effects are computed using the previous
overall and corrected main effects:
d−1 X
X d Z
+ µ2jj 0 (xj , xj 0 )wj (xj )wj 0 (xj 0 )dxj dxj 0
j=1 j 0 =j+1 χj ⊗χj 0
Z d
Y
+ ··· + µ21...d (x1 , . . . , xd ) wj (xj )dxj .
χ j=1
(4.11)
59
4.5. Functional Analysis of Variance
60
4.6. Franke Function
1.2
1.0
0.8
y
0.6
0.4
0.2
1.0
0.8
0.6 1.0
x2 0.8
0.4
0.6
0.2 0.4 x1
0.2
0.0 0.0
61
4.6. Franke Function
y8f (x) = yfra (x1 , x2 ) + yfra (x3 , x4 ) + yfra (x5 , x6 ) + yfra (x7 , x8 ). (4.13)
62
4.6. Franke Function
package CGP (Ba and Joseph, 2018). We aim to determine how competitive
the joint-effect E-JE2S and U-JE2S GaSPs are against MARS (up to 2-input
interactions), CGP, Std, and Oracle GaSPs. Table 4.2 shows the simulation
settings.
Setting Values
Dimensionality d=8
Inputs x ∈ [0, 1]8
Output y8f (x) as in (4.13)
GaSP Correlation
SqExp and PowExp
Functions
RStd (x, x0 ) with one GaSP model (Std),
GaSP Correlation
and four independent GaSPs (Oracle);
Structures
RE−JE2S (x, x0 ), and RU−JE2S (x, x0 )
Additional CGP and MARS
Approach up to 2-input interactions
D 20 different mLHDs with
Training Sets n = 80, 160, 320, 640 runs
H 1 rLHD with
Testing Set N = 10, 000 runs
Note that corresponding to Dpq , there is an output vector yfra (xp , xq ) (4.12)
to fit a GaSP model with only two inputs.
This arrangement will give us the following global prediction on the
(pq)
test input vector xt ∈ [0, 1]2 , based on the sum of four separate BLUPs
63
4.6. Franke Function
Given the form of function (4.13) and how the four GaSP predictions are
obtained, we will not get better prediction accuracy using any other form of
GaSP fitting considered here.
Returning to the practical case where the Oracle is unknown, since we
are dealing with selected 2-input interactions in the joint-effect correlation
structures (4.6) and (4.7), we have to define the pairs of inputs used in
the respective set B. Thus, under a Std GaSP using SqExp and PowExp
correlation functions for the sake of comparison, we run FANOVA up to
2-input interactions on each one of the 20 repeated experiments per training
set size. Note that this analysis is implemented via the GaSP software by
Welch (2014). Integration in FANOVA is with respect to uniform weights
over the inputs’ respective ranges; the same will be done for subsequent case
studies in this chapter.
SqExp PowExp
100
●
● ●
● ●
90 ● ● ● ● ● ● ●
80
Variance Percentage
70
60
50
40
30
20
10 ● ● ● ●
●
●
●
●
●
●
●
●
● ● ● ● ●
0 ● ● ● ● ● ● ●
●
Main Effects ●
2−Input Interactions ●
Higher−Order Interactions
Figure 4.3: FANOVA summary plot by type of effect for the expanded
Franke function (4.13) using a Std GaSP, with SqExp and PowExp cor-
relation functions. Each boxplot shows results from 20 mLHDs of n =
80, 160, 320, 640 runs for training.
64
4.6. Franke Function
Figure 4.3 summarizes the variance percentages for each type of effect
in all our repeated experiments. The lines depict the evolution of the mean
percentage across all training set sizes. Moreover, the plots are faceted by
correlation functions, and we have the following results:
• For both correlation functions, we can see that the main effects account
for 90% or more of the process variance on average. Nonetheless, there
is a subtle increasing trend in the SqExp case, unlike its PowExp
counterpart across all training set sizes. Either way, the percentage
assigned to the eight main effects is high for both functions.
• Figure 4.4a indicates that the main effects for inputs x2 , x4 , x6 , and
x8 each account for about 15% of the predictor variance on average;
the remaining four main effects account for about 7.5% each.
• For the 2-input interactions in Figure 4.4b, there are clearly two mag-
nitudes of effect:
65
4.6. Franke Function
SqExp PowExp
22.5
20.0
17.5
Variance Percentage
15.0
12.5 ●
●
10.0 ●
7.5
5.0
2.5
0.0
x2 x6 x4 x8 x1 x3 x5 x7 x2 x6 x4 x8 x1 x3 x5 x7
4.5
4.0
3.5
3.0
SqExp
2.5
2.0
1.5 ●
Variance Percentage
●
●
● ●
●
1.0 ● ●
●
●
● ●
●
●
0.5 ● ●
●
● ● ●
●
●
●
●
●
●
●
●
●
● ● ● ●
●
●
● ● ●
0.0 ●
4.5
4.0
●
3.5
3.0
PowExp
2.5
2.0
1.5
1.0
0.5 ● ● ●
●
● ● ●
● ● ● ● ● ● ●
● ● ● ● ● ● ●
● ●
●
● ● ●
0.0
x7 x3 x5 x1 x6 x1 x3 x2 x4 x4 x2 x6 x1 x2 x2 x3 x5 x1 x4 x1 x1 x2 x2 x3 x4 x5 x3 x1
x8 x4 x6 x2 x8 x6 x6 x8 x8 x6 x6 x7 x4 x5 x4 x8 x8 x3 x7 x7 x8 x3 x7 x7 x5 x7 x5 x5
Thus, the effects with the largest percentage contributions are our eight
main effects (x1 , x2 , x3 , x4 , x5 , x6 , x7 , and x8 ), plus four 2-input interactions
(x1 · x2 , x3 · x4 , x5 · x6 , and x7 · x8 ). Now, we can construct our set B for
RE−JE2S (x, x0 ) and RU−JE2S (x, x0 ) as:
B = {1, 2}, {3, 4}, {5, 6}, {7, 8} . (4.14)
Thus, the specific correlation structures for this test function are defined as:
8
E−JE2S 0 1 X
R (x, x ) = Rj (hj )
16
j=1
1
+ [R1 (h1 ) · R2 (h2 ) + R3 (h3 ) · R4 (h4 )
8
+ R5 (h5 ) · R6 (h6 ) + R7 (h7 ) · R8 (h8 )] ∈ [0, 1],
and
8
λ1 X
RU−JE2S (x, x0 ) = Rj (hj )
8
j=1
λ2
+ [R1 (h1 ) · R2 (h2 ) + R3 (h3 ) · R4 (h4 )
4
+ R5 (h5 ) · R6 (h6 ) + R7 (h7 ) · R8 (h8 )] ∈ [0, 1].
67
4.6. Franke Function
with a rough rate of 1/n4.12 . Note that the Oracle GaSP starts out with an
eN−RMSE of 0.06 at n = 80, and ends up with 0.00001 at n = 640.
In terms of the joint-effect GaSPs (E-JE2S and U-JE2S, which show
equivalent and overlapped performances), the improvement in prediction
accuracy versus all methods except the Oracle is noticeable. Regardless of
the correlation function, both joint-effect GaSPs show the same prediction
accuracy: an average eN−RMSE of 0.18 at n = 80 and 0.003 at n = 640
(showing a rate of 1/n1.92 ). As we previously pointed out, the prediction
accuracy of the Oracle GaSP will not be outperformed. Nonetheless, we
manage to be decently close with an appropriate GaSP correlation structure.
SqExp PowExp
100 0
10 −1
1 −2
log10(eN−RMSE)
eN−RMSE (%)
0.1 −3
0.01 −4
0.001 −5
0.0001 −6
80 160 320 640 80 160 320 640
n
68
4.7. Joint-Effect Correlation Structures up to All 3-Input Interactions
69
4.8. Joint-Effect Correlation Structure with a Residual Effect Term
The correlation structure (4.17) resembles the structure RE−JE2 (x, x0 ) (4.3),
but there is an additional addend RjRes (hj ) representing the residual com-
ponent under a standard multiplicative structure.
70
4.8. Joint-Effect Correlation Structure with a Residual Effect Term
Building on the previous ideas from Section 4.2, we have the following
set of pairs of input indices:
B = {j, k} for some 1 ≤ j < k ≤ d where |B| = b.
On the other hand, for the PowExp correlation function, the residual com-
ponent becomes:
d d d
!
PowExp 0 p0 p0
Y Y X
0 0 0
RjRes (hj , θ , p ) = exp(−θ hj ) = exp − θ hj .
j=1 j=1 j=1
71
4.9. Output Transformerless Circuit Function
where Vb1 = 12Rb2 /(Rb1 + Rb2 ). The inputs, units and ranges are
72
4.9. Output Transformerless Circuit Function
error function:
n
X 2
yi − β > f α > x .
min E =
β,f ,α
i=1
73
4.9. Output Transformerless Circuit Function
2. With the final tuning parameters, a single MARS and PPR model is
fitted with the training and test sets’ merge. The Std GaSP is directly
fitted using these two datasets as well.
Approach U-RMSE
PPR 0.060
MARS 0.024
Std GaSP under the
0.009
SqExp Correlation Function
Table 4.3 shows the U-RMSE results, obtained by Ben-Ari and Stein-
berg (2007), for assessing prediction accuracy of the three approaches in
the evaluation set of 20 runs. We can see that the Std GaSP outperforms
PPR and MARS. We would expect to obtain similar conclusions regarding
the accuracy differences concerning PPR and MARS compared to the Std
GaSP in our simulation studies.
74
4.9. Output Transformerless Circuit Function
Table 4.4 shows the simulation settings for this case study. The rule of
thumb for the smallest training set size is used, giving n = 6d = 60 with
20 different mLHDs for each of the three training set sizes: 60, 120, and
240 runs. Furthermore, the test set size is fixed at N = 10, 000. Note that,
unlike Ben-Ari and Steinberg (2007), we are using the PowExp correlation
function for the Std GaSP since this setting shows the most competitive
prediction accuracy for the OTL circuit function against our non-standard
correlation structures.
Setting Values
Dimensionality d=6
Inputs x as in (4.20)
Output Vo (x) as in (4.19)
RStd (x, x0 ) using PowExp versus
GaSP Correlation
RE−JE2R (x, x0 ), RE−JE2SR (x, x0 ),
Structures
R E−JE3 (x, x0 ), and RU−JE3 (x, x0 ) using SqExp
Additional PPR and MARS
Approaches up to 2 and 3-input interactions
D 20 different mLHDs with
Training Sets n = 60, 120, 240 runs
H 1 rLHD with
Testing Set N = 10, 000 runs
• The PPR model is fitted with the R base package stats via func-
tion ppr() with 6 as the maximum number of terms to select from
when fitting the model. The smoothing method is Friedman’s “super
smoother” found in Friedman (1984).
• For fitting the MARS model, we use the R package earth (Milborrow,
2019). Furthermore, MARS fitting is set up by allowing three evenly
spaced knots for each input. It is important to state that Ben-Ari and
Steinberg (2007) do not clarify the maximum order of interaction, so
we are setting it up according to our specific subsequent simulation
75
4.9. Output Transformerless Circuit Function
This case study aims to compare our joint-effect GaSPs with all 3-input
interactions (i.e., E-JE3 and U-JE3 from Section 4.7) and residual terms (i.e.,
E-JE2R and E-JE2SR from Section 4.8) against MARS (up to 2 or 3-input
interactions), PPR, and Std GaSP with PowExp. Note that the joint-effect
GaSPs are only implemented with the SqExp correlation function.
As in the case of the Franke function, we are dealing with selected 2-
input interactions in the joint-effect correlation structure RE−JE2SR (x, x0 )
(4.18). Hence, we need to define the pairs of inputs used in the respective
set B. Thus, under a Std GaSP using both SqExp and PowExp correla-
tion functions for the sake of comparison, we run FANOVA up to 3-input
interactions on each of the 20 repeated experiments with training set size
n = 60.
Figure 4.6 illustrates the variance percentages of the six main effects for
the 20 experiments with n = 60, as well as the top six input interactions
that stand out by increasing medians and all of them involve two inputs only.
Figure 4.6a shows the boxplots of the variance percentages corresponding to
the six main effects, where both correlation functions yield similar results.
One can see that input Rc1 is the most active with medians between 55 and
70% followed by Rf with medians between 30 and 40%. Note that Rb2 and
Rb1 are in third and fourth places with medians above 0.25%.
According to Figure 4.6b, which shows the top six input interactions, the
FANOVA results are following the Effect Heredity and Hierarchy principles.
The interaction Rf ·Rc1 is the most active with median variance contribution
slightly below 0.50% for SqExp and 0.75% for PowExp. If we take into
account the remaining five 2-input interactions that stand out (Rb2 · Rc1 ,
Rb1 · Rc1 , Rb2 · Rf , Rb1 · Rf , and Rb1 · Rb2 ) besides Rf · Rc1 , these are all the
possible 42 combinations of the top four aforementioned main effects.
Even though the last three 2-input interactions in Figure 4.6b show small
percentages, their inclusion in our set B is important for the E-JE2SR GaSP
in order to be competitive against the Std GaSP as we will further see along
with its residual component. Hence, by applying the Effect Sparsity princi-
ple, FANOVA suggests the following effect set for RE−JE2SR (x, x0 ) (4.18):
B = {Rf , Rc1 }, {Rb2 , Rc1 }, {Rb1 , Rc1 },
{Rb2 , Rf }, {Rb1 , Rf }, {Rb1 , Rb2 } .
76
4.9. Output Transformerless Circuit Function
90
●
3.0
80
70 2.5
Variance Percentage
60
2.0
50
1.5
40 ●
30 1.0 ●
20
0.5
10
●
0 0.0 ●
●
● ●
0.0035 ●
1.50 ●
0.0030
1.25
Variance Percentage
0.0025 ●
1.00
0.0020
0.75
0.0015
0.50
0.0010
0.25
0.0005
●
●
●
● ●
●
●
0.00 0.0000
Rf.Rc1 Rf.Rc1 Rb2.Rc1 Rb1.Rc1 Rb2.Rf Rb1.Rf Rb1.Rb2 Rb2.Rc1 Rb1.Rc1 Rb2.Rf Rb1.Rf Rb1.Rb2
31.62 −0.5
10 −1
3.16 −1.5
log10(eN−RMSE)
eN−RMSE (%)
1 −2
0.32 −2.5
0.1 −3
0.03 −3.5
60 120 240
n
78
4.9. Output Transformerless Circuit Function
Table 4.5: Summary statistics of eN−RMSE s (2.14) by type of GaSP for the
OTL circuit function (4.19). Each figure summarizes the results from 20
mLHDs of n = 240 runs for training and N = 10, 000 runs for testing. Std
GaSP is implemented with PowExp. E-JE2R, E-JE2SR, E-JE3, and U-JE3
GaSPs are implemented with SqExp.
Figure 4.8 shows the prediction accuracy in for this case study at n =
240. We compare the eN−RMSE s (2.14), in their original range of 0 to roughly
1, on the y-axis by GaSP approach on the x-axis with side-by-side boxplots.
Overall the E-JE2SR and U-JE3 GaSPs provide the best prediction accuracy
among all the GaSP approaches with medians of 0.0006 and 0.0007. The
medians are reduced by 33.33% and 22.22% respectively for E-JE2SR and U-
JE3 compared to the Std GaSP. To check whether this difference in medians
is not due to chance with respect to Std GaSP, a Wilcoxon test can be
applied. This results in a p-value < .001 for both joint-effect approaches.
79
4.10. Joint-Effect Correlation Structures up to Selected 3-Input Interactions
0.0018
●
0.0016
0.0014
0.0012
eN−RMSE
0.0010
0.0008
0.0006
0.0004
E−JE2R E−JE2SR E−JE3 U−JE3 Std
Figure 4.8: Prediction accuracy by type of GaSP for the OTL circuit func-
tion (4.19). Each boxplot shows results from 20 mLHDs of n = 240 runs for
training and N = 10, 000 runs for testing. Std GaSP is implemented with
PowExp. E-JE2R, E-JE2SR, E-JE3, and U-JE3 GaSPs are implemented
with SqExp.
80
4.10. Joint-Effect Correlation Structures up to Selected 3-Input Interactions
Function (4.21) has an additive form, but the inputs are interacting in spe-
cific pairs and triples. Therefore, it would be reasonable to make a joint-
effect correlation structure up to selected 3-input interactions.
With equal weights, a joint-effect correlation structure up to selected 3-
input interactions (E-JE3S) reduces its number of 2 and 3-input interactions
in the following way:
d
E−JE3S 0 1 X 1 X
R (x, x ) = Rj (hj ) + Rj (hj ) · Rk (hk )
3d 3b
j=1 {j,k}∈B
1 X
+ Rj (hj ) · Rk (hk ) · Rl (hl )
3c
{j,k,l}∈C
We assign equal weights to main effects and both interaction orders. Note
that as the numbers of elements in sets B and C decrease, the equal weight
assigned to each 2 and 3-input interaction respectively increases.
We can set up unequal weights for this class of correlation structure
(U-JE3S) as:
d
λ1 X λ2 X
RU−JE3S (x, x0 ) = Rj (hj ) + Rj (hj ) · Rk (hk )
d b
j=1 {j,k}∈B
λ3 X
+ Rj (hj ) · Rk (hk ) · Rl (hl )
c
{j,k,l}∈C
where the weights fulfil the constraint 3j=1 λj = 1. The weights and the
P
other correlation parameters are optimized as in Section 4.7.
81
4.11. Nilson-Kuusk Model
• input pairs in set B for correlation structures RE−JE2S (x, x0 ) (4.6) and
RU−JE2S (x, x0 ) (4.7), as well as,
82
4.11. Nilson-Kuusk Model
Setting Values
Dimensionality d=5
Inputs view zenith, LAI, sl, lmbd z, and lambda
Output CR
GaSP Correlation
SqExp and PowExp
Functions
RStd (x, x0 ),
GaSP Correlation
RE−JE2S (x, x0 ), RU−JE2S (x, x0 ),
Structures
R E−JE3S (x, x0 ), and RU−JE3S (x, x0 )
D 25 randomized LHDs with
Training Sets n = 200 runs
H 25 randomized LHDs with
Testing Set N = 50 runs
Figure 4.9 shows the boxplots of the variance percentages accounted for
by estimated main effects, 2 and 3-input interaction effects from all the 25
randomized LHDs of n = 200 runs. Both correlation functions yield similar
results, and the main effects together have median around 92% whereas the
2-input interactions have median around 8%. Note that 3-input interactions
show small percentages, but one of them will emerge for the selective joint-
effect structures RE−JE3S (x, x0 ) (4.22) and RU−JE3S (x, x0 ) (4.23).
Figure 4.10a shows the boxplots of the variance percentages correspond-
ing to the five main effects where both correlation functions yield similar
83
4.11. Nilson-Kuusk Model
results. One can see that input lambda is the most active one with medians
around 87% followed by LAI with medians close to 4%.
According to Figure 4.10b, which shows the top eight input interac-
tions, the interaction LAI · lambda is the most active with median variance
contribution about 7.5%. Furthermore, both inputs appear in the 3-input
interaction view zenith · LAI · lambda and either one of them appears in the
remaining six 2-input interactions.
Again, the FANOVA results are following the Effect Heredity and Hi-
erarchy principles. Furthermore, applying the Effect Sparsity principle,
FANOVA suggests the following effect sets:
B = {LAI, lambda}, {view zenith, lambda}, {sl, lambda},
{view zenith, LAI}, {lmbd z, lambda}, {view zenith, lmbd z},
{LAI, sl} ;
and
C = {view zenith, LAI, lambda} .
Regarding the input interactions for sets B and C, the boxplots in Fig-
ure 4.10b show that the variance percentages of these effects are never equal
to zero across the 25 randomized LHDs of n = 200 runs for training.
84
4.11. Nilson-Kuusk Model
SqExp PowExp
93.00
92.75
92.50
Variance Percentage
92.25
92.00
91.75
91.50
91.25
91.00
Main Main
Effect
SqExp PowExp
10
8
Variance Percentage
Figure 4.9: FANOVA summary plots for the Nilson-Kuusk model using a
Std GaSP, with SqExp and PowExp correlation functions. Each boxplot
shows results from 25 randomized LHDs of n = 200 runs for training.
85
4.11. Nilson-Kuusk Model
SqExp PowExp
88.0
Variance Percentage
87.5
87.0
86.5
86.0
85.5
85.0
lambda lambda
SqExp PowExp
5.0
Variance Percentage
4.5
4.0 ●
3.5
3.0
2.5
2.0
1.5
1.0 ●
0.5
0.0
LAI view_zenith sl lmbd_z LAI view_zenith sl lmbd_z
SqExp PowExp
8.0
Variance Percentage
7.9
7.8
7.7
7.6
7.5
7.4
7.3
7.2
7.1
7.0
LAI.lambda LAI.lambda
SqExp PowExp
0.200
Variance Percentage
0.175
0.150
●
0.125
0.100
0.075 ●
0.050
0.025
● ●
0.000
view_zenith view_zenith sl view_zenith lmbd_z view_zenith view_zenith view_zenith sl view_zenith lmbd_z view_zenith
lambda LAI lambda LAI lambda lmbd_z lambda LAI lambda LAI lambda lmbd_z
lambda lambda
Figure 4.10: FANOVA plots by type of effect for the Nilson-Kuusk model
using a Std GaSP, with SqExp and PowExp correlation functions. Each
boxplot shows results from 25 randomized LHDs of n = 200 runs for training.
86
4.11. Nilson-Kuusk Model
SqExp PowExp
0.11
●
●
0.10
●
● ●
0.09 ●
0.08
eN−RMSE
0.07
0.06 ●
●
0.05
0.04
● ● ●
0.03
Std E−JE2S U−JE2S E−JE3S U−JE3S Std E−JE2S U−JE2S E−JE3S U−JE3S
87
4.11. Nilson-Kuusk Model
SqExp
Approach Min Q1 Median Mean Q3 Max IQR
Std 0.061 0.065 0.069 0.069 0.073 0.084 0.008
E-JE2S 0.058 0.071 0.075 0.075 0.078 0.092 0.007
U-JE2S 0.059 0.071 0.074 0.075 0.077 0.091 0.006
E-JE3S 0.042 0.053 0.058 0.058 0.063 0.074 0.010
U-JE3S 0.041 0.053 0.058 0.057 0.061 0.072 0.008
PowExp
Approach Min Q1 Median Mean Q3 Max IQR
Std 0.036 0.057 0.061 0.060 0.066 0.073 0.009
E-JE2S 0.060 0.069 0.073 0.075 0.080 0.103 0.011
U-JE2S 0.061 0.068 0.072 0.075 0.080 0.102 0.012
E-JE3S 0.036 0.052 0.055 0.055 0.059 0.072 0.007
U-JE3S 0.036 0.053 0.055 0.056 0.061 0.071 0.008
Table 4.7: Summary statistics of eN−RMSE s (2.14) by type of GaSP for the
Nilson-Kuusk model, with SqExp and PowExp correlation functions. Each
figure summarizes the results from 25 randomized LHDs of n = 200 runs for
training and 25 randomized LHDs of N = 50 runs for testing.
88
4.12. Concluding Remarks
89
Chapter 5
Dimensional Analysis
Our second approach to improve prediction accuracy is DA, which pays at-
tention to fundamental physical dimensions when modelling scientific and
engineering systems. As stated in Chapter 1, early works in DA were done by
Buckingham (1914), Rayleigh (1915), and Bridgman (1931). These works do
not focus on mechanical tools for modelling but provide important DA foun-
dations. A critical point to highlight is that a dimensionally homogeneous
system of a natural phenomenon is depicted by relationships between the
seven fundamental physical dimensions. Nonetheless, a given physical law
coming from this system cannot depend on specific scales of measurement
known as “units”. Statistical modelling cannot be exempt from this fact.
Still, there are examples where the figure of the “unconscious statistician”
could arise (Lee and Zidek, 2020).
Even though statisticians usually do not take DA into account when
it comes to data modelling, we will see in this chapter its great potential.
Hence, DA can be explored as an alternative strategy in model training and
testing. DA allows us to deduce the physical relationships between a given
system’s variables while shedding light on their complex dynamics. Note
that “dimension” is being used in two senses here. DA is an analysis of
measurement units, and hence “dimensionless” refers to a possibly derived
variable with no units. A consequence of DA is that it also leads to dimension
reduction, where the dimensionality d of the inputs is reduced.
Some DA-related statistical modelling can be found in the literature
(Shen et al., 2014; Shen and Lin, 2018; Shen et al., 2018). These previous
works have addressed fundamental challenges in specific DA practices (e.g.,
the choice of base quantities or additional input/output transformations).
Shen et al. (2014) provide background and basic guidelines for DA, along
with examples of physical experiments. Furthermore, the work by Shen
et al. (2018) specifically applies DA work to computer experiments using
novel designs for model training.
The non-standard correlation structures in Chapters 3 and 4 are focused
on the prediction accuracy improvement related to interpolation, i.e., the
testing set H has the same d input ranges as its corresponding training set
90
Chapter 5. Dimensional Analysis
D. Therefore, predictions are obtained with input vectors under the same
ranges used for model training. Nonetheless, the experimenter might be
interested in predictions outside of the trained input ranges. This interest
leads to extrapolation. Note that Shen et al. (2018) argue that a model
under dimensional constraints performs better with extrapolated input vec-
tors, mainly when complexity arises in response surfaces. Hence, we will
target the use of DA to improve prediction accuracy for interpolation and
extrapolation.
DA can be thought of as a data pre-processing step. It analyzes in
terms of more fundamental dimensionless quantities derived from the orig-
inal variables, and possibly designs for them too. The derivation of these
dimensionless quantities takes into account the system’s fundamental physi-
cal dimensions out the total number of seven listed in Section 1.2. With the
“right” dimensionless quantities, prediction accuracy will hopefully improve.
While the above goals make much sense for scientific applications and
statistical modelling, DA’s implementation is far from straightforward; choos-
ing the derived quantities is particularly problematic. Empirical approaches
to finding “good” derived variables in computer experiments will be de-
scribed in this chapter, and the improvements in prediction accuracy will be
demonstrated in three case studies. These choices will be implemented via
FANOVA, as in our joint-effect correlation structures with selected input
interactions from Chapter 4. Note that constructing dimensionless derived
inputs resembles feature engineering from machine learning. However, our
implementation is not automatic but manual based on science.
The paradigm of choosing the “right” dimensionless quantities is more
unmistakable in the cases described in Sections 5.3 and 5.4. Firstly, we will
compare the accuracy of interpolated predictions with a trained model using
a FANOVA-based DA to others under Non-DA and alternative DAs from
the literature (Shen et al., 2018; Tan, 2017). The FANOVA-based DA out-
performs these other approaches. Then, since DA is related to the scaling
of variables, we also illustrate sustained accuracy gains when extrapolating
substantially outside the training data. The FANOVA-based DA’s outper-
formance, when compared to the other approaches, is even more striking for
extrapolated predictions.
Overall, we illustrate how the DA can improve prediction accuracy in
a computer experiment via three cases. Moreover, the use of sensitivity
analysis tools like FANOVA is helpful in key DA steps. This chapter is
structured as follows:
91
Chapter 5. Dimensional Analysis
• Section 5.3 introduces the Borehole function as our first DA case study.
The Borehole function is commonly used in computer experiments lit-
erature. We first show that logarithmic input transformations and
a further input space expansion can improve prediction accuracy in
a non-DA framework. Then, along with the use of a FANOVA-based
DA, we make additional accuracy improvements. We also compare our
FANOVA-based DA with an alternative DA (Shen et al., 2018). More-
over, we study extrapolation, where FANOVA-based DA substantially
improves prediction accuracy.
• Section 5.5 illustrates our final case study in this thesis, which is a
Storm Surge model with data coming from the Advanced Circulation
(ADCIRC) model (Luettich et al., 1992). Here GaSP is used as a
surrogate model to predict storm surges along the NC coast. This
model is an application of a full DA approach to a certain extent
to improve the prediction accuracy of a GaSP compared to a non-
DA framework. A further study shows the advantage of the novel
approaches in extrapolating a model trained only with low to moderate
intensity storms to predict surge from the highest intensity storms.
The main thrust of this thesis is the GaSP’s prediction accuracy im-
provement by either using more than one non-standard correlation structure
or different DA arrangements. FANOVA is the sensitivity analysis tool in
common between both strategies. Note that the experimenter could combine
both strategies, but at the cost of significantly increasing the number of pos-
92
5.1. Previous Applications of Dimensional Analysis
sible modelling settings. Furthermore, the previous case studies do not have
a suitable framework based on the seven fundamental physical dimensions.
Hence, unlike our previous simulation studies in Chapters 3 and 4, all
cases throughout this chapter will only involve the use of a Std GaSP. DA is
focused on strategies for choosing variables for modelling, i.e., input/output
transformations and input space expansions. Moreover, based on our pre-
vious results regarding the Std GaSP in Chapters 3 and 4, the use of the
PowExp correlation function provides better accuracy than the SqExp coun-
terpart for this class of GaSP. Therefore, we will also restrict our attention
to PowExp.
93
5.1. Previous Applications of Dimensional Analysis
y = a × xb
log(y) = log (a) + b log (x);
94
5.1. Previous Applications of Dimensional Analysis
95
5.2. Buckingham’s Π-Theorem
An important aspect of the theorem is that all the system inputs need
to be included in equation (5.1). Thus, this equation will depict the rela-
tionships between those inputs in terms of the seven fundamental physical
dimensions already listed in Section 1.2 whose coefficients (i.e., other con-
stants and parameters) will be dimensionless. These coefficients represent
the fixed interrelations of the units used to measure the fundamental physical
dimensions. However, they do not depend on any specific system of mea-
surement (e.g., SI or Imperial System). The theorem is explained by Shen
et al. (2014), from the statistical perspective, in the form of the following
four steps:
(I) System’s Dimensional Setup. The initial step in DA requires in-
spection of the fundamental dimensions involved in the system for
inputs and output. Define f (·) as the function to be estimated:
y = f (x1 , . . . , xd ). (5.2)
96
5.2. Buckingham’s Π-Theorem
where we have p terms dk,i that allow the kth row to be expressed in
terms of the first p rows. The rank p is the number of inputs that have
to be considered as base quantities in the system. As stated by Shen
et al. (2014), the dimensions of these base quantities have to fulfil the
following conditions:
97
5.3. Borehole Function
98
5.3. Borehole Function
where the inputs and their respective dimensions, units, and ranges are
detailed in Table 5.1.
Table 5.1: Inputs, dimensions, units, and ranges for the Borehole func-
tion (5.3).
The Borehole function does not pose significant predictive challenges for
a GaSP, mostly in interpolating predictions. Nevertheless, it will exemplify
the application of a DA by following the steps detailed in Section 5.2. As we
will see, prediction accuracy is greatly improved by conducting appropriate
input transformations and an input space expansion without DA. The appli-
cation of DA, along with the previous input arrangements, allow additional
prediction accuracy improvement. Section 5.3.1 provides the FANOVA re-
sults used for the selection of base quantities in this DA.
Our simulation studies start under a baseline arrangement, i.e. as train-
ing and testing sets are provided by the DoE.wrapper package with the
original variables, then we proceed with input logarithmic transformations
99
5.3. Borehole Function
d−p=6
100
5.3. Borehole Function
(rw ) roughly accounts for a median of 82.5%, which is the most dom-
inant main effect. In terms of dimension time T, the input hydraulic
conductivity of borehole (Kw ) is the most important main effect in-
volving this fundamental dimension with percentages around 0.95%.
90 1.0
●
80 0.9
0.8
70
0.7
Variance Percentage
60
0.6
50
0.5
40
0.4
30
0.3
20
0.2
10 0.1
●
0 0.0 ● ●
rw Hl Hu L Kw Tl r Tu
Figure 5.1: FANOVA percentage contributions of main effects for the Bore-
hole function (5.3) using a Std GaSP, with PowExp correlation function
and a constant regression term. Each boxplot shows results over 20 repeat
experiments with different mLHDs of n = 80 runs for training.
101
5.3. Borehole Function
1.3
1.2
1.1
1.0
0.9
Variance Percentage
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0.0
(F ) yb (x)
q0 = 2K
, (5.4)
rw w
102
5.3. Borehole Function
(F )
and the qi dimensionless inputs (i = 1, . . . , 6) are obtained as
(F ) r (F ) Tu (F ) Hu
q1 = q2 = q3 =
rw rw Kw rw
(F ) Tl (F ) Hl (F ) L
q4 = q5 = q6 = .
rw Kw rw rw
(5.5)
(S)
• Shen’s DA. Let q0 be the new transformed output defined as
(S) yb (x)
q0 = , (5.6)
Hu Tu
(S)
and the qi dimensionless inputs (i = 1, . . . , 6) are obtained as
i.e.
yb (x) r Tu Hu Tl Hl L
2K
=g , , , , ,
rw w rw rw Kw rw rw Kw rw rw
for FANOVA DA, and
(S) (S) (S) (S) (S) (S) (S)
q0 = h q1 , q2 , q3 , q4 , q5 , q6
i.e.
yb (x) rw r Tl Hl L Kw Hu
=h , , , , ,
Hu Tu Hu Hu Tu Hu Hu Tu
for Shen’s DA.
103
5.3. Borehole Function
Baseline Borehole
Baseline Borehole starts without any logarithmic transformations of the in-
puts. We compare the prediction accuracy of three different approaches:
Non-DA, FANOVA DA, and Shen’s DA. Table 5.2 shows the simulation set-
tings for this arrangement. The rule of thumb for the smallest training set
size is used here: n = 10d = 80 with 20 different mLHDs; similarly, the other
three training sizes. Moreover, the testing set size is fixed at N = 10, 000.
Setting Values
Non-DA: d = 8
Dimensionality
FANOVA DA and Shen’s DA: d = 6
Non-DA: x as in Table 5.1
Inputs FANOVA DA: q(F ) as in (5.5)
Shen’s DA: q(S) as in (5.7)
Non-DA: yb as in (5.3)
Outputs (F )
FANOVA DA: q0 as in (5.4)
(S)
Shen’s DA: q0 as in (5.6)
Regression Component Constant and Linear
D 20 different mLHDs with
Training Sets n = 80, 160, 320, 640 runs
H 1 rLHD with
Testing Set N = 10, 000 runs
104
5.3. Borehole Function
Log-Input Borehole
Taking logarithms of the inputs is particularly relevant for the input radius
of influence (r), which ranges over several orders of magnitude. Thus, the
baseline settings’ first variation is a logarithmic transformation on the inputs
with an untransformed output.
We compare the prediction accuracy of three approaches: Non-DA,
FANOVA DA, and Shen’s DA. A logarithmic input transformation in Non-
DA is not expected to improve its prediction accuracy as much as in a DA
setup. Recall the importance of setting up dimensionless inputs before a
logarithmic transformation, which is addressed in an example presented by
Shen et al. (2014).
We also keep all the simulation settings from Baseline Borehole, as well
as the same training and testing sets detailed in Table 5.2; except for the
inputs log(x) as in Table 5.1, log q(F ) as in (5.5) for FANOVA DA, and
Setting Values
Non-DA: d = 64
Dimensionality
FANOVA DA and Shen’s DA: d = 36
Non-DA:
log(x) as in Table (5.1),
and all pairwise sums and differences
of the elements of log(x)
FANOVA DA:
Inputs (F )
log q as in (5.5),
and all pairwise sums and differences
of the elements of log q(F )
Shen’s
DA:
log q (S) as in (5.7),
and all pairwise sums and differences
of the elements of log q(S)
105
5.3. Borehole Function
and
6
d=6+2 = 36 for FANOVA DA and Shen’s DA,
2
which targets all possible 2-input interactions (products and quotients) plus
the main effects on the logarithmic scale. Even though the processing times
in maximum likelihood optimization increase as the number of inputs in-
creases as well, this impact is not as large as the increase of the training set
size n.
106
5.3. Borehole Function
next two rows that all combinations of the three DA strategies and two lin-
ear models benefit considerably from the logarithmic transformation of the
inputs. These DA strategies benefit further from the expanded-log inputs.
FANOVA DA gains by far the most. It achieves an average eN−RMSE of
0.00006 at n = 80 and ends with 0.000002 at n = 640. Hence this method
has near-perfect accuracy even with n = 80. We also note that the inclusion
of a linear regression component does not improve prediction accuracy than
the constant case in any of our modelling strategies.
The two axes in Figure 5.3 are on logarithmic scales, and all lines join-
ing the sample means of eN−RMSE are approximately linear throughout the
range. Hence, the slope r of a line indicates the approximate rate of conver-
gence nr of eN−RMSE (or RMSE) to zero with sample size. With expanded-
log inputs, for example, all methods show similar rates of convergence: all
achieve eN−RMSE going down faster than n−1 (note this is for root MSE, not
MSE).
107
5.3. Borehole Function
Baseline Borehole
Constant Linear
1 −2
0.32 −2.5
log10(eN−RMSE)
eN−RMSE (%)
0.1 −3
0.03 −3.5
0.01 −4
80 160 320 640 80 160 320 640
n
Log−Input Borehole
Constant Linear
10 −1
1 −2
log10(eN−RMSE)
eN−RMSE (%)
0.1 −3
0.01 −4
0.001 −5
0.0001 −6
80 160 320 640 80 160 320 640
n
0.1 −3
log10(eN−RMSE)
eN−RMSE (%)
0.01 −4
0.001 −5
0.0001 −6
80 160 320 640 80 160 320 640
n
108
5.3. Borehole Function
5.3.4 Extrapolation
As clarified at the beginning of this chapter, we now turn to extrapolation.
Shen et al. (2018) argue that a model, satisfying dimensional constraints,
performs better when input ranges are extrapolated, especially for a complex
response surface. FANOVA DA and Shen’s DA both satisfy the dimensional
constraints and are compared with Non-DA.
We obtain extrapolated predictions based on these three modelling strate-
gies with the same 20 training sets per size n. Simulation settings are as
before, and the input ranges are as in Table 5.1 for training. Two inputs
have extended ranges, however, in the testing set H for extrapolation. We
keep the same testing set size: N = 10, 000 runs. Our FANOVA results
in Figure 5.1 show that radius of borehole (rw ) and potentiometric head
of lower aquifer (Hl ) are two important inputs, and H now has with the
following extrapolated ranges for them:
rw ∈ [0.15, 0.25]
Hl ∈ [820, 880]. (5.8)
109
5.3. Borehole Function
Baseline Borehole
Constant Linear
31.63 −0.5
10 −1
log10(eN−RMSE)
eN−RMSE (%)
3.16 −1.5
1 −2
0.32 −2.5
80 160 320 640 80 160 320 640
n
Log−Input Borehole
Constant Linear
31.63 −0.5
10 −1
log10(eN−RMSE)
eN−RMSE (%)
3.16 −1.5
1 −2
0.32 −2.5
0.1 −3
3.16 −1.5
1 −2
log10(eN−RMSE)
eN−RMSE (%)
0.32 −2.5
0.1 −3
0.03 −3.5
0.01 −4
0.003 −4.5
80 160 320 640 80 160 320 640
n
110
5.4. Heat Transfer in a Solid Sphere
Table 5.4: Inputs, dimensions, units, and ranges for the Solid Sphere func-
tion (5.11).
The second case study was used by Tan (2017) to illustrate DA for polyno-
mial emulation of a complex computer code. As in the case of a GaSP, a
polynomial metamodel is intended to provide quick output predictions with
training datasets composed of a small number of input runs. Nonetheless,
polynomial metamodels might be of high-degree if one wants to obtain a
good approximation of the computer code.
To make it flexible enough, when the polynomial may need to be high-
degree, Tan (2017) uses two approaches to reduce the number of terms:
111
5.4. Heat Transfer in a Solid Sphere
(T ) kt (T ) h2c
q4 = q5 = .
cρr2 ∆T c3 ρ2
(5.10)
(T ) (T ) (T )
Here q0 and q1 are dimensionless temperature ratios, q2 is a dimension-
(T )
less distance ratio from the center of the sphere relative to its radius, q3
(T )
is the dimensionless Biot number, q4 is the dimensionless Fourier number,
(T )
and q5 is the dimensionless convective heat transfer coefficient.
(T )
Then, q0 is given by
∞ (T )
(T ) (T )
X 4(sin ηi − ηi cos ηi ) −η2 q4(T ) sin ηi q2
q0 = q1 + e i (T )
; (5.11)
2ηi − sin(2ηi ) ηi q
i=1 2
112
5.4. Heat Transfer in a Solid Sphere
35 3.5 ●
30 3.0
25 2.5
Variance Percentage
Variance Percentage
20 2.0
15 1.5
10 1.0
5 0.5
●
● ●
0 0.0
Figure 5.5: FANOVA percentage contributions by type of effect for the Solid
Sphere function (5.11) using a Std GaSP, with PowExp correlation function
and a constant regression term. Each boxplot shows results from 20 mLHDs
of n = 70 runs for training.
113
5.4. Heat Transfer in a Solid Sphere
(t), with a median close to 30%. Finally, radius of sphere (r) has a
median of 15% and involves fundamental dimension L.
Based on the FANOVA results we select base quantities for DA. The
system has p = 4 fundamental dimensions: Θ, L, T, and M. Therefore, we
have to choose four inputs as base quantities from the set of nine in Table 5.4.
FANOVA suggests the inputs Tm , t, r, and hc . In the Appendix B we show
the main-effect plots for the seven varying inputs in this system. In these
figures, we clearly see that R and r have a non-linear effect on the output.
The input t also shows modest non-linear behaviour, and this matter will
be dealt in the DA setup as well.
We gain insight into the behaviour of the effect of r by inspection of its
main effect plot in Figure 5.6 from FANOVA. Additionally, this figure in-
cludes the fit of a simple linear regression of the observed main effect against
1/r. The good fit suggests that Ts is approximately linear in 1/r, which will
be taken into account in the DA when considering the dimensionless output.
Thus, the DA has five dimensionless inputs:
(F ) (F ) ∆T (F ) k
q1 =R q2 = q3 =
Tm + ∆T hc r
r s
(F ) ct2 Tm (F ) 3 hc t3 Tm
q4 = q5 = . (5.12)
r2 ρr3
There are two key points we have to highlight in the input setup (5.12):
(F )
• q2 is the proportion that ∆T represents in the initial sphere temper-
ature Tm + ∆T .
(F ) (F ) (F )
• Based on our findings in Figure 5.6; q3 , q4 , and q5 contain the
input r as a reciprocal. Note that we apply a square and cubic root
(F ) (F )
transformation on q4 and q5 respectively, so that r appears as 1/r.
Furthermore, these transformations make input t linear, as suggested
in the next step for the output.
114
5.4. Heat Transfer in a Solid Sphere
315
●
●
●
310
●
305 ●
Ts
300
●
●
295
290
Figure 5.6: Estimated main effect (solid line) of radius of sphere (r) on tem-
perature of sphere (Ts ) from FANOVA with PowExp correlation function
and a constant regression term. The dashed lines show approximate point-
wise 95% confidence limits, and the dotted line shows the fitted values from
a simple least-squares regression of Ts on 1/r.
(F ) Ts − Tm
q0 = , (5.13)
∆T
is shown in Figure 5.7c. Its maximum value is 1 by definition, and Tm and
∆t are used as suggested by FANOVA. Again there is a downward trend with
115
5.4. Heat Transfer in a Solid Sphere
t, but overall there is less scatter than in Figures 5.7a and 5.7b, particularly
for small values of t.
Finally, Meinsma (2019) provides a much simpler example involving the
temperature over time of an object in a medium of temperature 0 (pre-
sumably K, the author does not give units). According to Newton’s law of
cooling, the rate of cooling for this object is proportional to the difference
between its temperature and the temperature of the medium at time t. Ac-
(F )
cording to the idealized physics, log q0 should be approximately linear
in t. The LOESS regression shown in Figure 5.7d depicts an approximately
negative linear relationship. There is still scatter (due to the other inputs
varying in the design), but it is arguably slightly less than in any of the
(F )
other panels. The dimensionless output will therefore be log q0 in our
DA.
116
5.4. Heat Transfer in a Solid Sphere
●
340 ● ●
●
6.0
●
● ●
●
●●
● ●
●
● ● ●
●
320 ●
● 5.5 ●
● ● ●
● ● ●
● ●
● ● ● ●
● ● ●
● ●
● ● ●
● ● ● ● ●
● ● ● ●
● ● ●
● ● ● ●
● ● ● ●
●
● ●
● 5.0 ● ● ● ●
● ● ● ●
(T)
● ● ●
● ● ●
Ts
●
300 ●
q0
● ● ● ●
●
●
● ●
● ●
● ●
● ● ●
●
● ●
●
● 4.5 ● ●
●
●
● ● ● ●
● ●
● ● ●
● ●
● ● ●
● ● ●
● ● ● ●
280 ● ●
●●
●
● ● ● ●
4.0 ●
●
●
●
●
●
260 3.5
0 100 200 300 400 500 600 0 100 200 300 400 500 600
t t
(T )
(a) Ts against t. (b) q0 against t.
1.1 0.4
●● ● ● ●
●
● ● ●
● ●
●
●
●● ●
0.9 ●
● ●
● 0.0 ●● ● ● ●
● ●
●
●
● ● ● ●
●● ● ● ●
● ●● ●
● ● ● ●
● ●
● ●● ● ● ●
● ● ● ● ● ●
● ● ● ●
● ●
● ● ● ●
● ● ● ●
● ● ● ● ●
● ●● ●
●● ● ●
log(q0 )
●
( F)
● ● ●
●●
●
(F)
● ● ●
●
−0.4 ●
q0
●● ● ●
0.7 ●
●
●
● ●
● ●
● ● ●
●
●
●
● ●
● ● ●
●
● ●
●
● ●
●
● −0.8
0.5
● ●
●
●
●
●
●
● ●
●
● −1.2
0.3
0 100 200 300 400 500 600 0 100 200 300 400 500 600
t t
(F ) (F )
(c) q0 against t. (d) log q0 against t.
117
5.4. Heat Transfer in a Solid Sphere
Setting Values
Regression Component Constant and Linear
D 20 different mLHDs with
Training Sets n = 70, 140, 280, 560 runs
H 1 rLHD with
Testing Set N = 10, 000 runs
Table 5.5 shows all the simulation settings for this case study. The rule
of thumb for the starting training set size is used here with the number of
non-constant inputs in the Non-DA framework, resulting in n = 70 with 20
different mLHDs for each one of the four sizes: 70, 140, 280, and 560 runs.
Moreover, the testing set size is fixed at N = 10, 000. Table 5.6 shows the
specific simulation settings for each DA approach in terms of dimensionality,
inputs, and outputs. We highlight the following points regarding these DA
approaches:
118
5.4. Heat Transfer in a Solid Sphere
as in (5.12),
(F )
Expanded and all pairwise sums log q0
d = 25
Full-DA and differences as in (5.13)
of the elements of
log q (F )
119
5.4. Heat Transfer in a Solid Sphere
Constant Linear
100 0
31.62 −0.5
log10(eN−RMSE)
eN−RMSE (%)
10 −1
3.16 −1.5
1 −2
0.32 −2.5
70 140 280 560 70 140 280 560
n
120
5.4. Heat Transfer in a Solid Sphere
5.4.4 Extrapolation
As in the Borehole function, the Solid Sphere function can be used for an
additional simulation study on extrapolation. We obtain extrapolated pre-
dictions with the same 20 training sets per size n for Non-DA, Tan’s DA,
Full-DA, and Expanded Full-DA. Simulation settings in Tables 5.5 and 5.6
are kept in this study except for changes to four specific input ranges on the
testing set H, as described next.
Our FANOVA results in Figures 5.5a and 5.5b show that temperature
of medium (Tm ), time (t), initial sphere temperature minus temperature of
medium (∆T ), and radius of sphere (r) are the top four main effects; and the
2-input interaction r · t has the largest variance percentage among low-order
interactions. Therefore, we obtain predictions for a new testing set H with
the following extrapolated ranges:
Tm ∈ [270, 280]
t ∈ [600, 750]
∆T ∈ [40, 50]
r ∈ [0.2, 0.25]; (5.14)
121
5.4. Heat Transfer in a Solid Sphere
size. On the other hand, the Full-DA has a good prediction accuracy in
Extrapolation with a rate of convergence of 1/n0.83 , going from an average
eN−RMSE of 0.07 at n = 70 to 0.01 at n = 560. As we can see, the Expanded
Full-DA has the best prediction accuracy in Extrapolation with a rate of
convergence of 1/n0.97 going from an average eN−RMSE of 0.06 at n = 70 to
0.007 at n = 560.
Thus, both Full-DA and Expanded Full-DA have good extrapolation
accuracy with n = 70 and excellent accuracy with n = 560. Compared
to Interpolation, where Full-DA with a linear regression component is the
best strategy, the Extrapolation case is well approached with Full-DA and
Expanded Full-DA with a constant regression term.
Constant Linear
100 0
31.63 −0.5
log10(eN−RMSE)
10 −1
eN−RMSE (%)
3.16 −1.5
1 −2
0.32 −2.5
122
5.5. Storm Surge Model
40.0
37.5
1
2
Latitude
3
35.0 5 4
8 7 6
10 9
32.5
30.0
Figure 5.10: Storm surge locations on the North Carolina coast plotted
with ggmap (Kahle and Wickham, 2013).
123
5.5. Storm Surge Model
124
5.5. Storm Surge Model
which is constant over time for all 324 storm tracks. Moreover, the fixed
coordinates in longitude and latitude for each location (λi , φi ) are used to
create derived inputs. They are:
∆p = fp − cp ∈ [0.7, 90.2].
RE = 6371 km.
125
5.5. Storm Surge Model
We obtain the forward storm speed, in m/s, with the equation (5.18) for
each one of the runs in the storm track by using the average of the speeds
travelled by the eye of storm computed from three days before landfall to
one day after. The range of this input, all over the 324 storm tracks, is
vf ∈ [2.42, 8.51].
Thus Di is positive if the ith location is at the east of the eye of the storm at
landfall and negative otherwise, which is important because it takes account
of the onshore versus offshore direction of the wind at the location. The
range of this input, over all the 324 storm tracks, is
Di ∈ [−258.82, 492.79].
We compute the storm track angles before and after landfall, θb and θa
respectively. We obtain the true bearing to a point, which is the angle mea-
sured in degrees between the vector formed by two pairs of coordinates and
the north line in clockwise direction. Both angles are signed, e.g., NE = 45◦
and NW = −45◦ . Thus, both inputs use the following pairs of coordinates:
(−3) (−3)
• Storm track angle before landfall uses coordinates λ0 , φ0 of the
eye of the storm three days before landfall (i.e., −3), and coordinates
(0) (0)
λ0 , φ0 at landfall (i.e., 0); with the following range across the 324
storm tracks:
θb ∈ [−34.09, 5.38].
126
5.5. Storm Surge Model
(0.02) (0.02)
• Storm track angle after landfall uses coordinates λ0 , φ0 half
(1) (1)
an hour after landfall (i.e., 0.02 days) and λ0 , φ0
one day after
landfall (i.e., 1); with the following range across the 324 storm tracks:
θa ∈ [−28.01, 30.09].
We also add the signed bearing angle of the ith location with respect to
the hurricane’s location at landfall in degrees, αi . We use the coordinates of
(0) (0)
the ith location λi , φi along with λ0 , φ0 . The input has the following
range across the 324 storm tracks:
αi ∈ [−177.05, 179.97].
●
5.5
●
5.0 ●
●
● ●
● ●
●
4.5 ●
●
● ●
● ● ●
●
● ● ●
● ●
●
● ●
4.0 ●
● ●
● ●
● ●
●
● ●
●
3.5 ●
(m)
3.0
● ●
●
ηmax
2.5
i
●
●
●
● ●
●
●
2.0 ● ●
●
●
●
● ●
● ● ●
●
1.5 ● ●
●
●
●
● ●
● ● ●
● ● ●
1.0 ●
●
●
●
●
●
●
● ●
●
●
●
● ●
● ●
●
●
● ●
● ●
●
● ●
●
●
●
●
●
0.5 ●
●
0.0
1 2 3 4 5 6 7 8 9 10
Location
We extract the maximum storm surges by location and storm track. Hence,
we obtain ten different datasets for the ten locations, each with 324 data
points from the 324 tracks. Figure 5.11 shows the maximum storm surge by
location, and we can see that larger surges occur at the southern locations.
127
5.5. Storm Surge Model
Locations 7 and 8 have the largest medians above 1 m with larger spreads
(locations 9 and 10 also have large spreads).
Figure 5.12 details the locations at landfall in blue for each storm track,
and we can see that the landfalls are to the south of the NC coast or in
South Carolina. We will restrict our subsequent studies to the larger-surge
locations 7, 8, 9, and 10. Appendix C provides additional results on the rest
of the locations.
38.0
37.0
1
2
36.0
Latitude
4
35.0 5
8 7 6
34.0 9
10
33.0
128
5.5. Storm Surge Model
Table 5.7: Inputs, dimensions, units, and ranges for the Storm Surge model.
129
5.5. Storm Surge Model
130
5.5. Storm Surge Model
50 8
45 ●
7
40
6
35
Variance Percentage
5 ●
30
25 4
20 ●
3
15 ●
● 2
●
●
10
1 ●
0 0
∆p D7 vf B α7 Rmax θa θb D7.∆p D7.α7 ∆p.α7 B.D7 vf.∆p vf.D7
50 6
45
5
40 ●
35
Variance Percentage
4 ●
30
25 3
●
20
2
15
10
1 ●
5
●
0 ●
0
∆p α8 B D8 vf Rmax θa θb ∆p.α8 D8.α8 D8.∆p B.α8 vf.α8 B.D8
131
5.5. Storm Surge Model
60 16
55
14
50
45 12
Variance Percentage
40
10
35 ●
30 8
25
6
20
15 4
10
2
5 ●
0 ●
0
D9 ∆p vf B Rmax α9 θa θb D9.∆p vf.D9 B.D9 Rmax.D9 vf.∆p vf.D9.∆p
50 14
45 ●
12
40 ●
35 10
Variance Percentage
30
8
25
6
20
●
15 4 ●
10
●
2
5 ●
0 ●
●
●
0
∆p α10 D10 vf B θa Rmax θb ∆p.α10 D10.∆p B.α10 vf.α10 Rmax.α10 θa.α10
132
5.5. Storm Surge Model
3
3
●
●
●
●
●
2 ● ●
● 2 ●
●
ηmax
ηmax
●
7
8
●
●
●
●
●
1 ● 1 ●
● ●
● ●
0 0
25 50 75 25 50 75
∆p ∆p
2.0 3
1.5 ●
●
●
●
●
2 ●
● ●
1.0
ηmax
ηmax
●
10
●
9
●
●
●
●
0.5 ● 1
●
●
●
●
●
●
0.0
0
25 50 75 25 50 75
∆p ∆p
Figure 5.15: Estimated main effect (red) of central pressure deficit (∆p )
on maximum storm surge (ηimax ) from FANOVA, using a Std GaSP with
PowExp correlation function and a constant regression term with 95% con-
fidence limits (green). One randomized dataset of n = 200 runs is used for
training the Std GaSP.
133
5.5. Storm Surge Model
The map in Figure 5.16 shows the 324 storm tracks over time coloured
by the levels of ∆p , along with landfall locations in blue. We can identify
two subsets of locations:
• On the other hand, the remaining six southern locations are within the
landfall coastal region associated with higher values of ∆p . They show
a larger median maximum storm surges across the 324 storm tracks.
Figure 5.16: Spatial distribution of the 324 storm tracks along with pressure
deficit (∆p ) levels.
134
5.5. Storm Surge Model
where γ is the specific weight of water and mxi is a constant for the ith
location obtained from linear regression.
In our subsequent DA approach we disregard γ (whose fundamental di-
mensions are ML−2 T−2 ) and mxi , since they are taken as constants. Thus,
by using ∆p as our first base quantity, the maximum storm surge in the DA
approach suggested by FANOVA will be:
ηimax
ηimax = . (5.19)
(F )
∆p
We are using the absolute value of this base quantity i.e., the distance with-
out a sign as typically defined, since the inputs need to be considered as
positive by the Buckingham’s Π-theorem.
The use of ηimax
(F )
as a dimensionless output along with five dimensionless
inputs (Rmax(F ) , B, θb , θa , and αi ) would suffice in an usual DA setup. The
135
5.5. Storm Surge Model
inputs θb , θa , and αi are dimensionless since they are angles. They can take
on negative values, even though neither of them is used as a base quantity.
Note that the forward storm speed vf was initially considered as the third
base quantity, and it is not used in this DA arrangement. However, given
the system’s incompleteness, we have to include it as another dimensionless
input: s
vf2
vf(F ) = , (5.21)
| Di | g
which is adjusted by the constant acceleration due to gravity g = 9.81m/s2
on the surface of Earth.
We encounter an analogous matter to vf with the base quantity ∆p , in
terms of its lack of use in the set of dimensionless inputs. Thus, we add
an input depicting a ratio between the two base quantities ∆p and Di as
follows:
∆p
with ∆p(F ) = ML−2 T−2 ,
∆p(F ) = (5.22)
| Di |
which can also be adjusted with the constant γ to make it dimensionless.
The inclusion of this input is necessary for our DA in order to be competitive
against a non-DA approach.
It is essential to highlight that all base quantities appear among the
dimensionless inputs in our two previous case studies, and ∆p along with vf
would not appear unless we include ∆p(F ) and vf(F ) respectively.
136
5.5. Storm Surge Model
137
5.5. Storm Surge Model
138
5.5. Storm Surge Model
324 storm tracks used to train the GaSP models are not uniformly around all
ten locations. Additionally in Figure 5.12, the tracks’ landfalls are centred
between locations 5 and 10. We would expect this accuracy variability by
location since most of the inputs are completely associated with the track
(except for Di and αi ).
We also encounter local matters in certain locations behind this accuracy
variability in terms of topology and bathymetry. A detailed map of locations
9 and 10 is shown in Figure 5.17. The map shows an encounter of a river and
the ocean for each location: the Cape Fear River in location 9 and the Little
River in location 10. Recall this is a complex system that is incomplete
from the dimensional point of view. These limitations have an impact on
how much we can improve the GaSP’s accuracy in this study. However, this
issue is of lesser degree for our extrapolation study in Section 5.5.8.
34.1
34.0
Latitude
33.9
9
10
33.8
33.7
Figure 5.17: Storm surge locations 9 and 10 on the North Carolina coast.
139
5.5. Storm Surge Model
Non−DA DA
0.300 ●
0.275
0.250
0.225 ●
eN−RMSE
●
●
0.200
0.175
0.150
0.125 ●
0.100
(a) Location 7.
Non−DA DA
0.225
0.200
● ●
●
0.175 ●
eN−RMSE
0.150
0.125
0.100
(b) Location 8.
Figure 5.18: Prediction accuracy for the Storm Surge model, with PowExp
correlation function and a constant regression term, with 25 randomized
datasets of n = 200 runs for training and N = 124 runs for testing.
140
5.5. Storm Surge Model
Non−DA DA
●
0.300
0.275
0.250 ●
●
●
0.225 ●
eN−RMSE
0.200 ●
0.175
0.150
0.125
●
●
●
0.100 ●
(a) Location 9.
Non−DA DA
0.40
0.35
0.30
●
eN−RMSE
0.25
●
0.20
0.15
0.10
Figure 5.19: Prediction accuracy for the Storm Surge model, with PowExp
correlation function and a constant regression term, with 25 randomized
datasets of n = 200 runs for training and N = 124 runs for testing.
141
5.5. Storm Surge Model
5.5.8 Extrapolation
The implementation of an extrapolation study has to be conducted differ-
ently. New data are not available to test prediction accuracy over extended
input ranges. Instead, within the fixed dataset, we define training data and
extrapolated test data based on an identified key input coming from our
previous FANOVA results: the central pressure deficit, ∆p . Subject-matter
expertise (Irish et al., 2009) indicates that this input is a measure of the
storm intensity, and FANOVA shows that it is one of the most important
main effects for the four southern locations.
The input ∆p is the criterion used for defining the extrapolated runs
(among the 324 existing data points for each location). To obtain our testing
set H, we sort the tracks in descending order by ∆p . Then, the testing set
H corresponds to tracks with values in the first 20% while the rest is used as
the training set D. Thus this is a severe test, extrapolating to more extreme
hurricane events. We fit a single GaSP with the training set D and obtain
predictions with the testing set H. We use the Log-Input strategy for both
Non-DA and DA.
Figure 5.20 illustrates the prediction accuracy results. The y-axis shows
the difference between each prediction of the maximum storm surge and the
model’s output (η̂imax − ηimax ) in the testing set H, where negative values
indicate an under-prediction of flood hazard. Unlike previous plots, good
prediction accuracy is determined by values around zero and avoiding large
negative errors. We indicate an accuracy threshold band on the y-axis be-
tween -0.10 and 0.10 (i.e., 10 cm error in predicted surge) with orange dotted
lines and a solid line on 0, for reference purposes.
Table 5.9 provides the summary statistics of the prediction errors of
flood hazard (in m) from Non-DA and DA for the southern locations. We
can summarize the results in the following points:
• DA clearly outperforms Non-DA in locations 7, 8, and 10.
• Location 7. The difference between Non-DA and DA is clear, where
we can see than Non-DA tends to underpredict storm surges when
compared to DA. The median of DA’s prediction errors is -0.01, which
is closer to zero than the one corresponding to Non-DA, -0.13. In
terms of spread, the IQR is smaller for DA (0.25) than for Non-DA
(0.34).
• Location 8. The medians behave like those for location 7: DA has
a better median (0.02) than Non-DA (-0.09). The IQR corresponding
to DA is slightly smaller than with Non-DA, 0.23 versus 0.27.
142
5.5. Storm Surge Model
Location 7
Approach Min Q1 Median Mean Q3 Max IQR
Non-DA -0.69 -0.31 -0.13 -0.13 0.04 0.39 0.34
DA -0.45 -0.12 -0.01 -0.01 0.13 0.48 0.25
Location 8
Approach Min Q1 Median Mean Q3 Max IQR
Non-DA -0.70 -0.24 -0.09 -0.08 0.04 0.40 0.27
DA -0.36 -0.09 0.02 0.02 0.14 0.41 0.23
Location 9
Approach Min Q1 Median Mean Q3 Max IQR
Non-DA -0.75 -0.16 0.00 -0.08 0.06 0.31 0.21
DA -0.57 -0.13 -0.01 -0.02 0.06 0.71 0.19
Location 10
Approach Min Q1 Median Mean Q3 Max IQR
Non-DA -0.69 -0.14 -0.03 -0.07 0.05 0.48 0.19
DA -0.30 -0.03 0.04 0.06 0.12 0.66 0.14
Table 5.9: Summary statistics of the prediction errors (η̂imax −ηimax ) of flood
hazard (in m) by Non-DA and DA for the Storm Surge model, with PowExp
correlation function and a constant regression term, with the top 20% runs
in ∆p for testing and the remainder for training.
143
5.5. Storm Surge Model
0.50 0.50
0.25 0.25
^ max − ηmax
^ max − ηmax
0.00 0.00
7
8
7
8
−0.25 −0.25
η
η
−0.50 −0.50
−0.75 −0.75
Non−DA DA Non−DA DA
0.8 0.8
●
0.4 0.4 ●
●
●
^ max − ηmax
^ max − ηmax
10
9
0.0 0.0
10
9
η
●
●
●
−0.4 ●
−0.4
●
●
● ●
●
● ●
●
●
●
−0.8 −0.8
Non−DA DA Non−DA DA
Figure 5.20: Prediction error of flood hazard (in m) for the Storm Surge
model, with PowExp correlation function and a constant regression term,
with the top 20% runs in ∆p for testing and the remainder for training.
144
5.6. Concluding Remarks
145
5.6. Concluding Remarks
146
Chapter 6
147
6.1. Summary of the Thesis
148
6.1. Summary of the Thesis
• The Borehole function provides the ideal framework for applying the
Buckingham’s Π-theorem, where the dimensional system is complete
since it depicts all the inputs involved in the determination of the
output. Therefore, the application of a full DA and logarithmic input
transformations with further input expansion provides highly accurate
interpolated and extrapolated predictions (compared to a non-DA ap-
proach and another DA found in the literature).
• Finally, the Storm Surge model fully exploits the use of a DA approach.
As mentioned above, FANOVA is important in identifying those in-
puts considered as base quantities in the system. This example is of
particular attention since it is an extremely complex case study, where
the experimenter does not know all the inputs involved in the process.
Thus, the DA is considered full in the sense that we manage to obtain
dimensionless positive inputs when possible. Recall there are three
bearing angles that are already dimensionless, but they can take on
negative values. Of most practical importance, our DA performs best
149
6.2. Discussion and Future Work
150
6.2. Discussion and Future Work
tions in Chapter 4 with the TAAG Process: the 2-dimensional Franke (4.12)
and the OTL circuit (4.19) functions using a single replicate of training size
n = 10d.
In order to compare the performance of our non-standard correlation
functions against the TAG or TAAG Processes found in the R package TAG
(Lin and Joseph, 2020), each one of the case studies in Chapters 3 and 4 was
checked under the same training and testing conditions used for the other
GaSP approaches. As done by the authors, the starting Std GaSP in the
TAAG Process algorithm was obtained with the R package DiceKriging
(Roustant et al., 2012). We obtained the following results:
Table 6.1: Summary statistics of eN−RMSE s (2.14) by type of GaSP for the
OTL circuit function (4.19). Each figure summarizes the results from 20
mLHDs of n = 240 runs for training and N = 10, 000 runs for testing. Std
GaSP is implemented with PowExp. TAAG, E-JE2SR, and U-JE3 GaSPs
are implemented with SqExp.
• In the case of the OTL circuit function, Table 6.1 provides the sum-
mary statistics of the eN−RMSE s (2.14) by GaSP approach at n = 240
from our 20 training designs. We can see that the median corre-
sponding to the Std GaSP is 82% smaller than the one obtained with
TAAG. Moreover, Figure 6.1 illustrates the prediction accuracy in per-
cent (eN−RMSE ×100%) on the log10 (n)-log10 (eN−RMSE ) scale of TAAG
across the training sizes n = 60, 120, 240 and all the values are above
the rest of the GaSP approaches.
151
6.2. Discussion and Future Work
• The Nilson-Kuusk model also shows that a TAAG Process has a poor
prediction accuracy compared to the rest of the GaSP approaches.
10 −1
3.16 −1.5
log10(eN−RMSE)
eN−RMSE (%)
1 −2
0.32 −2.5
0.1 −3
0.03 −3.5
60 120 240
n
152
6.2. Discussion and Future Work
• Wilson and Adams (2013) introduce new base kernels, different from
the SqExp case, that allow pattern discovery. These kernels can ex-
trapolate accurate predictions in time series.
153
6.2. Discussion and Future Work
• On the other hand, using the ADCIRC model to generate the storm
surge data is just a single computational approach to deal with this
class of natural phenomenon. Other approaches include the Weather
Research and Forecasting (WRF) model, Wave Model (WAM), and
Simulating Waves Nearshore (SWAN) numerical model. We can find
works in the literature comparing the prediction accuracy of these
models (Bhaskaran et al., 2013). Since DA is a data pre-processing
tool, it can also be applied along with a surrogate GaSP with these
other models.
154
Bibliography
Abt, M. (1999). Estimating the prediction mean squared error in Gaussian
stochastic processes with exponential correlation structure. Scandinavian
Journal of Statistics, 26(4):563–578.
Ba, S. and Joseph, V. R. (2012). Composite Gaussian process models for em-
ulating expensive functions. The Annals of Applied Statistics, 6(4):1838–
1860.
Bhaskaran, P. K., Nayak, S., Bonthu, S. R., Murty, P., and Sen, D. (2013).
Performance and validation of a coupled parallel ADCIRC-SWAN model
for THANE cyclone in the Bay of Bengal. Environmental Fluid Mechanics,
13:601–623.
Box, G. E. P., Hunter, W. G., and Hunter, J. S. (1978). Statistics for Exper-
imenters: An Introduction to Design, Data Analysis, and Model Building.
Wiley Series in Probability and Mathematical Statistics: Applied Proba-
bility and Statistics. Wiley.
155
Bibliography
Chapman, W. L., Welch, W. J., Bowman, K. P., Sacks, J., and Walsh, J. E.
(1994). Arctic Sea ice variability: Model sensitivities and a multidecadal
simulation. Journal of Geophysical Research, 99C:919–935.
Chen, H., Loeppky, J. L., Sacks, J., and Welch, W. J. (2016). Analysis meth-
ods for computer experiments: How to assess and what counts? Statistical
Science, 31(1):40–60.
156
Bibliography
157
Bibliography
Jones, D. R., Schonlau, M., and Welch, W. J. (1998). Efficient global op-
timization of expensive black-box functions. Journal of Global Optimiza-
tion, 13(4):455–492.
Joseph, V. R., Hung, Y., and Sudjianto, A. (2008). Blind kriging: A new
method for developing metamodels. Journal of Mechanical Design, 130(3).
031102.
Kalaitzis, A., Honkela, A., Gao, P., and Lawrence, N. D. (2014). gptk:
Gaussian Processes Tool-Kit. R package version 1.08.
158
Bibliography
159
Bibliography
160
Stein, M. L. (1999). Interpolation of Spatial Data: Some Theory for Kriging.
Springer Series in Statistics. Springer New York.
Westerink, J., Luettich, Jr, R., Feyen, J., Atkinson, J., Dawson, C., Roberts,
H., Powell, M., Dunion, J., Kubatko, E., and Pourtaheri, H. (2008). A
basin to channel-scale unstructured grid hurricane storm surge model ap-
plied to southern Louisiana. Monthly Weather Review, 136:833–864.
161
Appendix A
Correlation Structures
We provide a summary of all the correlation structures used throughout
this thesis for easy reference. Let x and x0 be two input vectors in Rd , the
standard (Std) correlation structure is
d
Y
RStd (x, x0 ) = Rj (hj ) ∈ [0, 1] (A.1)
j=1
where Rj (hj ) is a chosen correlation function (see Section 2.3), and depends
on the distance hj = xj − x0j for the jth input.
Note that this structure assigns the equal weight of 1/d to each one of the
d main effects.
The main-effect correlation structure with unequal weights (U-ME) is
defined as:
d
X
RU−ME (x, x0 ) = ωj Rj (hj ) ∈ [0, 1], (A.3)
j=1
162
A.2. Joint-Effect Correlation Structures Up to All 2-Input Interactions
where
exp (τi )
ωi = Pd−1 for i = 1, . . . , d − 1;
1 + k=1 exp (τk )
and
1
ωd = Pd−1 .
1+ k=1 exp (τk )
Note that the two types of terms—main effects and 2-input interactions—
each receive weight 1/2, with the 1/2 distributed equally across all the cor-
relation functions in the type of term.
The joint-effect correlation structure with unequal weights using all 2-
input interactions (U-JE2) is defined as:
d d d
U−JE2 0λ1 X λ2 X X
R (x, x ) = Rj (hj ) + d Rj (hj ) · Rk (hk ) ∈ [0, 1],
d 2
j=1 j=1 k=j+1
(A.5)
τ ∈ (−∞, ∞),
163
A.3. Joint-Effect Correlation Structures up to Selected 2-Input Interactions
where λ1 is computed as
exp (τ )
λ1 = ;
1 + exp (τ )
and λ2 is defined as
1
λ2 = .
1 + exp (τ )
(A.6)
(A.7)
164
A.4. Joint-Effect Correlation Structures with a Residual Effect Term
This correlation structure resembles the structure RE−JE2 (x, x0 ), but there
is an additional third addend RjRes (hj ) representing the residual component
under a standard multiplicative structure.
Under equal weights on the effect framework, a joint-effect correlation
structure with selected 2-input interactions and a residual effect term (E-
JE2SR) is set up in the following way:
d
1 X 1 X
RE−JE2SR (x, x0 ) = Rj (hj ) + Rj (hj ) · Rk (hk )
3d 3b
j=1 {j,k}∈B
d
1 Y
+ RjRes (hj ) ∈ [0, 1], (A.9)
3
j=1
On the other hand, for the PowExp correlation function, a residual compo-
nent becomes:
d d d
" #
Y PowExp
Y p 0 X p0
RjRes (hj , θ0 , p0 ) = exp(−θ0 hj ) = exp − θ0 hj .
j=1 j=1 j=1
165
A.5. Joint-Effect Correlation Structures Up to All 3-Input Interactions
166
A.6. Joint-Effect Correlation Structures up to Selected 3-Input Interactions
167
A.6. Joint-Effect Correlation Structures up to Selected 3-Input Interactions
168
Appendix B
308 315
● ● ●
●
●
●
●
310
●
306 ●
●
●
●
●
305 ●
Ts
Ts
304 ●
●
300
●
●
302
●
●
295
●
●
300 290
0.0 0.2 0.4 0.6 0.8 1.0 0.050 0.075 0.100 0.125 0.150 0.175 0.200
R r
(a) Ratio of distance from center and (b) Radius of sphere (r).
sphere radius (R).
Figure B.1: Estimated main effect (red) on temperature of sphere (Ts ) from
FANOVA, using a Std GaSP with PowExp correlation function and a con-
stant regression term with approximate 95% confidence limits (green) for
the Solid Sphere function (5.11). One mLHD of n = 70 runs is used for
training the Std GaSP.
169
Appendix B. Heat Transfer in a Solid Sphere
●
320 ●
320
●
●
315
●
315
●
310
●
310 ●
●
305 ●
Ts
Ts
305
●
●
300
●
●
300
●
●
295
●
●
295
●
290 ●
●
290
0 100 200 300 400 500 600 240 245 250 255 260 265 270
t Tm
Figure B.2: Estimated main effect (red) on temperature of sphere (Ts ) from
FANOVA, using a Std GaSP with PowExp correlation function and a con-
stant regression term with approximate 95% confidence limits (green) for
the Solid Sphere function (5.11). One mLHD of n = 70 runs is used for
training the Std GaSP.
170
Appendix B. Heat Transfer in a Solid Sphere
320 307
315
●
●
306
●
●
310 ●
●
● ●
●
Ts
Ts
305 ● 305
●
●
●
● ●
300
●
●
●
304
●
●
295
●
290 303
50 55 60 65 70 75 80 30 40 50 60 70 80 90 100
∆T k
(a) Initial sphere temperature minus tem- (b) Thermal conductivity (k).
perature of medium (∆T ).
Figure B.3: Estimated main effect (red) on temperature of sphere (Ts ) from
FANOVA, using a Std GaSP with PowExp correlation function and a con-
stant regression term with approximate 95% confidence limits (green) for
the Solid Sphere function (5.11). One mLHD of n = 70 runs is used for
training the Std GaSP.
171
Appendix B. Heat Transfer in a Solid Sphere
308
●
●
307
306
●
●
Ts
305
●
304
●
303 ●
302
Figure B.4: Estimated main effect (red) of convective heat transfer coeffi-
cient (hc ) on temperature of sphere (Ts ) from FANOVA, using a Std GaSP
with PowExp correlation function and a constant regression term with ap-
proximate 95% confidence limits (green) for the Solid Sphere function (5.11).
One mLHD of n = 70 runs is used for training the Std GaSP.
172
173
Appendix C. Storm Surge Model
Appendix C
50 10
●
45
40 8
35
Variance Percentage
30 6
25
20 4
15
10 2
●
5 ●
0 0
α1 ∆p Rmax θa θb B D1 vf ∆p.α1 B.α1 Rmax.∆p θa.α1 Rmax.α1 θb.α1
(a) Location 1.
50
●
6
45
40
●
35
Variance Percentage
30 4
25 ●
20
15 ●
2
10 ●
●
5 ●
0 0
α2 ∆p Rmax D2 θa B θb vf ∆p.α2 Rmax.∆p B.α2 Rmax.α2 θa.α2 Rmax.B
(b) Location 2.
174
Appendix C. Storm Surge Model
60 10
55 9
50
8
45
7
Variance Percentage
40 ●
35 6
30 5
25 4
20 ●
3
15
2 ●
10
●
5 1
● ●
● ●
0 ●
●
●
0
D3 Rmax ∆p vf θa α3 B θb D3.∆p Rmax.D3 Rmax.∆p B.D3 Rmax.B.D3 Rmax.D3.∆p
60 20
55 18 ●
50
16
45
14
Variance Percentage
40
●
35 12
30 10
25 8
●
●
20
6
15 ●
4
10 ●
5 2
●
● ●
●
0 ●
●
0
D4 ∆p α4 Rmax vf B θa θb D4.∆p Rmax.∆p B.D4 Rmax.D4 D4.α4 Rmax.α4
175
Appendix C. Storm Surge Model
60 16
55
14
50
●
45 12
●
Variance Percentage
40 ●
●
10
35
30 8
25
6
20
15 4
10
2 ●
5 ●
●
●
●
●
●
●
● ●
0 ● ●
0
D5 ∆p vf B Rmax α5 θa θb D5.∆p B.D5 Rmax.D5 Rmax.∆p vf.D5 D5.α5
50 10
45 9
●
40 8
35 7
Variance Percentage
30 6
25 5
20 ●
4
15 3
● ●
10 2 ●
5 ●
1
0 ●
0
∆p D6 vf B α6 Rmax θa θb D6.∆p vf.∆p vf.D6 D6.α6 B.D6 Rmax.D6
176
Appendix C. Storm Surge Model
1.00 0.8
● ●
● ●
0.75 ●
0.6
●
●
● ●
●
●
0.50 0.4
ηmax
ηmax
●
1
2
●
●
●
●
0.25 ● 0.2 ●
●
●
●
●
0.00 0.0
25 50 75 20 40 60 80
∆p ∆p
0.8
0.4
●
●
● 0.6 ●
0.3 ● ●
●
●
●
ηmax
ηmax
●
0.4
3
0.2 ●
●
●
●
●
●
●
● ●
0.2 ●
0.1 ●
0.0 0.0
40 60 80 0 25 50 75
∆p ∆p
Figure C.4: Estimated main effect (red) of central pressure deficit (∆p ) on
maximum storm surge (ηimax ) from FANOVA, using a Std GaSP with Pow-
Exp correlation function and a constant regression term with 95% confidence
limits (green). One randomized dataset of n = 200 runs is used for training
the Std GaSP.
177
Appendix C. Storm Surge Model
1.5 1.5
●
●
●
● ●
1.0 1.0
● ●
ηmax
ηmax
● ●
5
● ●
● ●
0.5 0.5
●
● ●
● ●
●
●
● ●
0.0 0.0
0 20 40 60 80 25 50 75
∆p ∆p
Figure C.5: Estimated main effect (red) of central pressure deficit (∆p ) on
maximum storm surge (ηimax ) from FANOVA, using a Std GaSP with Pow-
Exp correlation function and a constant regression term with 95% confidence
limits (green). One randomized dataset of n = 200 runs is used for training
the Std GaSP.
178
Appendix C. Storm Surge Model
Non−DA DA
●
0.50
●
0.45 ● ●
0.40 ●
●
● ●
● ●
●
0.35
eN−RMSE
0.30
0.25
●
●
●
0.20
0.15
0.10
(a) Location 1.
Non−DA DA
0.70
●
0.65 ●
0.60
0.55
0.50
●
0.45
●
eN−RMSE
●
● ●
0.40
0.35 ●
0.30
0.25
0.20
0.15
0.10
(b) Location 2.
Figure C.6: Prediction accuracy for the Storm Surge model, with PowExp
correlation function and a constant regression term, with 25 randomized
datasets of n = 200 runs for training and N = 124 runs for testing.
179
Appendix C. Storm Surge Model
Non−DA DA
0.55
●
0.50
0.45
●
0.40
●
eN−RMSE
0.35
●
●
0.30
0.25
0.20
0.15
0.10
(a) Location 3.
Non−DA DA
●
0.50
●
●
0.45
●
0.40
●
0.35
●
eN−RMSE
0.30 ●
●
●
●
0.25
0.20
0.15
0.10
(b) Location 4.
Figure C.7: Prediction accuracy for the Storm Surge model, with PowExp
correlation function and a constant regression term, with 25 randomized
datasets of n = 200 runs for training and N = 124 runs for testing.
180
Appendix C. Storm Surge Model
Non−DA DA
0.45
0.40
0.35
●
●
0.30 ●
●
eN−RMSE
0.25 ●
0.20
0.15
0.10
(a) Location 5.
Non−DA DA
0.40
●
0.35 ●
0.30
eN−RMSE
0.25
●
0.20
0.15
0.10
(b) Location 6.
Figure C.8: Prediction accuracy for the Storm Surge model, with PowExp
correlation function and a constant regression term, with 25 randomized
datasets of n = 200 runs for training and N = 124 runs for testing.
181