Naser 2024 Integrating Machine Learning Models Into Building Codes and Standards Establishing Equivalence Through

You might also like

Download as pdf or txt
Download as pdf or txt
You are on page 1of 14

Integrating Machine Learning Models into Building Codes

and Standards: Establishing Equivalence through


Engineering Intuition and Causal Logic
M. Z. Naser, Ph.D., P.E., M.ASCE 1
Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

Abstract: The traditional approach to formulating building codes often is slow and labor-intensive, and may struggle to keep pace with the
rapid evolution of technology and domain findings. Overcoming such challenges necessitates a methodology that streamlines the mod-
ernization of codal provisions. This paper proposes a machine learning (ML) approach to append a variety of codal provisions, including
those of empirical, statistical, and theoretical natures. In this approach, a codal provision (i.e., equation) is analyzed to trace its properties
(e.g., engineering intuition and causal logic). Then a ML model is tailored to preserve the same properties and satisfy a collection of
similarity and performance measures until declared equivalent to the provision at hand. The resulting ML model harnesses the predictive
capabilities of ML while arriving at predictions similar to the codal provision used to train the ML model, and hence it becomes possible to
use in lieu of the codal expression. This approach was examined successfully for seven structural engineering phenomena contained within
various building codes, including those in North America and Australia. The findings suggest that the proposed approach could lay the
groundwork for implementing ML in the development of future building codes. DOI: 10.1061/JSENDH.STENG-12934. © 2024 American
Society of Civil Engineers.
Author keywords: Building codes; Machine learning; Prediction.

Introduction Many codal guidelines are communicated via equations (formu-


las). These formulas display the key parameters and the relationship
Building codes are guiding documents that house rules and recom- by which such parameters are tied to a specific phenomenon. For
mendations aimed to standardize civil engineering guidelines and example, in Eq. (1), the moment capacity of compact W-shaped
processes pertaining to design, analysis, construction, and so forth steel beams is a function of the plastic section modulus, Z, and the
(ASCE 2016). Such codes remain the cornerstone of civil engineer- yield strength, f y , of the structural steel. In this format, an engineer
ing, providing a framework to establish building occupants’ safety, can simply predict and/or evaluate the expected moment capacity of
health, and general welfare. a given beam via Eq. (1). More importantly, an engineer also may
The process of creating codal provisions begins with collecting causally infer that an increase in either (or both) parameters can
and analyzing data, which often are gathered from past building increase the moment capacity of the beam (and vice versa)
failures, tests, simulations, and empirical or practical experiences
gained from well-documented incidents. All such data are system- M ¼ fy × Z ð1Þ
atically archived, studied, and evaluated to understand and learn whereas some formulas stem from the deterministic nature rooted in
from them to avoid repeating past failures (Ellingwood 1994). This physics and engineering principles, others are probabilistic, empiri-
information and incorporated into the building codes. For example, cal, or arrived at via statistical methods (Lie 1992; Marasco et al.
the International Building Code (IBC) (IBC 2018) stipulates the 2021). Despite the origin of codal provisions, variants of such pro-
minimum standards for various aspects, such as structural integrity visions exist in building codes. Building committees have been noted
and fire safety. By complying with such codes, engineers can min- to accept such variants, because without these committees, provi-
imize the risk of building collapse, fire hazards, and other potential sions would not have been accepted and included in building codes.
accidents. Thus, adherence to codal provisions is not merely vol- In recent years, machine learning (ML) has shown great success
untary but is a legal obligation and ethical responsibility (Law in many fields, including civil engineering. ML models can analyze
2021). The literature presents various cases in which noncompli- vast data sets, identify complex patterns, and make accurate predic-
ance with building codes resulted in catastrophic loss of life and tions. The existing literature includes studies wherein predictions
property (Windapo and Cattell 2010; Yakubu 2019). from ML are reported to exceed those obtained from national and
international building codes (Aladsani et al. 2022; Cakiroglu et al.
1
Assistant Professor, School of Civil and Environmental Engineering 2022; Degtyarev and Tsavdaridis 2022; Naser 2023a; Tarawneh
and Earth Sciences, Artificial Intelligence Research Institute for Science et al. 2021; Yahiaoui et al. 2023). This leads to the question of how
and Engineering (AIRISE), College of Engineering, Computing and we can leverage such models.
Applied Sciences, Clemson Univ., 312 Lowry Hall, Clemson, SC 29634. However, ML models differ significantly from codal provisions.
ORCID: https://orcid.org/0000-0003-1350-3654. Email: mznaser@clemson
At the moment, most ML models remain driven by the supplied
.edu
Note. This manuscript was submitted on June 29, 2023; approved on data. For example, if a building code specifies an equation to cal-
December 11, 2023; published online on March 6, 2024. Discussion per- culate wind loads based on the height and location of a building, a
iod open until August 6, 2024; separate discussions must be submitted for ML model first would be trained to use a database of historical wind
individual papers. This paper is part of the Journal of Structural Engi- data and building performance to predict the same phenomenon
neering, © ASCE, ISSN 0733-9445. easily. Although ML models may excel at making predictions based

© ASCE 04024039-1 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


on the patterns in the training data that may surpass predictions An Examination of Codal Provisions
obtained from codal provisions, the same models can struggle when
confronted with scenarios or data that significantly deviate from Codal provisions often undergo a series of iterations and must with-
their training set (Krueger et al. 2021). stand rigorous discussions and voting by regulatory bodies, commit-
Furthermore, the dynamic nature of ML models, which often are tee members, and the general public. Such provisions commonly
updated and changed as they learn from new data, contrasts sharply are developed from first principles (i.e., mechanics). If first princi-
with the static nature of building codes (which are updated, but at a ples are absent or immature, provisions also can be derived through
much more infrequent rate). For example, the specifics of the ML statistical treatment of the outcome of physical tests or, to a lesser
model (algorithmic architecture, hyperparameters, and so forth) used extent, numerical simulations.
to predict a specific phenomenon may change significantly over time In physical tests, a series of specimens is designed with preset
as the model is fed new data. This may present a potential issue for features identified by various means (including empirically) and
engineers who rely on consistency in their design processes (Paleyes tested to generate an understanding that ties the examined features
et al. 2022). to the phenomenon at hand (i.e., flexural response). The outcome of
Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

Perhaps of most importance is this lack of transparency of some such an investigation then is analyzed via statistical regression. To
ML models. This opaqueness can pose difficulties when attempting a certain extent, civil engineers favor using linear and nonlinear
to integrate ML models into building codes, because these codes regression (Benjamin and Cornell 2014). Both forms of regression
need to have clearly defined and understandable criteria so that en- comprises a number of assumptions that need to satisfy the data
gineers can follow and verify compliance (Naser and Ross 2022). at hand as well as those pertaining to preselected parametric forms
Although the data-driven nature of ML is its strength, the value of (e.g., the use of linear regression implies the validity of linear and
engineering principles, intuition, and domain knowledge cannot be additive nature of features, and so forth) some of which may or may
overlooked—especially in our field, in which decisions can impact not be satisfied.
occupants’ safety (Naser 2021a). Evidently, a proper statistical regression analysis boils down
Understandably, engineers may remain skeptical of ML models to the best coefficients that can improve the accuracy of the
due to the aforementioned challenges and the absence of a regula- preselected parametric form from the available tests or simulations
tory framework that allows such models to be included in building data (Ribarič et al. 2008). Such an expression is treated further to
codes. This absence of a framework means there are no defined attain the expected reliability measures inherently required by a
standards to follow. This lack of standards makes it difficult for building code’s technicalities. When such an expression is fully
engineers to fully trust and adopt ML models. For example, engi- verified, vetted, and voted upon, this expression is likely to be in-
neers might be wary of using a ML model if there is not a univer- corporated into a guiding document. This procedure has been duly
sally accepted standard for such ML models. Given the absence of noted in the literature, and one of the recent examples is that de-
a structured way to integrate ML models into the design and con- veloped by Kuchma et al. (2019), credited with updating the one-
struction processes outlined by the building codes, engineers often way shear design provisions for RC in the latest ACI 318 code
opt for traditional, code-approved methods—even in scenarios (ACI 2019) adopted by the American Concrete Institute.
requiring performance-based principles or out-of-the-box solutions. Depending on the comprehensiveness of a testing campaign, it
On a more positive note, the rapid advancements in ML have is quite likely that additional tests or numerical simulations can be
allowed the integration of eXplainable AI (XAI), which can added to supplement the available data. For example, Franssen and
help create more reliable and interpretable models that could be Dotreppe (2003) followed a similar approach when developing
adopted into building codes. This concept refer to methods and the basis for Eurocode 2’s assessment method for the fire resistance
techniques that allow users (i.e., engineers) to understand ML of columns. Similarly, Wade et al. (1997) combined tests with
models’ decision-making processes (Emmert-Streib et al. 2020). numerical simulations to derive an equation to evaluate the fire re-
For example, if a ML model could explain that it predicts a specific sistance of RC columns for the Australian building code AS 3600
outcome (e.g., sectional capacity) based on specific characteristics (Standards Australia 1994). Other examples were presented by
of the building design (such as material properties and geometric Feldman and Bartlett (2005) and Ollgaard et al. (1971).
features), engineers would be more confident integrating these in- This paper argues that, with the presence and/or absence of
sights into their work than if the ML was a traditional black-box first principles, and when the primary motivation behind utilizing
model. Simply, if a ML model can explain its decisions in a way statistical-based expressions in building codes is their capacity to
that aligns with the stipulations of codal provisions, this could sig- best fit existing data, it is of merit to utilize higher-order models
nificantly enhance the potential for ML integration into these codes. (such as tree-based and neural network models, and so forth) that
Providing transparency and a deeper understanding of the under- are capable of best fitting the data in a more accurate manner than
lying relationships in the data can bridge the gap between the dynamic, that provided by the traditional counterparts. Noting how such mod-
data-driven world of ML and the static, calmer nature of building els can better handle high-dimensional data, has nonparametric free-
codes. This paper proposes such a methodology. In this methodology, dom, and less adherence to data related assumptions enables this
the first stage analyzes a codal equation to trace its properties method to create suitable codal provisions.
(e.g., logic). Then a ML model is tailored to learn and map the same A supporting philosophical view is that, for the same data and
logic—thereby transforming into a higher-order model that preserves features, the statistical-based procedure of iteratively fitting a codal
the same properties of the codal equation. Finally, the ML model is expression by best fitting the coefficients of its predefined form can
tuned on big data to achieve superior performance, given the ability of be viewed similarly to training a ML model to learn patterns in the
ML to accommodate highly nonlinear relations. Our findings indicate available data (Fig. 1). A primary difference is that the available
that the proposed approach could set the stage for ML implementation data tune the coefficients for the already-defined parametric form
in future building codes. Although significant research and develop- of the former and the algorithmic logic (internal model structure) for
ment are needed to refine these techniques and demonstrate their reli- the latter. One then can view the parametric form to be equivalent to
ability and effectiveness in the context of civil engineering and the the algorithmic logic because both connect the key parameters to
involvement of regulatory and code-developing organizations, this realize the prediction. However, a critical difference between the
work and its findings could be an important step toward such an effort. two approaches is that the procedure of statistical regression is

© ASCE 04024039-2 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


Fig. 1. Simplified illustration depicting two approaches for prediction. Additional steps could be involved (i.e., training, fitting, and so forth) in a
prediction problem.

familiar, and hence the steps that are carried out to interpret the out- 1X n

come of such analysis can be traced. As is explained in the section PDPXS ¼ fðX S ; X Ci Þ ð2Þ
n i¼1
“Proposed Methodology,” recent advances in explainability have
Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

made this possible for ML (Molnar 2019). where X Ci = remaining features for ith observation; and n = num-
ber of observations.
The key insight of PDPs is that by averaging out the effects of all
Proposed Methodology
other features, this tool can isolate the effect of the feature of in-
terest. In this way, the relationship between the target and a feature
General Description can be revealed. Fig. 2 illustrates a hypothetical PDP obtained from
Suppose the same logic and properties of an empirical codal pro- a ML model and from a physical equation. This PDP shows how
vision (i.e., equation) are imposed on a superior high-dimensional changing a specific feature while keeping all other features constant
ML model. In this case, such a model can be considered to be a affects the predicted target. The hypothetical model’s PDP has a
natural substitute for the statistically derived equation, because the similar trend and behavior, to some extent, to that of the hypotheti-
model can arrive at predictions and interpret their outcomes sim- cal equation. This implies that model and equation predictions are
ilarly to the codal equation. within certain a certain bound (i.e., percentage).
This is of importance because, in reality, although many of the
phenomena addressed by civil engineers may not conform fully to Equivalence through Accumulated Local Effects
the assumptions inherent in statistical methods used to derive em-
Although PDPs provide valuable insights, they can be biased when
pirical expressions, engineers still apply domain knowledge to as-
features are correlated, because PDPs treat features independently.
sign the position and interaction between the involved parameters
ALEs also can be used to explain ML models. However, unlike
fitted into such expressions. Thus, the proposed methodology starts
PDPs, ALEs provide a more accurate representation when features
by constraining the algorithmic structure of a candidate ML model
are correlated. For example, ALEs provide a way to visualize fea-
to one that imitates the same characteristics (in terms of engineering
ture effects by taking into account the average effects of other fea-
logic, domain knowledge, and so forth) and the essence of the codal
tures. Hence, ALE plots can be especially useful when there are
provision.
interactions between features. ALEs handle interactions between
This process is defined herein as model equivalence. Although
features by considering only the local effects; hence, they are not
there could be multiple ways to establish equivalence (as is de-
biased by global correlations as are PDPs. They also are centered
tailed in the section “Additional Thoughts, Limitations, and Future
around the average prediction over the training data, which makes
Directions”), this work establishes equivalence by leveraging two
them easier to interpret in terms of the average prediction.
eXplainable artificial intelligence methods, namely partial depend-
The ALE of a feature is calculated by accumulating the local
ence plots (PDPs) and accumulated local effects (ALEs), to con-
effects (i.e., the difference between the prediction for a slightly
strain the algorithmic structure to yield PDPs or ALEs similar to
higher value of the feature and the prediction for a slightly lower
those obtained from the codal equation (to a prespecified degree).
value for the same feature, averaged over the data set). To calculate
Such a degree is outlined via several similarity measures, five of
the ALE, the feature of interest is divided into intervals. For each
which are presented subsequently. Finally, when equivalence is
established, model predictions are compared with those obtained
from the codal equation via statistical measures. This section starts
by covering the principles of PDPs and ALEs and then addresses
the similarity measures selected to establish equivalence.

Equivalence through Partial Dependence Plots


A PDP explains ML models by depicting an individual feature’s (or
several features’) marginal effect on the prediction of a ML model
while holding other features constant (Friedman 2001). The same
plot can be used to infer the type of relationship between each fea-
ture and the target. A PDP is created by varying the feature of in-
terest across its range, and for each value, a prediction is made
using the ML model or codal equation, holding all other features
constant at their average values. The average prediction then is
plotted against the values of the feature of interest to show the mar-
ginal effect of that feature on the predicted outcome. Mathemati-
Fig. 2. Demonstration of PDPs. The target and feature are assumed to
cally, the partial dependence of a model f with respect to a set of
be unitless for demonstration purposes.
features X S can be defined as

© ASCE 04024039-3 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


interval, the difference in the model’s predictions at the upper and However, the proposed approach ensures that a ML will realize
lower bounds of the interval is computed while keeping other fea- matching PDPs and ALEs and high predictive accuracy.
tures constant. This difference then is accumulated across the inter- Specifically, the proposed approach entails the following steps:
vals to obtain the ALE plot. Mathematically, the ALE for a feature xi 1. An initial ML model is trained and the PDP and ALE are cal-
can be expressed as culated for each feature.
2. The differences between the PDPs and ALEs of the model and
1X n those of the equation the engineer wants the model to match are
ALExi ¼ ½fðxi;jþ1 ; X −i Þ − fðxi;j ; X −i Þ ð3Þ computed.
n i¼1
3. The calculated differences are used to adjust the target variable
by iterating over each unique value in a feature and adjusting the
where xi;j = value of feature xi for jth observation; and X -i = target variable for all instances with a feature value in a certain
remaining features. interval.
The following example compares PDPs and ALEs. A PDP might 4. The adjustment is calculated as the product of the learning
rate (calculated using a decay formula ¼ learning rate=ð1 þ
Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

show how the predicted sectional capacity changes with the cross-
sectional area while averaging out the effect of the material strength. decay rate × iterationÞ and the difference in PDPs or ALEs, and
If there is a strong interaction between the cross-sectional area and a sensitivity parameter that represents how sensitive the target is
the material (e.g., if changing the material significantly affects how to changes in this feature. The adjustment is calculated as follows:
the cross-sectional area influences the capacity), this interaction ef-
fect could be hidden in a PDP. Conversely, suppose that there is an adjustmentpdp ¼ learning ratepdp × pdp difference½i
interaction between the cross-sectional area and the material of the × sensitivitypdp
load-bearing member. An ALE plot can capture this interaction by
showing the effect of changing the cross-sectional area on the pre-
adjustmentale ¼ learning rateale × ale difference½i
dicted capacity, considering the material feature’s average effect. For
visualization purposes, both PDP and ALE are quite similar. × sensitivityale

where learning ratepdp and learning rateale = learning rates for


Detailed Description of the Proposed Approach PDP and ALE, respectively, calculated using a decay formula,
The key challenge herein is that both PDPs and ALEs are generated which measures how quickly the model should adjust its pre-
only when model training is completed. In other words, these tools dictions based on the differences; pdp difference½i and ale
cannot be manipulated during the training process, and there is no difference½i = absolute difference between model’s PDP or ALE
direct way to train ML models to yield PDPs and ALEs that fit and equation’s PDP or ALE, respectively, for the current feature
some criteria. Similarly, the traditional way of tuning model param- value; and sensitivitypdp and sensitivityale = sensitivities of output
eters (i.e., hyperparameters) to improve model accuracy cannot be to the feature for the PDP and ALE, respectively. These measure
applied herein to arrive at prespecified PDPs and ALEs. Simply how much the output changes with respect to a small change in
adjusting model parameters directly to match the PDPs and ALEs the feature. The adjustments are then applied to the target values
of the physical equation can be quite challenging because PDPs and (FR) in the training data where the feature value is between the
ALEs are, as mentioned previously, not connected directly to the current value and the next value.
model’s parameters. Therefore, this challenge is inherently nontri- 5. This adjustment is added to the target variable for all instances in
vial, and involves creating a loss function that incorporates the dis- which the feature value falls within a certain range.
crepancy between the model’s PDPs or ALEs and the target PDPs 6. The process is repeated until the maximum number of iterations
or ALEs and then optimizing this loss function to find the best (the default is set to 100) is satisfied or until the differences be-
parameters. tween the model’s PDPs and ALEs and the equation’s PDPs and
Thus, a new approach is proposed herein. This approach utilizes ALEs fall below a certain similarity threshold defined for each
learning by indirect feedback. Rather than adjusting the model pa- measure.
rameters directly to fit the PDPs and ALEs of the equation, the • If the model’s PDPs and ALEs are similar to the equation
training data are adjusted slightly based on the discrepancy between effect’s PDPs and ALEs (as per the predefined similarity
the ML model’s PDPs and ALEs and the equation’s PDPs and measures), the process continues to the next feature. If not,
ALEs, and then the model is retrained on these adjusted data. Simply, the targets in the training data are adjusted based on the dif-
the ML model is tuned, and one of the means of tuning the model ferences, and the model is retrained.
(among other standard methods often used in ML analyses) in- • If similarity is not achieved for all features within the maxi-
cludes a step wherein the discrepancy between the PDPs and ALEs mum number of iterations, the algorithm moves on to the
is minimized. next measure.
As is demonstrated in the case studies, this approach has the 7. After looping over all measures, a similarity check is examined
potential to guide the ML model to learn the correct behavior (as to verify if the similarity was achieved for all features for all
defined by the PDPs and ALEs of the physical equation) while still measures. If so, each model that satisfies a specific measure is
being able to fit the actual data. This approach can transfer the em- retained.
bedded empirical scientific knowledge from codal provisions into 8. Finally, an ensemble model is created using all the retained
the ML, and can be particularly useful in situations in which there is models. The PDP and ALE for the ensemble model are calcu-
good understanding of the underlying physics or other processes lated. This ensemble model represents a combination of models,
but the actual data are noisy or incomplete, so a purely data-driven each of which is trained for a different measure and adjusted for
model may not be accurate or reliable. In other words, although a each feature, based on the differences between the model’s
traditionally-trained ML model can achieve high accuracy (as is PDPs and ALEs and the equation’s PDPs and ALEs.
shown in the presented case studies), such a model generally does There are three main loops: a similarity measures loop, an itera-
not attain PDPs and ALEs that match those from codal provisions. tion loop, and a feature loop. Depending on the scenario, an engineer

© ASCE 04024039-4 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


can opt to choose if each loop needs to satisfy one or all measures, Fréchet Distance
iterations, and/or features. Similarly, the number of iterations may The Fréchet distance is another measure of similarity between
need to be adjusted for the problem or equation. After testing this two curves that takes into account the location and ordering of the
approach for a number of problems, it was found that a balance of points along the curves (Aronov et al. 2006). This measure is for-
the three loops is needed. In most cases, arriving at an ensemble mally described as the length of the shortest leash required to let a
with models satisfying two or three measures and all features pet and its owner walk along their separate paths, such that the pet
within the maximum number of iterations seems to yield balanced and the owner can vary their speed but cannot go backward (Fig. 4).
results. Additional testing and verification are likely to continue in The Fréchet distance function implemented here calculates the dis-
the future. crete Fréchet distance between two curves, represented as arrays,
and results in a single value (with the same units as the target value)
Similarity Measures representing the similarity between the two curves. This measure
assumes that the curves are continuous and connected, and that the
As mentioned previously, the relationships between features in order of points along the curve matters. Because this measure re-
Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

building codes are governed by physical equations/provisions. Thus, turns similarity in the same units as the target, a default setting was
if a ML model is developed to mimic these physical processes, the chosen arbitrarily as 10% of the observed target observation in this
PDP and ALE of the model also need to mimic the logic of the study.
physical equation. The elemental idea is that if the PDP and/or ALE
of a feature in the model follows the same trend as the physical equa- Bounding Measure
tion, this implies that the ML model has learned the underlying The custom bounding measure calculates an upper and a lower
engineering intuition and causal logic of the physical process. For bound based on a specified threshold (Fig. 5). The idea is to de-
example, if the physical equation describes a linear relationship termine whether the PDP or ALE values from the ML model fall
between a feature and its target, and the PDP or ALE of these var- within a certain range of values defined by a physical equation.
iables in the model also shows a linear relationship, this suggests The measure returns the proportion of model values that fall within
that the model has learned this relationship. the bounds. This measure assumes that the physical equation pro-
This section presents the selected similarity measures. These mea- vides a ground truth range of acceptable values. The choice of thresh-
sures were selected to be largely independent of each other to maxi- old is arbitrary (it was taken as 10% in this study), and this measure
mize their strengths and minimize their limitations. The derivations does not take into account the order of the values.
of these measures are presented in their respective references.

Normalized Euclidean Distance


The normalized Euclidean distance function calculates the Euclidean
distance between two vectors and then normalizes this distance by
the number of elements in the vectors (Fig. 3) (Bernardin et al.
2017). Normalizing the distance by the number of elements is a
way to make this measure scale independent. This can be useful
when comparing distances across data sets with different numbers
of elements. The Euclidean distance is the straight-line distance
between two points in space, and is calculated as
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
dðp; qÞ ¼ ðp1 − q1 Þ2 þ : : : þ ðpn − qn Þ2 ð4Þ

where p and q are two points in an n-dimensional space. This mea-


sure returns a zero if the shapes are identical, and unity when they
are completely dissimilar (the default setting was chosen arbitrar-
ily as 0.25 in this study).
Fig. 4. Illustration of the Fréchet distance measure (dashed lines).

Fig. 3. Illustration of the normalized Euclidean distance measure


(dashed lines). Fig. 5. Illustration of the bounding measure (dashed lines).

© ASCE 04024039-5 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


Coefficient of Determination and Correlation Coefficient Extreme Gradient Boosted Trees
In addition to the preceding four measures, two commonly used Extreme gradient boosting (XGBoost) is based on the gradient
statistical measures can be used to establish similarity between boosting framework (scikit-learn 2020a; XGBoost Python Package
PDPs or ALEs of a model and a physical equation: the coefficient 2020). Gradient boosting is a method that goes through cycles to
of determination (R2 ) and the correlation coefficient (r) also are iteratively add models into an ensemble. It begins by initializing an
used. The coefficient of determination represents the proportion of ensemble with a single model, the predictions of which can be quite
the variance for a dependent variable that is explained by a number naive. Then the algorithm begins to add new models that aim to
of features in a model. The correlation coefficient measures the correct the errors of the combined ensemble. In this way, the models
strength and direction of a linear relationship between two varia- are added sequentially until no further improvements can be made.
bles. In the context of comparing PDPs or ALEs of a model and a One of the main improvements that XGBoost brings for tree-like
physical equation, a high R2 or r approaching unity suggest a high models is a more regularized model formalization to control over-
degree of similarity. A similarity of more than 70% was deemed to fitting, which tends to be a problem with gradient boosting.
be strong, as noted by Smith (1986)
Light Gradient Boosting Machine
Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

X
n X
n The light gradient boosting machine (LGBM) is an improvement
R2 ¼ 1 − ðPi − Ai Þ2 ðAi − Amean Þ2 ð5Þ over XGboost because it combines the strengths of gradient boost-
i¼1 i¼1 ing and the LightGBM framework to achieve high performance
(Ke et al. 2017). This optimizes the objective function by iteratively
Pn fitting new trees to the negative gradient of the loss function with
i¼1 ðAi − Āi ÞðPi − P̄i Þ
R ¼ qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
Pn Pn ð6Þ respect to the current predictions. The final prediction is obtained
2 2 by summing the predictions of all the trees. The exclusive feature
i¼1 ðAi − Āi Þ i¼1 ðPi − P̄i Þ
bundling technique in LGBM groups data points in a near-lossless
way, leveraging similarities within the data. This technique improves
where A = actual measurements; P = predictions; and n = number computational efficiency by reducing the number of distinct feature
of data points. values and enabling efficient memory usage during the training
process.
Machine Learning Models Support Vector Machines
Five ML algorithms were used herein, and these were selected be- The support vector machines (SVM) algorithm aims to find the
cause they are commonly used algorithms in our domain, as noted optimal hyperplane that maximally separates the data points (scikit-
in a survey by Tapeh and Naser (2022). All ML algorithms were learn 2020b). The objective is to minimize the deviation or error
used in their default settings (including hyperparameters) for trans- between the actual data points and the hyperplane while consid-
parency and to allow repetition of the proposed approach. The fol- ering the regularization parameters. To handle nonlinearly sepa-
lowing sections briefly describe such algorithms, and their full rable data, SVM utilizes kernel functions to transform the input
description and scripts are presented in their respective references. data into a higher-dimensional feature space in which it becomes
The specific script used in creating the provided analysis can be linearly separable.
requested from the author.
Performance Metrics
Decision Trees
A decision tree (DT) is a supervised ML algorithm that uses a tree- After the ensemble was created, it was trained by randomly shuf-
like graph of decisions and their possible consequences (Naser fling and training the data into training (T), validation (V), and test-
2021b). It is a flowchart-like structure, in which each internal node ing (S) sets. The ensemble was trained and validated against the
denotes a test of a feature, each branch represents an outcome, and T and V sets and then examined on the S set following a k-folds
each leaf node holds a class label. The process of constructing a cross-validation procedure, wherein the collected data set was split
decision tree involves selecting features to test at each node in the randomly into test and training sets of k ¼ 10 groups. The ensem-
tree to minimize the total variance of the target variable among the ble was trained using nine unique sets and validated on the remain-
instances at each node. There are several strategies for making this ing set until each unique set was used as the validation set.
choice, but a common strategy is to use information gain, which Then, in lieu of the aforementioned measures selected to verify
is related to the concept of entropy (a measure of impurity) (Kent and compare the similarity between the PDPs and ALEs of the ML
1983). model and codal provision, the validity of the created ML models
also was examined using another independent performance metric,
Random Forest namely the mean absolute error (MAE), as well as R2 (Naser and
Random forest (RF) is a versatile ensemble learning algorithm that Alavi 2021; Tapeh and Naser 2022)
combines the power of decision trees and randomization techniques Pn
jP − Ai j
(Breiman 2001). The RF algorithm constructs an ensemble of de- MAE ¼ i¼1 i ð7Þ
cision trees by randomly selecting subsets of the training data and n
features. The individual trees are trained independently using the
selected subsets, and their predictions are aggregated to produce Case Studies
the final prediction. Each decision tree in the RF ensemble is built
using a random subset of the training data (known as bootstrap Seven phenomena and their respective codal provisions were
aggregating or bagging). This random sampling process helps to used herein to showcase and verify the proposed methodology.
reduce the risk of overfitting and improves the model’s generali- These are presented in the order of their complexity (i.e., the num-
zation ability. Additionally, the algorithm selects a random subset ber of involved features). Two phenomena were taken from ACI
of features at each split in the tree construction, introducing diver- 318. Two phenomena were curated from ASCE 29 (ASCE 2006);
sity among the trees and reducing the impact of individual features. one of these phenomena also appears in NDS (2018) for wood

© ASCE 04024039-6 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


construction. The fifth phenomenon was taken from the Australian pffiffiffiffiffi
f r ¼ 7.5 f c0 ð8Þ
Building Code (AS 3600) (AISC 2016) for concrete construction.
The sixth case study was taken from the American Institute of Steel
Construction (AISC) Steel Construction Manual. The final case This is a simple statistical equation, and is a suitable candidate for
study is present in multiple building codes because it deals with a first case study. Thus, the presented approach was applied to creat-
the elastic deformation of beams. ing a ML ensemble that mimics this equation. Due to the simplicity
Each phenomenon was examined using data that satisfied the of this problem, one similarity measure (normalized Euclidean dis-
practical range of the specific codal, and these data also satisfied tance) was used. Upon the successful tuning of the ensemble, the
the recommendations of the recent researchers aimed to establish R2ensemble was 1 and MAEensemble was calculated to be 0.192, and the
the health of data [van Smeden et al. (2018) with a minimum set of same metric, when calculated for the RF (without tuning), was
10 observations per feature, Riley et al. (2019) with a minimum set 0.203, implying a 5.4% improvement in accuracy. The outcome of
of 23 observations per feature, and Frank and Todeschini (1994) this approach is presented in Fig. 6. The PDP and ALE of the equiv-
with a minimum ratio of the number of observations to the number alent ensemble (made from RF) matched the codal equation well
throughout the full series. In addition, the PDPs and ALEs from
Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

of features of 3 and 5). In addition, a random ML algorithm from


those listed previously was selected for each case study to create the all tree-based algorithms also matched those of the codal equation
tuned ensemble, and the other algorithms were used without tuning well. This indicates that such models can predict this phenomenon
(i.e., without being incorporated into the proposed equivalence without actually being tuned. This working hypothesis currently is
approach) for comparison. This was done to show the proposed ap- being examined in a separate work and, at the time of the present
proach’s wide applicability and maintain the page limitation of this paper appears to be true for simple phenomena.
work, because examining each individual ML model for each case
study increased the size of this paper beyond what is commonly ASCE 29 Section 3.3 on the Design of Fire-Resistive
expected. All case studies used the default settings listed previously. Exposed Wood Members
In the PDP and ALE figures, the vertical axis represents the partial
and accumulated effects on the target, instead of the target itself. ASCE 29 is a standard for calculating the fire resistance of struc-
tural members and barrier assemblies made from structural steel,
reinforced and unreinforced concrete, timber, and masonry, and
ACI 318 Section 19.2.3.1 on Modulus of Rupture was developed as a joint effort between ASCE’s Structural Engi-
In the context of building design and construction, the tensile strength neering Institute and the Society of Fire Protection Engineers
of concrete is particularly important because it helps to determine the (SFPE). This standard contains a number of empirical codal pro-
concrete’s ability to resist tensile and torsional forces, as well as visions that can be used to evaluate and predict the fire resistance of
cracking (for example, due to environmental factors such as temper- structural members.
ature changes and freeze–thaw cycles, and so forth). The tensile Section 3.3 in this standard provides guidance on calculating the
strength of concrete generally is only about 10% of its compressive fire resistance rating (or time to failure, t), measured in minutes, of
strength, and measures the force required to break or crack concrete. timber beams with a least nominal dimension of 6 in. (140 mm).
Although the tensile strength of concrete increases with an For the severe case of beams exposed on four sides, t is calcu-
increase in the compressive strength, the ratio of the tensile to lated as
the compressive strength decreases as the compressive strength in-
creases. The tensile strength often is approximated proportionally to t ¼ γzb½4 − ðb=dÞ ð9Þ
the square root of the compressive strength of concrete. ACI 318
Section 19.2.3.1 presents the modulus of rupture for normal strength where γ is a constant [2.54 min=in. (0.1 min=mm)], z is a load
concrete for use in calculating deflections as obtained from statis- factor that equals 1.3 for lightly loaded beams with a loading ratio
tical analysis (McNeely and Stanley 1963; Raphael 1984) <50%; and b and d = actual breadth and depth of a beam. The same

Fig. 6. Comparison for Case study 1 (unit: psi). (f = compressive strength of concrete. Base ML mode: RF.)

© ASCE 04024039-7 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


empirical equation also appears in NDS (2018) for wood construc- ACI 318 Table 22.5.5.1 on Shear Strength for Beams
tion. This empirical expression contains four parameters, two of without Minimum Shear Reinforcement
which are considered constants and hence are treated as such, ACI 318 Table 22.5.5.1 lists the following equation for calculating
and therefore only b and d are considered to be features. This case the shear strength for beams without minimum shear reinforcement,
study applied the proposed methodology using LGBM as the base which resulted from an iterative process led by the Joint ACI-ASCE
ML model. Committee 445 and Subcommittee E of ACI Committee 318:
In this case study, only three similarity measures converged
pffiffiffiffiffi
with default settings (i.e., Fréchet distance, the bounding measure, V c ¼ ½8λs λρ1=3 fc þ N u =6Ag bd ð10Þ
and the normalized Euclidean distance). Special tuning was not ap-
plied to converge the other measure as the realized results were where ρw = flexural reinforcement ratio for tension reinforcement
deemed acceptable (Fig. 7). The MAEensemble was calculated to in As =bw d; N u = member-factored axial load; and Ag = gross
be 0.859 min, and the same metric calculated for the same algorithm member area. The lightweight-aggregate concrete factor, λ, was de-
without tuning was 1.111 min. The value of R2 between the fined previously. The value of V c should not be taken as less than
Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

ensemble and the codal equation was 1. It is clear that the tuned zero, nor as greater than λ5ðfc0 Þ0.5 bw d, and N u =6Ag should not be
ensemble outperformed all individual models in matching the codal taken as greater than 0.05fc0 . In this case study, beams with applied
equation’s PDPs and ALEs. In this case study, and in the next case axial forces were not considered. In addition, λs (the size effect
study, matching PDPs often is found to be faster (and more afford- term) and λ were taken as unity for simplicity.
able) than matching ALEs. This stems from the more complex The LGBM was used in this case study as the base model. Fig. 8
derivation nature of ALEs than of PDPs. Given this observation, shows the behavior of the tuned ensemble in terms of PDPs and
the proposed methodology was appended to allow the user or en- ALEs after satisfying two similarity measures, namely the Fréchet
gineer to select whether they seek to match PDP, ALE, or both. Such distance and the bounding measure. The MAEensemble was calculated
an amendment did not alter the analysis of the first case study. to be 3,444.348 lbs (1 lbs = 4.45 N), and that for the untuned model

Fig. 7. Comparison for Case study 2 (unit: min). (b and d = breadth and depth of timber beams, respectively. Base ML mode: LGBM.)

© ASCE 04024039-8 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

Fig. 8. Comparison for Case study 3 (unit: lbs). (Base ML mode: LGBM.)

was 3,620.0 lbs. The R2 for the ensemble was 99%. Clear matching rating of hollow steel columns filled with unreinforced normal-
between the ensemble and codal equation is evident. weight concrete. The equation was established empirically through
To examine further the performance of the tuned ensemble (as a number of publications and has been verified by computer sim-
well as the proposed approach), an additional layer of validation ulations at the National Research Council of Canada (NRC) (Kodur
was performed. In this process, 100 random beams were selected 1999; Kodur and MacKinnon 2000). This expression is also listed
from ACI’s shear database (Reineck et al. 2003) for RC members under Section D-2.6.6. in the National Building Code of Canada
without shear reinforcement used in deriving the codal expression and also is applicable to tubular shapes; however, only the case of
and examined against the codal expression and tuned ensemble. circular tubes is presented herein
Fig. 9 shows the results of this comparison; the ensemble achieved pffiffiffiffiffiffiffiffiffiffi
ðf þ 20Þ
better predictive capability than the codal expression in pre- R¼a D2 D=C ð11Þ
dicting the shear strength of the selected beams (MAEensemble ¼ 60ðLe − 1000Þ
4,930.51 lbs, MAEcodalexpression ¼ 5,219.73 lbs, R2ensemble ¼ 84.1%, where R = fire resistance rating (h); a ¼ 0.07 for circular columns
and R2codalexpression ¼ 80.83%). This additional external verification filled with siliceous aggregate concrete, and 0.08 for circular
could not be conducted in the other case studies because the data columns filled with carbonate aggregate concrete; f = specified
used to create the codal provisions of interest were not available. 28-day compressive strength of concrete (MPa); Le = column ef-
fective length (mm); D = outside diameter for circular columns
(mm); and C = compressive force due to unfactored dead load
ASCE 29 Section 5.2.3 on Concrete-Filled Hollow and live load (kN).
Steel Columns This codal provision is accompanied by additional stipulations
Another complex codal provision is addressed in this case study. that were enforced in this analysis, including that the required R
ASCE 29 also presents guidance for evaluating the fire resistance rating should be less than or equal to 2 h, the specified f should
be equal to or greater than 20 MPa and should not exceed 40 MPa,
the column’s effective length should be at least 2,000 mm and
should not exceed 4,000 mm, the outside diameter should be at
least 140 mm and should not exceed 410 mm, and the compressive
force should not exceed the design strength of the concrete core
determined in accordance with AISC LRFD-94 (AISC 1994).
The RF algorithm was used as the base model. The convergence
of the ensemble on the PDPs and ALEs was difficult to obtain,
and hence the results in Fig. 10 are shown for a converged run
wherein only the PDPs were matched using all aforementioned
similarity measures. Despite matching only the PDPs, the ALEs
also seem to have attained some degree of matching to that from
the equation. The MAEensemble of the tuned model was calculated to
be 0.281 h, versus 0.471 h for the untuned model—an improvement
of about 40%.

AS 3600 Fire Resistance of Concrete Columns


This case study addresses a more complex codal provision with
Fig. 9. Comparison against 100 randomly selected beams from ACI
six variables. A former edition of the Australian concrete code
shear database (unit: lbs). (Data from Reineck et al. 2003.)
(AS 3600) used the following statistical-based codal equation to

© ASCE 04024039-9 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

Fig. 10. Comparison for Case study 4 (unit: h). (Base ML mode: LGBM.)

evaluate the fire resistance of concrete columns (Wade et al. 1997); AISC Manual Section F6.1 for I-Shaped Members and
this equation was fitted using the results of standard fire tests and Channels Bent about Their Minor Axis
numerical simulations: Section F6.1 in the AISC manual (AISC 2016) presents the follow-
ing simple equation to evaluate the nominal flexural strength, Mn ,
k × f1.3 × B3.3 D1.8 in yielding:
R¼ ð12Þ
105 × N 1.5 × L0.9
M ¼ fy × Z ≤ 1.6fy Sy ð13Þ
where R = fire resistance (min); k is a constant dependent on the
cover and steel reinforcement ratio, and equals 1.47 and 1.48 for where Sy = elastic section modulus about the y-axis.
cover less than 35 mm and greater than or equal to 35 mm, respec- This simple codal provision has two main features, and the data
tively; f c ¼ 28-day compressive strength of concrete (MPa); B = used in this case study were selected to satisfy those conditions.
least dimension of column (mm); D = greatest dimension of column XGBoost was selected to be the base ML model with two simi-
(mm); N = axial load during fire (kN); and L = effective length (mm). larity measures (i.e., Fréchet distance and the Bounding measure).
This case study utilized the RF algorithm as a base model. In The outcome of this analysis is shown in Fig. 12. Both similarity
addition, R2 and the bounding measure were satisfied in this case metrics were satisfied, and MAEensemble was calculated to be
study. MAEensemble (with default settings) was calculated to be 1,637.95 kip · in (1 kip.in = 1,355.8 N.m), and that for the same
56.957 min, and that for the same algorithm without tuning was algorithm without tuning was 1,659.74 kip.in. The tuned ensemble
68.03 min (the R2 between the ensemble and the codal equation matched the PDP and ALE of the codal equation and outperformed
was 89%, and it is quite possible to improve the performance of the those obtained from the other models (Fig. 12). Although this was a
tuned ensemble by tuning the default settings). The tuned ensemble simple case study, and there is a need for additional tests in future
matched a good degree of similarity in terms of trends between work, this finding seems to indicate the suitability and applicability
PDPs and ALEs with the former codal provision (Fig. 11). In some of the proposed approach to the ability of ML to mimic principle-
features (namely k, f, and L), there was a gap between the PDPs based codal provisions.
and ALEs obtained from all ML models and the equation. This gap
seems to have a shifting effect, wherein the models capture the Various Building Codes: On the Elastic Deformation
trends of the PDPs and ALEs, but not the values in some of the of Beams
involved features. Another theoretically derived expression was used as a case study.
The equation for elastic deformation in simply supported beams
loaded with a point load placed at the midspan, which is used in
Additional Case Studies Based on Provisions
a variety of building codes and standards [such as Table 3-23 in the
Stemming from a Theoretical Derivation
AISC manual (AISC 2016)], is
These two case studies explored the applicability of the proposed
approach to codal provisions that are not empirical in nature but PL
Elastic deformation ¼ ð14Þ
rather are derived from first principles. 48EI

© ASCE 04024039-10 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

Fig. 11. Comparison for case study no. 7 (in mm) (k = constant for cover and steel reinforcement ratio; fc = 28-day compressive strength of concrete;
B = least dimension of column; D = greatest dimension of column; N = axial load during fire; and L = effective length [Note: The base ML mode:
XGBoost].

where P = applied load (N); L = span of the beam (mm); E = modu- importance measures the total effect of shuffling a feature on pre-
lus of elasticity (Pa); and I = moment of inertia (mm4 ). diction accuracy. Although this is a common approach in ML, the
For this case study, the XGBoost model was selected as the base perception of feature importance in an equation is somewhat of a
model. All five similarity measures were used, but only one mea- nuanced concept simply because all features in a given equation
sure (R2 ) was satisfied (given the default settings). Fig. 13 shows must hold true (Naser 2023b). Therefore, one could argue that
the PDPs and ALEs generated from the model and Eq. (14). Despite all features in an equation are equally important.
satisfying one similarity measure, the achieved performance can be Other explainability measures [i.e., SHapley Additive exPlana-
deemed acceptable. MAEensemble was calculated to be 2.461 mm, tions (SHAP) (Lundberg and Lee 2017), local interpretable model-
and that for the same algorithm without tuning was 2.391 mm. The agnostic explanations (LIME) (Ribeiro et al. 2016)), prototypes,
R2 value between the ensemble and the equation was 96%. and so forth] could be explored to build upon the current analysis.
The author believes that, despite their computational complexity,
such measures could be proven successful, and hence can be stream-
Additional Thoughts, Limitations, and lined and added to the proposed methodology. There is room to in-
Future Directions vestigate the influence of model type (i.e., tree-based, ANN, and so
forth) versus the degree of explainability. It is possible that some
This section documents some of the main limitations of the proposed models might fit a phenomenon more easily, more accurately, or
methodology and shares insights into future research directions. with higher explainability than other models. Therefore, there is a
possibility of a trade-off between accuracy and explainability.
A Note on Other Matching Methods
Limitations of the Proposed Methodology and
An engineer might entertain the idea of achieving equivalence
Future Directions
through matching feature importance (e.g., a measure of how valu-
able a feature is in the construction of the model) between the codal The presented approach can be viewed as complex and somewhat
provision and the ML model. For example, if a given model’s error lengthy. As such, this approach may struggle to guarantee the con-
increases after systematically permuting a feature, the so-called fea- vergence of a ML model that both fits the data well and matches the
ture is important, because permuting adversely affected the model’s physical equation’s behavior. In reality, the proposed approach con-
predictions (Altmann et al. 2010). verged within 2 min of analysis for all the case studies mentioned
Although this concept could possibly be integrated into the in this work using a typical desktop that can be found in a com-
proposed methodology, and was successful in the first two studies mon engineering design office. However, in lengthy simulations or
(these results were not shown in the present work), feature multiphased codal provisions, it is possible that the proposed

© ASCE 04024039-11 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

Fig. 12. Comparison for Case study 6 (unit: kip·in.). (Base ML mode: XGBoost.)

Fig. 13. Comparison for Case study 7 (unit: mm). (Base ML mode: XGBoost.)

© ASCE 04024039-12 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


approach might suffer from instability due to chasing a target that ASCE. 2016. Minimum design loads for buildings and other structures.
continues to change slightly but constantly based on the model’s ASCE 7. Reston, VA: ASCE.
own performance. Benjamin, J. R., and C. A. Cornell. 2014. Probability, statistics, and de-
Fortunately, there are potential solutions to these problems. One cision for civil engineers. Chelmsford, MA: Courier Corporation.
potential approach is to attempt to optimize model parameters via Bernardin, A., L. Hoyet, A. Mucherino, D. Gonçalves, and F. Multon.
2017. “Normalized Euclidean distance matrices for human motion re-
gradient-based methods by computing the gradients of the PDPs
targeting.” In Proc., 10th Int. Conf. on Motion in Games. New York:
and ALEs of the model parameters. Alternatively, an engineer might Association for Computing Machinery.
consider employing evolutionary or population-based optimization Breiman, L. 2001. “Random forests.” Mach. Learn. 45 (10): 5–32. https://doi
algorithms to search the parameter space for a set that minimizes the .org/10.1023/A:1010933404324.
discrepancy between the PDPs and ALEs. If neural networks are Cakiroglu, C., K. Islam, G. Bekdaş, U. Isikdag, and S. Mangalathu. 2022.
used, prespecified layers can be trained to satisfy similarity metrics “Explainable machine learning models for predicting the axial compres-
separately. When a layer passes the similarity check, it can be frozen sion capacity of concrete filled steel tubular columns.” Constr. Build.
to maintain it before moving into another measure. This approach Mater. 356 (Nov): 129227. https://doi.org/10.1016/j.conbuildmat.2022
Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

could be an alternative to the ensemble approach used herein. It .129227.


would be interesting to see what other possible solutions (and lim- Degtyarev, V. V., and K. D. Tsavdaridis. 2022. “Buckling and ultimate load
itations) this research area might realize in the future. prediction models for perforated steel beams using machine learning
algorithms.” J. Build. Eng. 51 (Jul): 104316. https://doi.org/10.1016/j
.jobe.2022.104316.
Ellingwood, B. R. 1994. “Probability-based codified design: Past accom-
Conclusions
plishments and future challenges.” Struct. Saf. 13 (3): 159–176. https://
doi.org/10.1016/0167-4730(94)90024-8.
Building codes provide definitive standards to guide civil engineer-
Emmert-Streib, F., O. Yli-Harja, and M. Dehmer. 2020. “Explainable ar-
ing practices and ensure safety. Although ML has shown the poten-
tificial intelligence and machine learning: A reality rooted perspective.”
tial to enhance predictive abilities, its integration into building codes Wiley Interdiscip. Rev.: Data Min. Knowl. Discovery 10 (6): e1368.
is complicated due to transparency and regulatory approval issues. https://doi.org/10.1002/widm.1368.
As civil engineering moves toward a more datacentric future, blend- Feldman, L. R., and F. M. Bartlett. 2005. “Bond strength variability in pull-
ing traditional engineering wisdom with innovative ML approaches out specimens with plain reinforcement.” ACI Struct. J. 102 (6): 860.
could unlock new frontiers in designing and constructing safer, more- https://doi.org/10.14359/14794.
efficient buildings. This work proposes an approach to modernizing Frank, I., and R. Todeschini. 1994. The data analysis handbook. Amsterdam,
building codes via ML. In this approach, empirical and regression- Netherlands: Elsevier.
based codal equations can be used to create equivalent ML models Franssen, J.-M., and J.-C. Dotreppe. 2003. “Fire tests and calculation meth-
that retain the logic and properties of such equations. This approach ods for circular concrete columns.” Fire Technol. 39 (1): 89–97. https://
has been shown to be successful in structural and fire engineering doi.org/10.1023/A:1021783311892.
Friedman, J. H. 2001. “Greedy function approximation: A gradient boosting
phenomena adopted in notable codal provisions.
machine.” In Annals of statistics, 1189–1232. Waite Hill, OH: Institute of
Mathematical Statistics.
IBC (International Building Code). 2018. International building code.
Data Availability Statement Washington, DC: International Building Code.
Ke, G., Q. Meng, T. Finley, T. Wang, W. Chen, W. Ma, Q. Ye, and T. Y. Liu.
Some or all data, models, or code that support the findings of this 2017. “LightGBM: A highly efficient gradient boosting decision tree.”
study are available from the corresponding author upon reasonable Adv. Neural Inf. Process. Syst. 30: 30–39.
request. Kent, J. T. 1983. “Information gain and a general measure of correlation.”
Biometrika 70 (1): 163–173. https://doi.org/10.1093/biomet/70.1.163.
Kodur, V. K. R. 1999. “Performance-based fire resistance design of concrete-
Acknowledgments filled steel columns.” J. Constr. Steel Res. 51 (1): 21–36. https://doi.org
/10.1016/S0143-974X(99)00003-6.
The author thanks the kind reviewers and editors for their help and Kodur, V. K. R., and D. H. MacKinnon. 2000. “Design of concrete-filled
support in improving the state of this work. hollow structural steel columns for fire endurance.” Eng. J. Am. Inst.
Steel Constr. 37 (1): 13–24.
Krueger, D., E. Caballero, J.-H. Jacobsen, A. Zhang, J. Binas, D. Zhang,
References R. Priol Le, and A. Courville. 2021. “Out-of-distribution generalization
via risk extrapolation (REx).” In Proc., 38th Int. Conf. on Machine
ACI (American Concrete Institute). 2019. Building code requirements for Learning, 5815–5826.
structural concrete. ACT 318-19. Farmington Hills, MI: ACI. Kuchma, D. A., S. Wei, D. H. Sanders, A. Belarbi, and L. C. Novak. 2019.
AISC. 1994. Specification for structural steel buildings. Chicago: AISC. “Development of the one-way shear design provisions of ACI 318-19
AISC. 2016. Specification for structural steel buildings. ANSI/AISC 360- for reinforced concrete.” ACI Struct. J. 116 (4): 285–296. https://doi.org
16. Chicago: AISC. /10.14359/51716739.
Aladsani, M. A., H. Burton, S. A. Abdullah, and J. W. Wallace. 2022. Law, T. 2021. “Beyond minimum: Proposition for building surveyors to
“Explainable machine learning model for predicting drift capacity of exceed the minimum standards of the construction code.” J. Leg. Aff.
reinforced concrete walls.” ACI Struct. J. 119 (3): 191–204. https://doi Dispute Resolut. Eng. Constr. 13 (2): 03721001. https://doi.org/10.1061
.org/10.14359/51734484. /(ASCE)LA.1943-4170.0000463.
Altmann, A., L. Toloşi, O. Sander, and T. Lengauer. 2010. “Permutation Lie, T. T. 1992. Structural fire protection. Edited by T. T. Lie. Reston, VA:
importance: A corrected feature importance measure.” Bioinformatics ASCE.
26 (10): 1340–1347. https://doi.org/10.1093/bioinformatics/btq134. Lundberg, S. M., and S. I. Lee. 2017. “A unified approach to interpreting
Aronov, B., S. Har-Peled, C. Knauer, Y. Wang, and C. Wenk. 2006. model predictions.” In Vol. 30 of Proc., 31st Conf. on Neural Informa-
“Fréchet distance for curves, revisited.” In Proc., Algorithms–ESA 2006: tion Processing Systems (NIPS 2017). Red Hook, NY: Curran
14th Annual European Symp. Berlin: Springer. Associates.
ASCE. 2006. Standard calculation methods for structural fire protection. Marasco, S., A. Fiore, R. Greco, G. P. Cimellaro, and G. C. Marano. 2021.
ASCE29-05. Reston, VA: ASCE. “Evolutionary polynomial regression algorithm enhanced with a robust

© ASCE 04024039-13 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039


formulation: Application to shear strength prediction of RC beams with- Int. Conf. on Knowledge Discovery and Data Mining, 1135–1144. New
out stirrups.” J. Comput. Civ. Eng. 35 (6): 04021017. https://doi.org/10 York: Association for Computing Machinery. https://doi.org/10.1145
.1061/(ASCE)CP.1943-5487.0000985. /2939672.2939778.
McNeely, D. J., and D. L. Stanley. 1963. “Tensile strength of concrete.” Riley, R. D., K. I. E. Snell, J. Ensor, D. L. Burke, F. E. Harrell Jr, K. G. M.
J. Am. Concr. Inst. 60 (6): 751–762. https://doi.org/10.14359/7876. Moons, and G. S. Collins. 2019. “Minimum sample size for developing
Molnar, C. 2019. Interpretable machine learning. A guide for making black a multivariable prediction model: Part II-binary and time-to-event out-
box models explainable. British Columbia, Canada: Leanpub. comes.” Stat. Med. 38 (7): 1276–1296. https://doi.org/10.1002/sim.7992.
Naser, M. Z. 2021a. “Demystifying ten big ideas and rules every fire scikit-learn. 2020a. “sklearn.ensemble.GradientBoostingRegressor.” Ac-
scientist & engineer should know about blackbox, whitebox & causal cessed February 9, 2021. https://scikit-learn.org/stable/modules/generated
artificial intelligence.” Fire Technol. 58 (3): 1075–1085. https://doi.org /sklearn.ensemble.GradientBoostingRegressor.html.
/10.1007/s10694-021-01210-1. scikit-learn. 2020b. “1.4. Support Vector Machines.” Accessed April 5,
Naser, M. Z. 2021b. “Observational analysis of fire-induced spalling of 2021. https://scikit-learn.org/stable/modules/svm.html.
concrete through ensemble machine learning and surrogate modeling.” Smith, G. 1986. Probability and statistics in civil engineering. London:
J. Mater. Civ. Eng. 33 (1): 04020428. https://doi.org/10.1061/(ASCE) Collins.
MT.1943-5533.0003525.
Downloaded from ascelibrary.org by Universidad de la Sabana on 03/12/24. Copyright ASCE. For personal use only; all rights reserved.

Standards Australia. 1994. Concrete structures. AS 3600. Sydney,


Naser, M. Z. 2023a. Leveraging artificial intelligence in engineering, man- Australia: Standards Australia.
agement, and safety of infrastructure. Edited by M. Naser. London: Tapeh, A., and M. Z. Naser. 2022. “Artificial intelligence, machine learning,
Taylor & Francis. and deep learning in structural engineering: A scientometrics review of
Naser, M. Z. 2023b. Machine learning for civil and environmental engi-
trends and best practices.” Arch. Comput. Methods Eng. 30 (1): 115–159.
neers: A practical approach to data-driven analysis, explainability, and
https://doi.org/10.1007/s11831-022-09793-w.
causality. New York: Wiley.
Tarawneh, A., Y. Momani, and R. Alawadi. 2021. “Leveraging artificial
Naser, M. Z., and H. Alavi. 2021. “Error metrics and performance fitness
intelligence for more accurate and reliable predictions of anchors shear
indicators for artificial intelligence and machine learning in engineering
breakout capacity in thin concrete members.” Structures 32 (Aug):
and sciences.” In Architecture, Structures and construction 2021, 1–19.
1005–1014. https://doi.org/10.1016/j.istruc.2021.03.074.
New York: Springer.
van Smeden, M., K. G. Moons, J. A. H. de Groot, G. S. Collins, D. G.
Naser, M. Z., and B. Ross. 2022. “An opinion piece on the dos and don’ts of
artificial intelligence in civil engineering and charting a path from data- Altman, M. J. C. Eijkemans, and J. B. Reitsma. 2018. “Sample size for
driven analysis to causal knowledge discovery.” Civ. Eng. Environ. Syst. binary logistic prediction models: Beyond events per variable criteria.”
39 (1): 1–11. https://doi.org/10.1080/10286608.2022.2049257. Stat. Methods Med. Res. 28 (8): 2455–2474. https://doi.org/10.1177
NDS (National Design Specification). 2018. Calculating the fire resistance /0962280218784726.
of exposed wood members. Technical Rep. No. 10. Leesburg, VA: Wade, C. A., G. S. Cowles, R. J. Potter, and P. Sanders. 1997. “Concrete
American Wood Council. blade columns in fire wade.” In Proc., Concrete 97 Conf. Adelaide,
Ollgaard, J. G., R. G. Slutter, and J. W. Fisher. 1971. “Shear strength of stud Australia: Steel Reinforcement Institute of Australia.
connectors in lightweight and normal-weight concrete.” AISC Eng. J. Windapo, A., and K. Cattell. 2010. “A study of building contractors’ com-
9 (2): 86. pliance with national building regulations in Cape Town.” In Proc.,
Paleyes, A., R. G. Urma, and N. D. Lawrence. 2022. “Challenges in COBRA 2010-Construction, Building and Real Estate Research Conf.
deploying machine learning: A survey of case studies.” ACM Comput. of the Royal Institution of Chartered Surveyors. London: Royal
Surv. 55 (6): 1–29. https://doi.org/10.1145/3533378. Institution of Chartered Surveyors.
Raphael, J. M. 1984. “Tensile strength of concrete.” J. Am. Concr. Inst. XGBoost Python Package. 2020. “Python package introduction.” Accessed
81 (Jun): 158–165. https://doi.org/10.14359/10653. February 10, 2021. https://xgboost.readthedocs.io/en/latest/python/python
Reineck, K. H., D. A. Kuchma, K. S. Kim, and S. Marx. 2003. “Shear _intro.htm.
database for reinforced concrete members without shear reinforcement.” Yahiaoui, A., S. Dorbani, and L. Yahiaoui. 2023. “Machine learning tech-
ACI Struct. J. 100 (2): 240–249. https://doi.org/10.14359/12488. niques to predict the fundamental period of infilled reinforced concrete
Ribarič, M., L. Šušteršič, and J. Stefan. 2008. “Empirical relationship as a frame buildings.” Structures 54 (Jun): 918–927. https://doi.org/10.1016
stepping-stone to theory.” Preprint, submitted May 12, 2021, https://doi /j.istruc.2023.05.052.
.org/10.48550/arXiv.0810.0905. Yakubu, S. 2019. “Determinants of non-compliance with structural build-
Ribeiro, M. T., S. Singh, and C. Guestrin. 2016. “‘Why should I trust you?’ ing code standards in Nigeria.” Proc. Inst. Civ. Eng. Manage. Procure.
Explaining the predictions of any classifier.” In Proc., ACM SIGKDD Law 172 (2): 47–59. https://doi.org/10.1680/jmapl.18.00033.

© ASCE 04024039-14 J. Struct. Eng.

J. Struct. Eng., 2024, 150(5): 04024039

You might also like