<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="style/jpub3-html-trans.xsl"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xml:lang="en" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMES</journal-id>
<journal-id journal-id-type="nlm-ta">CMES</journal-id>
<journal-id journal-id-type="publisher-id">CMES</journal-id>
<journal-title-group>
<journal-title>Computer Modeling in Engineering &#x0026; Sciences</journal-title>
</journal-title-group>
<issn pub-type="epub">1526-1506</issn>
<issn pub-type="ppub">1526-1492</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">72200</article-id>
<article-id pub-id-type="doi">10.32604/cmes.2025.072200</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Predicting Concrete Strength Using Data Augmentation Coupled with Multiple Optimizers in Feedforward Neural Networks</article-title>
<alt-title alt-title-type="left-running-head">Predicting Concrete Strength Using Data Augmentation Coupled with Multiple Optimizers in Feedforward Neural Networks</alt-title>
<alt-title alt-title-type="right-running-head">Predicting Concrete Strength Using Data Augmentation Coupled with Multiple Optimizers in Feedforward Neural Networks</alt-title>
</title-group>
<contrib-group>
<contrib id="author-1" contrib-type="author">
<name name-style="western"><surname>Choudhary</surname><given-names>Sandeerah</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western"><surname>Abbas</surname><given-names>Qaisar</given-names></name><xref ref-type="aff" rid="aff-2">2</xref></contrib>
<contrib id="author-3" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Akram</surname><given-names>Tallha</given-names></name><xref ref-type="aff" rid="aff-3">3</xref><xref rid="cor1" ref-type="corresp">&#x002A;</xref><email>t.akram@psau.edu.sa</email></contrib>
<contrib id="author-4" contrib-type="author">
<name name-style="western"><surname>Qureshi</surname><given-names>Irshad</given-names></name><xref ref-type="aff" rid="aff-4">4</xref></contrib>
<contrib id="author-5" contrib-type="author">
<name name-style="western"><surname>Aldajani</surname><given-names>Mutlaq B.</given-names></name><xref ref-type="aff" rid="aff-2">2</xref></contrib>
<contrib id="author-6" contrib-type="author">
<name name-style="western"><surname>Salahuddin</surname><given-names>Hammad</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<aff id="aff-1"><label>1</label><institution>Department of Civil Engineering, COMSATS University Islamabad, Wah Campus</institution>, <addr-line>Wah, 47040</addr-line>, <country>Pakistan</country></aff>
<aff id="aff-2"><label>2</label><institution>College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU)</institution>, <addr-line>Riyadh, 11432</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-3"><label>3</label><institution>Department of Information Systems, College of Computer Engineering and Sciences, Prince Sattam bin Abdulaziz University</institution>, <addr-line>Al-Kharj, 11942</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-4"><label>4</label><institution>Department of Civil Engineering, University of Engineering &#x0026; Technology</institution>, <addr-line>Taxila, 47040</addr-line>, <country>Pakistan</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Tallha Akram. Email: <email>t.akram@psau.edu.sa</email></corresp>
</author-notes>
<pub-date date-type="collection" publication-format="electronic">
<year>2025</year>
</pub-date>
<pub-date date-type="pub" publication-format="electronic">
<day>26</day><month>11</month><year>2025</year>
</pub-date>
<volume>145</volume>
<issue>2</issue>
<fpage>1755</fpage>
<lpage>1787</lpage>
<history>
<date date-type="received">
<day>21</day>
<month>08</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>16</day>
<month>10</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2025 The Authors.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Published by Tech Science Press.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMES_72200.pdf"></self-uri>
<abstract>
<p>The increasing demand for sustainable construction practices has led to growing interest in recycled aggregate concrete (RAC) as an eco-friendly alternative to conventional concrete. However, predicting its compressive strength remains a challenge due to the variability in recycled materials and mix design parameters. This study presents a robust machine learning framework for predicting the compressive strength of recycled aggregate concrete using feedforward neural networks (FFNN), Random Forest (RF), and XGBoost. A literature-derived dataset of 502 samples was enriched via interpolation-based data augmentation and modeled using five distinct optimization techniques within MATLAB&#x2019;s Neural Net Fitting module: Bayesian Regularization, Levenberg-Marquardt, and three conjugate gradient variants&#x2014;Powell/Beale Restarts, Fletcher-Powell, and Polak-Ribiere. Hyperparameter tuning, dropout regularization, and early stopping were employed to enhance generalization. Comparative analysis revealed that FFNN outperformed RF and XGBoost, achieving an R<sup>2</sup> of 0.9669. To ensure interpretability, accumulated local effects (ALE) along with partial dependence plots (PDP) were utilized. This revealed trends consistent with the pre-existent domain knowledge. This allows estimation of strength from the properties of the mix without extensive lab testing, permitting designers to track the performance and sustainability trends in concrete mix designs while promoting responsible construction and demolition waste utilization.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Feedforward neural networks</kwd>
<kwd>recycled aggregates</kwd>
<kwd>compressive strength prediction</kwd>
<kwd>optimization techniques</kwd>
<kwd>data augmentation</kwd>
<kwd>grid search</kwd>
</kwd-group>
<funding-group>
<award-group id="awg1">
<funding-source>Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University</funding-source>
<award-id>IMSIU-DDRSP2503</award-id>
</award-group>
</funding-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>Concrete is the most widely utilized construction material, crucially essential for infrastructure development and building construction. The components of concrete are cement, aggregates (sand and gravel), and water, which make it strong, durable, and versatile. The aggregates are joined together by the cement paste, resulting in a solid, long-lasting material with improved qualities. Still, the manufacturing and use of concrete have substantial environmental, financial, and resource concerns [<xref ref-type="bibr" rid="ref-1">1</xref>,<xref ref-type="bibr" rid="ref-2">2</xref>]. The annual consumption of natural aggregates (NAs) for concrete production is estimated at 8&#x2013;12 billion tons globally [<xref ref-type="bibr" rid="ref-3">3</xref>]. NAs in concrete manufacturing, either as gravel or sand, have certain environmental drawbacks like habitat destruction, water pollution, resource depletion, <inline-formula id="ieqn-2"><mml:math id="mml-ieqn-2"><mml:mi>C</mml:mi><mml:msub><mml:mi>O</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula> emissions, and land use. The construction industry in the United Kingdom supplies about 44% of landfill waste, while in Australia the industry&#x2019;s land waste is approximately 44%, 29% in the United States, and 35% across the globe [<xref ref-type="bibr" rid="ref-4">4</xref>]. China, on average, generates 8 billion tons of construction and demolition waste (CDW) every year [<xref ref-type="bibr" rid="ref-5">5</xref>]. Nonetheless, it is undeniable that construction and demolition debris is notorious for its detrimental impact on the environment [<xref ref-type="bibr" rid="ref-6">6</xref>]. CDW remains are generally classified into five categories, i.e., unsorted mixed fractions, wood, metal, concrete, and minerals [<xref ref-type="bibr" rid="ref-7">7</xref>]. Concrete waste constitutes the largest proportion of construction waste, comprising approximately 35% of the grand [<xref ref-type="bibr" rid="ref-8">8</xref>]. One of the biggest polluters, the construction sector, is beginning to show concern for the problem [<xref ref-type="bibr" rid="ref-9">9</xref>]. Also, in recent years, there has been an increasing interest in discovering sustainable alternatives to construction materials [<xref ref-type="bibr" rid="ref-10">10</xref>]. CDW can be used as aggregates in recycled aggregate (RA) form as a substitute for natural resources in concrete mixes. It could be one way to lessen the environmental impact. Also, this enhances performance and conform to sustainable construction principles, rendering them essential for future advancements in the sector [<xref ref-type="bibr" rid="ref-11">11</xref>]. Numerous academics have investigated methods for producing concrete with recycled aggregates (RAs), which recommends that more than 75% of construction debris might be recycled to produce concrete, mainly to lower <inline-formula id="ieqn-3"><mml:math id="mml-ieqn-3"><mml:mi>C</mml:mi><mml:msub><mml:mi>O</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula> emissions [<xref ref-type="bibr" rid="ref-5">5</xref>].</p>
<p>Many researchers are focusing on predicting various properties of concrete with different components. Numerous concrete attributes, covering mechanical, thermal, and durability aspects, have been predicted in-depth in the literature, and the effectiveness of several prediction techniques has been studied and documented [<xref ref-type="bibr" rid="ref-12">12</xref>,<xref ref-type="bibr" rid="ref-13">13</xref>]. The compressive strength (CS) of concrete, being the most crucial property, has been the foremost priority of researchers to predict, as concrete made with recycled aggregates (RAs) exhibits complex relationships between design mix components and CS, and traditional methods of concrete mix design are material, cost, labor, and time intensive [<xref ref-type="bibr" rid="ref-10">10</xref>]. The manual testing procedure is susceptible to human error as well, and even minor mistakes may significantly extend the typical wait time [<xref ref-type="bibr" rid="ref-14">14</xref>]. Since RAC is mixed with different types of recycled materials, it is hard to accurately forecast its performance using conventional regression strategies [<xref ref-type="bibr" rid="ref-15">15</xref>]. The poor qualities of RAs, such as higher affinity for water, higher brittleness of the old mortar, cracks created during the crushing process, and weak bonding between old and new mortar, are the main reasons for the unpredictable behavior of concrete having recycled aggregates [<xref ref-type="bibr" rid="ref-16">16</xref>]. Concrete manufactured using RAs has a 30% to 40% lower CS than natural aggregate concrete (NAC) due to the inferior qualities of recycled particles [<xref ref-type="bibr" rid="ref-17">17</xref>,<xref ref-type="bibr" rid="ref-18">18</xref>]. In the civil engineering discipline, the development and application of machine learning (ML) techniques have garnered significant interest recently [<xref ref-type="bibr" rid="ref-19">19</xref>]. These techniques have primarily been used for optimization and prediction [<xref ref-type="bibr" rid="ref-20">20</xref>]. Recent ML model-based studies on experimental data can anticipate the influence of CDW aggregate in the concrete behavior with an acceptable error range, improving one-factor-at-a-time experimental method studies by lowering the amount of materials used, testing efforts, time consumed, and cost incurred [<xref ref-type="bibr" rid="ref-21">21</xref>]. Furthermore, a key benefit of using ML approaches is their ability to consider a multitude of input factors [<xref ref-type="bibr" rid="ref-22">22</xref>], and ML encompasses a wide variety of algorithms that can grasp patterns in data [<xref ref-type="bibr" rid="ref-23">23</xref>]. These models have demonstrated significant efficacy in recreating both experimental and numerical simulations, while minimizing the time and computational expenses typically linked to conventional methods [<xref ref-type="bibr" rid="ref-24">24</xref>]. The most popular artificial intelligence approach for assessing the CS of various concrete families is ANN [<xref ref-type="bibr" rid="ref-25">25</xref>].</p>
<p>Numerous researchers have considered multiple input factors while using ML algorithms for the prediction of the CS of concrete. Xu et al. [<xref ref-type="bibr" rid="ref-26">26</xref>] have only considered water-to-cement ratio, maximum aggregate size, and aggregate-to-cement ratio as influencing factors for the prediction of CS of concrete using multiple nonlinear regression (MNLR) and artificial neural network (ANN). Dantas et al. [<xref ref-type="bibr" rid="ref-21">21</xref>] examined several input variables in the prediction of CS of recycled aggregate concrete (RAC) using artificial neural networks (ANNs): water to cement ratios, cement content, dry mortar ratio, total dry aggregate concrete, the aggregate substitution ratio of recycled fine aggregates and coarse aggregates, chemical admixture rate, composition of CDW, ratio of recycled mortar and concrete, ratio of recycled materials, fineness modulus of natural and recycled fine aggregates and coarse aggregates, maximum aggregate size of fine aggregates and coarse aggregates, water absorption rates of recycled fine aggregates (RFA) and coarse aggregates (RCA), and the compensation rates of water absorption. The analysis also incorporates the age of the concrete. Overall cement content, sand content and recycled coarse aggregates. Water content, w/c%, and substitution ratio of RCA were the factors studied as input variables by Hu et al. [<xref ref-type="bibr" rid="ref-5">5</xref>]. The fineness modulus, densities, particle sizes of coarse and natural aggregates, and age are other parameters infrequently employed by some researchers [<xref ref-type="bibr" rid="ref-25">25</xref>].</p>
<p>Many material designers and researchers employ ANNs as the approximation method for forecasting various properties of construction materials. Allegedly, ANNs mimic the learning patterns of human brains by repeating actions in different contexts and replicate these patterns to make decisions in unforeseen situations [<xref ref-type="bibr" rid="ref-27">27</xref>]. ANNs employ a learning process that assigns random values (<italic>weights</italic>) to each unit, known as a neuron, to assess the impact of input data on the prediction of an output. After acquiring and examining a signal, these artificial neurons transmit information to the next consecutive neurons connected to it. Non-linear functions produce each output by aggregating their inputs, with the &#x201C;signal&#x201D; at a connection represented as a real number. Links are referred to as edges [<xref ref-type="bibr" rid="ref-25">25</xref>]. As learning advances, the relative weights of neurons and edges are updated until the best solution&#x2014;one with the minimum mean-squared error (MSE) is found. Other evaluation parameters most frequently used are coefficient of determination (<inline-formula id="ieqn-4"><mml:math id="mml-ieqn-4"><mml:msup><mml:mi>R</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:math></inline-formula>), mean absolute error (MAE) or mean absolute percentage error (MAPE), and root mean square error (RMSE) for validation of the performance of the model [<xref ref-type="bibr" rid="ref-25">25</xref>]. A more accurate model reveals lower values of RMSE, MAE, and MSE but with a larger <inline-formula id="ieqn-5"><mml:math id="mml-ieqn-5"><mml:msup><mml:mi>R</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:math></inline-formula> indicates enhanced accuracy. The training of the ANNs is a nondeterministic process because, primarily as a result of this unpredictability in the coefficient generation, the ANNs tend to converge to one local minima/maxima from a pool of potential solutions [<xref ref-type="bibr" rid="ref-28">28</xref>]. The most crucial hyperparameters in ANN models are frequently the starting weights, learning rate, number of epochs/iterations, activation functions, number of layers, and number of neurons [<xref ref-type="bibr" rid="ref-29">29</xref>]. Different activation functions like sigmoid, hyperbolic tangent (Tanh), rectified linear unit (ReLU), leaky ReLU, exponential linear unit (ELU) are available in literature. It is apparent from literature that the sigmoid function was utilized as an activation function by the majority of researchers and material designers [<xref ref-type="bibr" rid="ref-30">30</xref>]. A feed-forward neural net (FFNN) is an architecture, in which each layer is simply coupled to its immediate neighbor layers, is the most basic architecture that an ANN model may adopt. As a result, serial and linear mapping is accomplished from the input to the output layers through the hidden layers [<xref ref-type="bibr" rid="ref-27">27</xref>]. Feedforward neural networks (FFNNs) are easier to understand than more complex neural architectures due to their simple structure, and their training process is typically straightforward and efficient. Numerous studies have demonstrated the efficacy of FFNNs, particularly in material modeling [<xref ref-type="bibr" rid="ref-22">22</xref>]. Reference [<xref ref-type="bibr" rid="ref-31">31</xref>] used ANNs to predict the concrete CS having RAs with one hidden layer and 14 input variables and model generated represents fairly high accuracy. Reference [<xref ref-type="bibr" rid="ref-32">32</xref>] forecasted the CS of RAC using a nonlinear, regression-based ML model, namely an ANN. In order to forecast the mechanical properties of RAC, reference [<xref ref-type="bibr" rid="ref-33">33</xref>] examined the suitability of a number of regression-based ML models, including least squares support vector regression (LSSVR) models, M5 model tree (M5Tree), and multivariate adaptive regression splines (MARS). CS of RAC using a convolutional ANN-based deep learning algorithm was predicted by [<xref ref-type="bibr" rid="ref-34">34</xref>]. ANNs and response surface methodology (RSM) have been used by [<xref ref-type="bibr" rid="ref-35">35</xref>] to estimate the concrete compressive strength having RAs and the outputs of these models were assessed, which revealed that the precision of ANNs is better than RSM. ANN is the most often used ML model [<xref ref-type="bibr" rid="ref-36">36</xref>], and it is anticipated that the precision of ML approaches will increase with the number of input variables and data points [<xref ref-type="bibr" rid="ref-36">36</xref>].</p>
<p>Compiling huge datasets from the literature is a hectic and tedious task. ML models demand big and premium quality datasets. Overfitting issues are common in models trained on datasets with fewer than 1000 data points [<xref ref-type="bibr" rid="ref-37">37</xref>]. Models trained on fewer than 1000 data points are particularly prone to overfitting issues [<xref ref-type="bibr" rid="ref-37">37</xref>]. According to [<xref ref-type="bibr" rid="ref-38">38</xref>], only 11% of studies applying ML to concrete science utilize datasets exceeding 1000 samples. To mitigate data scarcity in predicting concrete mix design properties, researchers have increasingly turned to generative approaches. For instance, generative adversarial networks (GANs) have been employed to generate synthetic data for the CS of concrete incorporating industrial waste materials [<xref ref-type="bibr" rid="ref-39">39</xref>]. Conditional GANs (CGANs), cycle-consistent deep GANs (CDGANs) [<xref ref-type="bibr" rid="ref-40">40</xref>], and tabular GANs (TGANs) [<xref ref-type="bibr" rid="ref-41">41</xref>,<xref ref-type="bibr" rid="ref-42">42</xref>] have been used to simulate data for ultra-high-performance concrete. Similarly, reference [<xref ref-type="bibr" rid="ref-38">38</xref>] applied TGANs to generate synthetic datasets for predicting the CS of geopolymer concrete.</p>
<p>Hence, this study presents a novel approach by simultaneously utilizing five different in-built optimizers in MATLAB Neural Net fitting module for the development of an ANN model, specifically focusing on feedforward neural networks (FFNN) due to their simplicity, ease of implementation, and time efficiency. By utilizing multiple optimizers, the study aims to enhance the performance and robustness of model. Random forest (RF) and extreme gradient (XGBoost) models were also trained and evaluated. RF functions as a bagging technique that mitigates variance by concurrently training multiple decision trees on various bootstrapped samples and averaging their outcomes. While XGBoost employs a boosting approach, constructing trees in a sequential manner where each subsequent tree aims to reduce the errors made by its predecessors, thereby minimizing bias and improving accuracy. NNs, however, differ logically from both methods, as it function as a singular, robust model that learns intricate, hierarchical patterns directly from the data by modifying interconnected layers of neurons through the process of backpropagation. Additionally, the incorporation of advanced interpolation-based data augmentation techniques allows the model to achieve high accuracy with a relatively smaller dataset collected from literature, illustrating the efficacy of techniques in refining model generalization and performance. This dual strategy of optimizer diversity and data augmentation represents a significant advancement in FFNN model development. Next, models developed utilizing synthetic data have been compared with models using original data with evaluation metrics such as R-squared (<inline-formula id="ieqn-6"><mml:math id="mml-ieqn-6"><mml:msup><mml:mi>R</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:math></inline-formula>), correlation coefficient (R), mean absolute error (MAE), mean absolute percentage error (MAPE), root mean squared error (RMSE) and mean squared error (MSE). Additionally, sensitivity analysis employing perturbation and weight-partitioning techniques was carried out to enhance understanding of the performance of model and to assess the relative importance of the input parameters. To further elucidate the influence of these variables on model output and predictions, partial dependence plots (PDP) and accumulated local effects (ALE) were incorporated, offering deeper interpretability of feature impacts.</p>
<sec id="s1_1">
<label>1.1</label>
<title>Artificial Neural Network and Training Algorithms</title>
<p>ANNs are designed mathematical frameworks that imitate the neural biological systems identical human brain, enabling them to learn and solve problems. An ANN can be conceived of as a mesh of several processors called units or neurons [<xref ref-type="bibr" rid="ref-43">43</xref>]. Structure and function of a neuron or node&#x2014;the fundamental building element of an artificial neural network&#x2014;have been explained numerous times by various researchers [<xref ref-type="bibr" rid="ref-44">44</xref>]. A connection between input and output layers, facilitated by several hidden layers, may adhere to one of the three most commonly utilized architectures: feed-forward, cascaded, and layer-recurrent neural networks [<xref ref-type="bibr" rid="ref-27">27</xref>]. An ANN which consists of three main types of layers: the input layer, one or more hidden layers, and the output layer with no cycle or loop in the connections between nodes is called a feed-forward neural network. Data travels unidirectionally from the input layer to the output layer, routing via any intervening hidden layers. The reason it is named &#x201C;feed-forward&#x201D; is that there are no feedback loops and the data moves through the network in a forward direction only. <xref ref-type="fig" rid="fig-1">Fig. 1</xref> depicts simplest ANN architecture with one hidden layer <bold>H</bold>. <italic>P</italic> number of inputs are connected to input layer <bold>I</bold> with <italic>p</italic> number of neurons, whereas outputs generated by the input layer act as a source for the hidden layer, having <italic>n</italic> number of neurons. Here <italic>f</italic> is termed as activation or threshold function, which quantizes the network output. The most typically used activation functions are step, linear, sigmoid, tansigmoid, hyperbolic tangent (tanh), or rectified linear unit (ReLU), etc. ANN training algorithms employ MSE, MAE, RMSE, and <inline-formula id="ieqn-7"><mml:math id="mml-ieqn-7"><mml:msup><mml:mi>R</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:math></inline-formula> as their evaluation parameters.
<disp-formula id="eqn-1"><label>(1)</label><mml:math id="mml-eqn-1" display="block"><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover><mml:msubsup><mml:mi>W</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mi>l</mml:mi></mml:mrow><mml:mrow><mml:mi>Y</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi>H</mml:mi><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover><mml:msubsup><mml:mi>W</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>H</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi>I</mml:mi><mml:mi>P</mml:mi></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>ANN architecture with single hidden layer</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-1.tif"/>
</fig>
<p>Choosing the number of hidden layers and neurons in each hidden layer is a core element that has a high influence on the system soundness. The training algorithm is also the crucial component of the FFNN [<xref ref-type="bibr" rid="ref-45">45</xref>]. MSE is used as the objective function in the majority of ANN training algorithms. The difference between the expected (model results) and observed (lab testing) responses is known as the MSE.</p>
<p>As far as the training of ANN with feedforward architecture is concerned, literature proposes various training algorithms. The most popularly employed training algorithms are Bayesian Regularization (BR), Conjugate Gradient with Powell/Beale Restarts (CGB), Fletcher-Powell Conjugate Gradient (FP-CG), Polak-Ribi&#x00E9;re Conjugate Gradient (PR-CG) and Levenberg-Marquardt (LM) which are also considered in the present study.</p>
</sec>
<sec id="s1_2">
<label>1.2</label>
<title>Conjugate Gradient (CG)</title>
<p>Conjugate Gradient (CG) is a frequently used iterative process for determining the values of weights and bias (parameters of ANN) and is reputable for its prompt convergence rate. Some of its editions include CGB, FP-CG, PR-CG, and scaled conjugate gradient (SCG), etc. Fundamentally training starts with initialization process where random values <inline-formula id="ieqn-8"><mml:math id="mml-ieqn-8"><mml:msub><mml:mi>x</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:math></inline-formula> are assigned to parameters linked to individual variables in input layer along with initialization of search direction as
<disp-formula id="eqn-2"><label>(2)</label><mml:math id="mml-eqn-2" display="block"><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mrow><mml:mtext mathvariant="bold">w</mml:mtext></mml:mrow><mml:mn>0</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>In each iteration, values of parameters are updated by defining the step size in terms of <inline-formula id="ieqn-9"><mml:math id="mml-ieqn-9"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:math></inline-formula> and gradient to update the search direction using following equations
<disp-formula id="eqn-3"><label>(3)</label><mml:math id="mml-eqn-3" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>w</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:msub><mml:mi>d</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="eqn-4"><label>(4)</label><mml:math id="mml-eqn-4" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:msub><mml:mi>g</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p>Basically conjugate gradient method is focused to calculate coefficient of conjugate gradient <inline-formula id="ieqn-10"><mml:math id="mml-ieqn-10"><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:math></inline-formula> which is the updated search direction in each iteration as
<disp-formula id="eqn-5"><label>(5)</label><mml:math id="mml-eqn-5" display="block"><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>g</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:msub><mml:mi>d</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:math></disp-formula></p>
<sec id="s1_2_1">
<label>1.2.1</label>
<title>Conjugate Gradient with Powell/Beale Restarts</title>
<p>In conjugate gradient with Powell/Beale Restarts iterations are resumed periodically to accelerate and to resolve probable convergence problems. The intrinsic idea is to restart the iterative process when the optimization process starts to lag. Usually a restart condition is defined based on the number of iterations, convergence threshold or step-size threshold. Algorithm 1 explains the basic steps of the conjugate gradient algorithm.</p>
<fig id="fig-18">
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-18.tif"/>
</fig>
</sec>
<sec id="s1_2_2">
<label>1.2.2</label>
<title>Fletcher-Powell Conjugate Gradient (FP-CG)</title>
<p>The Fletcher-Powell version of the conjugate gradient method uses a different formula for calculation of the conjugate gradient coefficients <inline-formula id="ieqn-14"><mml:math id="mml-ieqn-14"><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:math></inline-formula>.
<disp-formula id="eqn-6"><label>(6)</label><mml:math id="mml-eqn-6" display="block"><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:msubsup><mml:mi>f</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msubsup><mml:mo>.</mml:mo><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:msubsup><mml:mi>f</mml:mi><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msubsup><mml:mo>.</mml:mo><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:msub><mml:mi>f</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow></mml:mfrac><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:math></disp-formula></p>
</sec>
<sec id="s1_2_3">
<label>1.2.3</label>
<title>Polak-Ribi&#x00E9;re Conjugate Gradient (PR-CG)</title>
<p>This variant of conjugate gradient employs different equation for conjugacy coefficient.
<disp-formula id="eqn-7"><label>(7)</label><mml:math id="mml-eqn-7" display="block"><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:msub><mml:mi>&#x03B2;</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:msub><mml:mi>f</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mi>T</mml:mi></mml:msup><mml:mo>.</mml:mo><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:msubsup><mml:mi>f</mml:mi><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msubsup><mml:mo>.</mml:mo><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:msub><mml:mi>f</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow></mml:mfrac><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:math></disp-formula></p>
</sec>
</sec>
<sec id="s1_3">
<label>1.3</label>
<title>Bayesian Regularization</title>
<p>The regularization parameters tuning with Bayesian requires calculation of minimum point <inline-formula id="ieqn-15"><mml:math id="mml-ieqn-15"><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> from Hessian matrix of a given function <italic>F</italic>(<italic>w</italic>) [<xref ref-type="bibr" rid="ref-46">46</xref>]. The aim of the Hessian matrix-based update in Bayesian regularization is to include the curvature of the loss function in the weight updates. The update rule in consecutive iterations can be expressed as follows:
<disp-formula id="eqn-8"><label>(8)</label><mml:math id="mml-eqn-8" display="block"><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:msub><mml:mi>W</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mo stretchy="false">[</mml:mo><mml:msup><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mi>T</mml:mi></mml:msup><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03B6;</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mi>I</mml:mi><mml:msup><mml:mo stretchy="false">]</mml:mo><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:msup><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mi>T</mml:mi></mml:msup><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mi>V</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>W</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:math></disp-formula>where <inline-formula id="ieqn-16"><mml:math id="mml-ieqn-16"><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>W</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> is a Jacobian Matrix and <inline-formula id="ieqn-17"><mml:math id="mml-ieqn-17"><mml:mi>&#x03B6;</mml:mi></mml:math></inline-formula> is a step controlling parameter and <inline-formula id="ieqn-18"><mml:math id="mml-ieqn-18"><mml:msub><mml:mi>V</mml:mi><mml:mi>w</mml:mi></mml:msub></mml:math></inline-formula> is the sum of squares function:
<disp-formula id="eqn-9"><label>(9)</label><mml:math id="mml-eqn-9" display="block"><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:mi>F</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>w</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:msubsup><mml:mi>V</mml:mi><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:msubsup></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>W</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:msup><mml:mi>V</mml:mi><mml:mi>T</mml:mi></mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mi>W</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mi>V</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>W</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>The regularization parameters <inline-formula id="ieqn-19"><mml:math id="mml-ieqn-19"><mml:mi mathvariant="normal">&#x03A6;</mml:mi></mml:math></inline-formula> and <inline-formula id="ieqn-20"><mml:math id="mml-ieqn-20"><mml:mi mathvariant="normal">&#x03A8;</mml:mi></mml:math></inline-formula> with relations are defined as;
<disp-formula id="eqn-10"><label>(10)</label><mml:math id="mml-eqn-10" display="block"><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:msub><mml:mi mathvariant="normal">&#x03A6;</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mi>&#x03BE;</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:msub><mml:mi mathvariant="normal">&#x03A5;</mml:mi><mml:mi>D</mml:mi></mml:msub><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:math></disp-formula>
<disp-formula id="eqn-11"><label>(11)</label><mml:math id="mml-eqn-11" display="block"><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:msub><mml:mi mathvariant="normal">&#x03A8;</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>N</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>&#x03BE;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:msub><mml:mi mathvariant="normal">&#x03A5;</mml:mi><mml:mi>D</mml:mi></mml:msub><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:math></disp-formula>where <inline-formula id="ieqn-21"><mml:math id="mml-ieqn-21"><mml:mi>&#x03BE;</mml:mi></mml:math></inline-formula> is number of effective parameters used to reduce error function defined as;
<disp-formula id="eqn-12"><label>(12)</label><mml:math id="mml-eqn-12" display="block"><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:mi>&#x03BE;</mml:mi><mml:mo>=</mml:mo><mml:mi>n</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>2</mml:mn><mml:msub><mml:mi mathvariant="normal">&#x03A6;</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mrow><mml:mtext mathvariant="bold">H</mml:mtext></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:math></disp-formula>where <italic>n</italic> is the total number of parameters in network and <inline-formula id="ieqn-22"><mml:math id="mml-ieqn-22"><mml:mrow><mml:mtext mathvariant="bold">H</mml:mtext></mml:mrow></mml:math></inline-formula> is the Hessian matrix which can be calculated as;
<disp-formula id="eqn-13"><label>(13)</label><mml:math id="mml-eqn-13" display="block"><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:mrow><mml:mtext mathvariant="bold">H</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mn>2</mml:mn><mml:mi mathvariant="normal">&#x03A8;</mml:mi><mml:msup><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mo>+</mml:mo><mml:mn>2</mml:mn><mml:mi mathvariant="normal">&#x03A6;</mml:mi><mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>With updated values of <inline-formula id="ieqn-23"><mml:math id="mml-ieqn-23"><mml:mi mathvariant="normal">&#x03A6;</mml:mi></mml:math></inline-formula> and <inline-formula id="ieqn-24"><mml:math id="mml-ieqn-24"><mml:mi mathvariant="normal">&#x03A8;</mml:mi></mml:math></inline-formula>, the objective function can be defined as;
<disp-formula id="eqn-14"><label>(14)</label><mml:math id="mml-eqn-14" display="block"><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:mi>F</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>W</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mi mathvariant="normal">&#x03A8;</mml:mi><mml:msub><mml:mi mathvariant="normal">&#x03A5;</mml:mi><mml:mi>D</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:mi mathvariant="normal">&#x03A6;</mml:mi><mml:msub><mml:mi mathvariant="normal">&#x03A5;</mml:mi><mml:mi>E</mml:mi></mml:msub><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>Algorithm 2 summarizes the fundamental steps of the Bayesian regularization algorithm.</p>
<fig id="fig-19">
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-19.tif"/>
</fig>
</sec>
<sec id="s1_4">
<label>1.4</label>
<title>Levenberg-Marquardt Framework</title>
<p>The Levenberg-Marquardt (LM) algorithm is an optimization technique most frequently used to solve nonlinear least squares problems that aspires to reduce the sum of the squares of the variations between observed and predicted values and usually works in conjunction with the steepest descent method [<xref ref-type="bibr" rid="ref-27">27</xref>]. This algorithm starts with assigning of initial values of hyper parameters (weights and bias) and a small positive value for the damping component <inline-formula id="ieqn-30"><mml:math id="mml-ieqn-30"><mml:mi>&#x03BB;</mml:mi></mml:math></inline-formula>. This algorithm reportedly offers faster convergence, improved stability and owing to adaptive damping, it is reliably more resilient when data is noisy or contains outliers. Let us assess the output return of FFNN, calculated using <xref ref-type="disp-formula" rid="eqn-1">(1)</xref>, where initial output response is given as <inline-formula id="ieqn-31"><mml:math id="mml-ieqn-31"><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and the error of the network is calculated using <xref ref-type="disp-formula" rid="eqn-29">(29)</xref>. Originated from the steepest descent method, the rule to update LM algorithm is defined as;
<disp-formula id="eqn-15"><label>(15)</label><mml:math id="mml-eqn-15" display="block"><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mi>W</mml:mi><mml:mo>=</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:msubsup><mml:mi>J</mml:mi><mml:mrow><mml:mi>w</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mi>T</mml:mi></mml:msubsup><mml:msub><mml:mi>J</mml:mi><mml:mrow><mml:mi>w</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03B4;</mml:mi><mml:mi>r</mml:mi></mml:msub><mml:mi>I</mml:mi><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:msubsup><mml:mi>J</mml:mi><mml:mrow><mml:mi>w</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mi>T</mml:mi></mml:msubsup><mml:mi>r</mml:mi></mml:math></disp-formula>and updated parameters in each iteration can be given as;
<disp-formula id="eqn-16"><label>(16)</label><mml:math id="mml-eqn-16" display="block"><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mo stretchy="false">[</mml:mo><mml:msubsup><mml:mi>J</mml:mi><mml:mrow><mml:mi>w</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mi>T</mml:mi></mml:msubsup><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mi>J</mml:mi><mml:mrow><mml:mi>w</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mi>x</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03B4;</mml:mi><mml:mi>r</mml:mi></mml:msub><mml:mi>I</mml:mi><mml:msup><mml:mo stretchy="false">]</mml:mo><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:msubsup><mml:mi>J</mml:mi><mml:mrow><mml:mi>w</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mi>T</mml:mi></mml:msubsup><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mi>v</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></disp-formula>where <inline-formula id="ieqn-32"><mml:math id="mml-ieqn-32"><mml:msub><mml:mi>J</mml:mi><mml:mrow><mml:mi>w</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> has dimensions of <inline-formula id="ieqn-33"><mml:math id="mml-ieqn-33"><mml:mo stretchy="false">(</mml:mo><mml:mi>P</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>Q</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>R</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> and the matrix of dimensions <inline-formula id="ieqn-34"><mml:math id="mml-ieqn-34"><mml:mo stretchy="false">(</mml:mo><mml:mi>P</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>Q</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> is defined as error matrix. So the Jacobian matrix is established using the following association:
<disp-formula id="eqn-17"><label>(17)</label><mml:math id="mml-eqn-17" display="block"><mml:msub><mml:mrow><mml:mtext mathvariant="bold">J</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mtext mathvariant="bold">wr</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mtable columnalign="center center center center center" rowspacing="4pt" columnspacing="1em"><mml:mtr><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mn>11</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mn>11</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mo>&#x2026;</mml:mo></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mn>11</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mi>R</mml:mi></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mn>11</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>b</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mn>12</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mn>12</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mo>&#x2026;</mml:mo></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mn>12</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mi>R</mml:mi></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mn>12</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>b</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mi>Q</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mi>Q</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mo>&#x2026;</mml:mo></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mi>Q</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mi>R</mml:mi></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mi>Q</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>b</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mo>&#x2026;</mml:mo></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mi>R</mml:mi></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>b</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mo>&#x2026;</mml:mo></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mi>R</mml:mi></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>b</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EE;</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mi>Q</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mi>Q</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mo>&#x2026;</mml:mo></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mi>Q</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>w</mml:mi><mml:mi>R</mml:mi></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>P</mml:mi><mml:mi>Q</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>b</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mstyle></mml:mtd></mml:mtr></mml:mtable><mml:mo>]</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>Here, <italic>R</italic> denotes the total number of weights and elements in the error vector, while <italic>r</italic> is computed using <xref ref-type="disp-formula" rid="eqn-17">Eq. (17)</xref>. <italic>P</italic> represents the number of training patterns, each associated with <italic>Q</italic> output values. Traditionally, the Jacobian matrix <inline-formula id="ieqn-35"><mml:math id="mml-ieqn-35"><mml:msub><mml:mi>J</mml:mi><mml:mrow><mml:mi>w</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is first computed, and then calculations for updating weights and biases are carried out using values that have been stored. This approach operates smoothly and effectively with lesser patterns, but memory constraints arise while calculating the Jacobin matrix for big patterns. Algorithm 3 enlists the fundamental steps of the standard LM algorithm that generates the Jacobian matrix and later calculates the sensitivities. We can also conclude that the LM algorithm&#x2019;s performance decreases with large training sets.</p>
<fig id="fig-20">
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-20.tif"/>
</fig>
</sec>
<sec id="s1_5">
<label>1.5</label>
<title>XGBoost Model</title>
<p>Boosting is an ensemble learning strategy which generates a powerful regression model by combining several weak learners with the idea that every new model is trained to reduce the errors of the ones that were trained before. The final ensemble model achieves excellent accuracy and robustness by upgrading one step at a time. It greatly enhances traditional gradient-boosting models by integrating second-order gradient information, hence refining the optimization process [<xref ref-type="bibr" rid="ref-47">47</xref>]. The training of the dataset <inline-formula id="ieqn-40"><mml:math id="mml-ieqn-40"><mml:mo stretchy="false">(</mml:mo><mml:mi>X</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> starts with defining the number of boosting rounds <italic>M</italic>, the learning rate <inline-formula id="ieqn-41"><mml:math id="mml-ieqn-41"><mml:mi>&#x03B7;</mml:mi></mml:math></inline-formula>, and a specified loss function <italic>L</italic>. The model is first initialized with a constant prediction that minimizes the loss:
<disp-formula id="eqn-18"><label>(18)</label><mml:math id="mml-eqn-18" display="block"><mml:mrow><mml:msub><mml:mi>F</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mi>arg</mml:mi><mml:mo>&#x2061;</mml:mo><mml:munder><mml:mo movablelimits="true" form="prefix">min</mml:mo><mml:mi>c</mml:mi></mml:munder><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:mi>L</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:mi>c</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>As the error is measured in terms of mean squared error (MSE):
<disp-formula id="eqn-19"><label>(19)</label><mml:math id="mml-eqn-19" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:msub><mml:mi>F</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mi>arg</mml:mi><mml:mo>&#x2061;</mml:mo><mml:munder><mml:mo movablelimits="true" form="prefix">min</mml:mo><mml:mi>c</mml:mi></mml:munder><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mi>c</mml:mi><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mn>2</mml:mn></mml:msup></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mfrac><mml:mi>d</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>c</mml:mi></mml:mrow></mml:mfrac><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mi>c</mml:mi><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mn>2</mml:mn></mml:msup><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mn>2</mml:mn><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>+</mml:mo><mml:mn>2</mml:mn><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mi>c</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>n</mml:mi></mml:munderover><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mi></mml:mi><mml:mo>&#x2234;</mml:mo><mml:mspace width="1em" /><mml:msub><mml:mi>F</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p>At each boosting round, a new decision tree is constructed to correct the errors of the current model. For each data point <inline-formula id="ieqn-42"><mml:math id="mml-ieqn-42"><mml:mi>i</mml:mi></mml:math></inline-formula>, compute the gradient as first derivative and hessian as second derivative of the loss function with respect to the current prediction:
<disp-formula id="eqn-20"><label>(20)</label><mml:math id="mml-eqn-20" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:mi>L</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mfrac></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="eqn-21"><label>(21)</label><mml:math id="mml-eqn-21" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:mn>2</mml:mn></mml:msup><mml:mi>L</mml:mi><mml:mspace width="negativethinmathspace" /><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p>A classification and regression tree (CART) decision tree is built using the computed gradients and hessians from <xref ref-type="disp-formula" rid="eqn-20">(20)</xref> and <xref ref-type="disp-formula" rid="eqn-21">(21)</xref>.</p>
<p>For each candidate split, the gain is calculated as:
<disp-formula id="eqn-22"><label>(22)</label><mml:math id="mml-eqn-22" display="block"><mml:mrow><mml:mtext>Gain</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:mrow><mml:mo>[</mml:mo><mml:mfrac><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x2211;</mml:mo><mml:mi>g</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup><mml:mrow><mml:mo>&#x2211;</mml:mo><mml:mi>h</mml:mi><mml:mo>+</mml:mo><mml:mi>&#x03BB;</mml:mi></mml:mrow></mml:mfrac><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x2211;</mml:mo><mml:msub><mml:mi>g</mml:mi><mml:mrow><mml:mrow><mml:mtext>left</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup><mml:mrow><mml:mo>&#x2211;</mml:mo><mml:msub><mml:mi>h</mml:mi><mml:mrow><mml:mrow><mml:mtext>left</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:mi>&#x03BB;</mml:mi></mml:mrow></mml:mfrac><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x2211;</mml:mo><mml:msub><mml:mi>g</mml:mi><mml:mrow><mml:mrow><mml:mtext>right</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup><mml:mrow><mml:mo>&#x2211;</mml:mo><mml:msub><mml:mi>h</mml:mi><mml:mrow><mml:mrow><mml:mtext>right</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:mi>&#x03BB;</mml:mi></mml:mrow></mml:mfrac><mml:mo>]</mml:mo></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>&#x03B3;</mml:mi></mml:math></disp-formula>where <inline-formula id="ieqn-43"><mml:math id="mml-ieqn-43"><mml:mi>&#x03BB;</mml:mi></mml:math></inline-formula> is the regularization term on leaf weights, and <inline-formula id="ieqn-44"><mml:math id="mml-ieqn-44"><mml:mi>&#x03B3;</mml:mi></mml:math></inline-formula> is the penalty for additional leaves. The split with the highest gain is chosen. For each terminal leaf <inline-formula id="ieqn-45"><mml:math id="mml-ieqn-45"><mml:mi>j</mml:mi></mml:math></inline-formula>, the optimal weight is computed as:
<disp-formula id="eqn-23"><label>(23)</label><mml:math id="mml-eqn-23" display="block"><mml:mrow><mml:msub><mml:mi>w</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mrow><mml:mo>&#x2211;</mml:mo><mml:mi>g</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2211;</mml:mo><mml:mi>h</mml:mi><mml:mo>+</mml:mo><mml:mi>&#x03BB;</mml:mi></mml:mrow></mml:mfrac></mml:mrow></mml:math></disp-formula></p>
<p>This determines the prediction adjustment that the leaf contributes. The new tree is added to the model with learning rate <inline-formula id="ieqn-46"><mml:math id="mml-ieqn-46"><mml:mi>&#x03B7;</mml:mi></mml:math></inline-formula>:
<disp-formula id="eqn-24"><label>(24)</label><mml:math id="mml-eqn-24" display="block"><mml:mrow><mml:msub><mml:mi>F</mml:mi><mml:mi>m</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>+</mml:mo><mml:mi>&#x03B7;</mml:mi><mml:mo>&#x22C5;</mml:mo><mml:msub><mml:mrow><mml:mtext>Tree</mml:mtext></mml:mrow><mml:mi>m</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>After <italic>M</italic> iterations, the final boosted model is obtained:
<disp-formula id="eqn-25"><label>(25)</label><mml:math id="mml-eqn-25" display="block"><mml:mrow><mml:msub><mml:mi>F</mml:mi><mml:mi>M</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>This ensemble of decision trees provides strong predictive power, balancing accuracy and generalization through regularization and learning rate scaling. <xref ref-type="fig" rid="fig-2">Fig. 2</xref> describes the detailed workflow of the development of XGBoost regression model, whereas the fundamental steps are given in Algorithm 4.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>XGBoost Workflow (Newly added figure)</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-2.tif"/>
</fig>
<fig id="fig-21">
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-21.tif"/>
</fig>
</sec>
<sec id="s1_6">
<label>1.6</label>
<title>Random Forest Model</title>
<p>Random forest (RF) is an ensemble techniques developed for classification and regression by Leo Breiman at the University of California, Berkeley [<xref ref-type="bibr" rid="ref-48">48</xref>]. The training of the dataset <inline-formula id="ieqn-57"><mml:math id="mml-ieqn-57"><mml:mo stretchy="false">(</mml:mo><mml:mi>X</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> commences with the specification of the number of trees <italic>M</italic>, the number of features to sample <inline-formula id="ieqn-58"><mml:math id="mml-ieqn-58"><mml:mi>m</mml:mi></mml:math></inline-formula> at each split, and the number of bootstrap samples per tree <inline-formula id="ieqn-59"><mml:math id="mml-ieqn-59"><mml:mi>n</mml:mi></mml:math></inline-formula>. Draw a bootstrap sample of size <inline-formula id="ieqn-60"><mml:math id="mml-ieqn-60"><mml:mi>n</mml:mi></mml:math></inline-formula> from the training set <inline-formula id="ieqn-61"><mml:math id="mml-ieqn-61"><mml:mo stretchy="false">(</mml:mo><mml:mi>X</mml:mi><mml:mo>,</mml:mo><mml:mi>y</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> for every tree <inline-formula id="ieqn-62"><mml:math id="mml-ieqn-62"><mml:mi>t</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:mi>M</mml:mi></mml:math></inline-formula>. As every tree does not see the same data, this adds randomization and lowers variance. To grow each decision tree, the bootstrap sample is used. Choose a random subset of <inline-formula id="ieqn-63"><mml:math id="mml-ieqn-63"><mml:mi>m</mml:mi></mml:math></inline-formula> features from the entire feature set at each split node. As a result, there is less association between trees and tree diversity is guaranteed. For regression, the split is chosen to minimize the MSE. The impurity at a node with dataset <italic>S</italic> is measured as:
<disp-formula id="eqn-26"><label>(26)</label><mml:math id="mml-eqn-26" display="block"><mml:mrow><mml:mrow><mml:mtext>MSE</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mi>S</mml:mi><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:munder><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:munder><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow><mml:mi>S</mml:mi></mml:msub><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:math></disp-formula>where <inline-formula id="ieqn-64"><mml:math id="mml-ieqn-64"><mml:msub><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow><mml:mi>S</mml:mi></mml:msub></mml:math></inline-formula> is the mean of the target values in <italic>S</italic>. The best split is the one that maximizes the reduction in impurity:
<disp-formula id="eqn-27"><label>(27)</label><mml:math id="mml-eqn-27" display="block"><mml:mrow><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mtext>MSE</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mfrac><mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mrow><mml:mtext>left</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mi>S</mml:mi><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:mrow><mml:mtext>MSE</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mrow><mml:mtext>left</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>+</mml:mo><mml:mfrac><mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mrow><mml:mtext>right</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mi>S</mml:mi><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:mrow><mml:mtext>MSE</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>S</mml:mi><mml:mrow><mml:mrow><mml:mtext>right</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:math></disp-formula></p>
<p>In its terminal leaf, each tree outputs the mean value of the training samples in order to make a prediction. The average of each tree yields the RF prediction:
<disp-formula id="eqn-28"><label>(28)</label><mml:math id="mml-eqn-28" display="block"><mml:mrow><mml:mfrac><mml:mn>1</mml:mn><mml:mi>M</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>M</mml:mi></mml:munderover><mml:msub><mml:mrow><mml:mtext>Tree</mml:mtext></mml:mrow><mml:mi>t</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>After <italic>M</italic> trees, the final trained random forest regression model <inline-formula id="ieqn-65"><mml:math id="mml-ieqn-65"><mml:msub><mml:mi>F</mml:mi><mml:mi>M</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> is obtained.</p>
<p>Algorithm 5 explains the fundamental steps of the random forest algorithm&#x2014;starting from the initialization to the final output of all tree predictions.</p>
<fig id="fig-22">
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-22.tif"/>
</fig>
</sec>
</sec>
<sec id="s2">
<label>2</label>
<title>Methodology</title>
<p><xref ref-type="fig" rid="fig-3">Fig. 3</xref> summarizes the research technique used to develop a predictive model for determining CS of concrete having RAs.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>Methodology for concrete compressive strength prediction using FFNN</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-3.tif"/>
</fig>
<sec id="s2_1">
<label>2.1</label>
<title>Dataset Description</title>
<p>Concrete with RAs has distinct fresh and hardened characteristics from concrete manufactured using NAs. When compared with NAs, RAs absorb more water and have lower density [<xref ref-type="bibr" rid="ref-9">9</xref>]. Other possible reasons for reduced strength in concrete are the presence of residual mortar on the aggregate surface, poor gradation due to improper crushing, age, and strength of the demolished structure, etc. Owing to these factors, with an increase in the RA content, the mechanical characteristics decrease [<xref ref-type="bibr" rid="ref-49">49</xref>]. Thus, the fundamental elements of regular concrete, which are cement, water, sand, superplasticizer, and coarse aggregates (in terms of their bulk density, water absorption capacities, and size), along with other parameters describing the properties of RAs, i.e., size, density, and water absorption of RAs, are considered input factors influencing the hardened concrete CS. Present research considers a data set that includes 501 data points collected from the published literature. The statistics of the input and output variables (mean, median, mode, standard deviation, minimum, and maximum) are listed in <xref ref-type="table" rid="table-1">Table 1</xref>. The distribution of the data used is shown in <xref ref-type="fig" rid="fig-4">Fig. 4</xref>. The replacement ratio of RAs with natural coarse aggregates (NCA) mainly takes values around 0 %, 25%, 50%, and 100%. The 0% indicates samples having 100% natural aggregates as reference samples. The bulk density of recycled coarse aggregate and the bulk density of natural coarse aggregates ranges from 2200 to 2800 kg/m<sup>3</sup>; the water absorption capacity of recycled coarse aggregates is mainly distributed in 2%&#x2013;10% while the water absorption of natural is distributed in 0.2%&#x2013;3%.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>Dataset statistics</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th rowspan="2">Type</th>
<th align="center" rowspan="2">Variable names</th>
<th colspan="6">Initial data</th>
<th colspan="6">Augmented data</th>
</tr>
<tr>
<th>Mean</th>
<th>Median</th>
<th>Mode</th>
<th>SD</th>
<th>Min</th>
<th>Max</th>
<th>Mean</th>
<th>Median</th>
<th>Mode</th>
<th>SD</th>
<th>Min</th>
<th>Max</th>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="14">Input</td>
<td>Replacement Level %</td>
<td>57.26</td>
<td>50</td>
<td>100</td>
<td>36.81</td>
<td>0</td>
<td>100</td>
<td>57.09</td>
<td>50</td>
<td>100</td>
<td>35.46</td>
<td>0</td>
<td>100</td>
</tr>
<tr>
<td>w/c (%)</td>
<td>0.50</td>
<td>0.49</td>
<td>0.5</td>
<td>0.11</td>
<td>0.23</td>
<td>1.02</td>
<td>0.495</td>
<td>0.498</td>
<td>0.5</td>
<td>0.1084</td>
<td>0.23</td>
<td>1.02</td>
</tr>
<tr>
<td>Water (kg/m<sup>3</sup>)</td>
<td>183.98</td>
<td>180</td>
<td>180</td>
<td>26.78</td>
<td>117.6</td>
<td>271</td>
<td>183.796</td>
<td>180</td>
<td>180</td>
<td>25.964</td>
<td>117.6</td>
<td>271</td>
</tr>
<tr>
<td>Cement (kg/m<sup>3</sup>)</td>
<td>383.12</td>
<td>380</td>
<td>380</td>
<td>78.97</td>
<td>210</td>
<td>600</td>
<td>384.14</td>
<td>380</td>
<td>380</td>
<td>77.95</td>
<td>210</td>
<td>600</td>
</tr>
<tr>
<td>Sand (kg/m<sup>3</sup>)</td>
<td>712.53</td>
<td>708</td>
<td>693</td>
<td>177.50</td>
<td>0</td>
<td>1800</td>
<td>712.134</td>
<td>709.02</td>
<td>693</td>
<td>175.313</td>
<td>0</td>
<td>1800</td>
</tr>
<tr>
<td>NCA (kg/m<sup>3</sup>)</td>
<td>492.24</td>
<td>506</td>
<td>0</td>
<td>428.59</td>
<td>0</td>
<td>1800</td>
<td>492.024</td>
<td>504.0</td>
<td>0</td>
<td>417.422</td>
<td>0</td>
<td>1800</td>
</tr>
<tr>
<td>RCA (kg/m<sup>3</sup>)</td>
<td>578.55</td>
<td>523</td>
<td>0</td>
<td>384.90</td>
<td>0</td>
<td>1778</td>
<td>573.071</td>
<td>524.55</td>
<td>0</td>
<td>380.282</td>
<td>0</td>
<td>1778</td>
</tr>
<tr>
<td>SP (kg/m<sup>3</sup>)</td>
<td>0.82</td>
<td>0</td>
<td>0</td>
<td>1.73</td>
<td>0</td>
<td>7.8</td>
<td>0.861</td>
<td>0</td>
<td>0</td>
<td>1.743</td>
<td>0</td>
<td>7.8</td>
</tr>
<tr>
<td><inline-formula id="ieqn-78"><mml:math id="mml-ieqn-78"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mi>R</mml:mi><mml:mi>C</mml:mi><mml:mi>A</mml:mi></mml:math></inline-formula> (mm)</td>
<td>17.99</td>
<td>20</td>
<td>20</td>
<td>7.12</td>
<td>0</td>
<td>38</td>
<td>17.841</td>
<td>19</td>
<td>20</td>
<td>7.06</td>
<td>0</td>
<td>38</td>
</tr>
<tr>
<td><inline-formula id="ieqn-79"><mml:math id="mml-ieqn-79"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mi>N</mml:mi><mml:mi>C</mml:mi><mml:mi>A</mml:mi></mml:math></inline-formula> (mm)</td>
<td>12.03</td>
<td>19</td>
<td>20</td>
<td>10.23</td>
<td>0</td>
<td>38</td>
<td>12.152</td>
<td>19</td>
<td>20</td>
<td>9.85</td>
<td>0</td>
<td>38</td>
</tr>
<tr>
<td><inline-formula id="ieqn-80"><mml:math id="mml-ieqn-80"><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mi>R</mml:mi><mml:mi>C</mml:mi><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> (%)</td>
<td>4.74</td>
<td>4.9</td>
<td>5.3</td>
<td>2.49</td>
<td>0</td>
<td>14.9</td>
<td>4.68</td>
<td>4.9</td>
<td>5.3</td>
<td>2.43</td>
<td>0</td>
<td>14.9</td>
</tr>
<tr>
<td><inline-formula id="ieqn-81"><mml:math id="mml-ieqn-81"><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mi>N</mml:mi><mml:mi>C</mml:mi><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> (%)</td>
<td>0.84</td>
<td>0.7</td>
<td>0</td>
<td>1.08</td>
<td>0</td>
<td>6.28</td>
<td>0.85</td>
<td>0.7</td>
<td>0</td>
<td>1.05</td>
<td>0</td>
<td>6.28</td>
</tr>
<tr>
<td><inline-formula id="ieqn-82"><mml:math id="mml-ieqn-82"><mml:mi>S</mml:mi><mml:mi>S</mml:mi><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>N</mml:mi><mml:mi>C</mml:mi><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> (kg/m<sup>3</sup>)</td>
<td>1580.23</td>
<td>2545</td>
<td>0</td>
<td>1277.00</td>
<td>0</td>
<td>2834</td>
<td>1601.97</td>
<td>2540</td>
<td>0</td>
<td>1233.43</td>
<td>0</td>
<td>2834</td>
</tr>
<tr>
<td><inline-formula id="ieqn-84"><mml:math id="mml-ieqn-84"><mml:mi>S</mml:mi><mml:mi>S</mml:mi><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>R</mml:mi><mml:mi>C</mml:mi><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> (kg/m<sup>3</sup>)</td>
<td>2116.71</td>
<td>2350</td>
<td>2320</td>
<td>742.95</td>
<td>0</td>
<td>2705</td>
<td>2097.59</td>
<td>2340</td>
<td>2320</td>
<td>736.51</td>
<td>0</td>
<td>2705</td>
</tr>
<tr>
<td>Output</td>
<td>Compressive Strength (MPa)</td>
<td>43.25</td>
<td>42.8</td>
<td>41</td>
<td>12.48</td>
<td>13.4</td>
<td>80.8</td>
<td>43.38</td>
<td>42.9</td>
<td>41</td>
<td>12.17</td>
<td>13.4</td>
<td>80.8</td>
</tr>
</tbody>
</table>
</table-wrap><fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>Histograms of the variables in the dataset</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-4.tif"/>
</fig>
<p><xref ref-type="fig" rid="fig-5">Fig. 5</xref> demonstrates the multiple correlation matrix of the input features and output used in the present study. The Pearson correlation coefficient between variables is displayed visually in each cell of the heatmap, which offers a depiction of the correlation matrix [<xref ref-type="bibr" rid="ref-50">50</xref>]. The color tones represent the strength of the relationships. Most of the correlations between variables are pretty weak (less than 0.5), but some have strong correlations, like the highest correlation found between R.L% and content of RAs (R &#x003D; 0.94). Similarly, water absorption capacity, density, and content of RAs are highly correlated, and the same is visible for water absorption capacity, density, and content of NAs. Since ANNs can handle correlated features without compromising model stability, and because their inclusion enables the network to learn complex, nonlinear interactions that could be lost if variables were removed, multicollinearity among some input parameters was kept in place despite their existence.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Correlation matrix for variables</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-5.tif"/>
</fig>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Data Preprocessing</title>
<p>While training ANNs, normalization is crucial since it assure that all features contribute equitably to the learning process and avoids some attributes from dominating others due to disparities in their magnitude.</p>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Model Development</title>
<p>The evolution of the prediction model intricate numerous intervals, each targeted on improving the performance of model.</p>
<sec id="s2_3_1">
<label>2.3.1</label>
<title>Grid Search for Hyperparameter Tuning</title>
<p>The formulation of ML models entails identifying the ideal values of hyperparameters for model [<xref ref-type="bibr" rid="ref-14">14</xref>]. To achieve best combination of hyperparameters such as the number of layers and their respective neurons, and the training function, a grid search approach was employed. Grid search is a comprehensive, autonomous search technique that determines the optimal performance by analyzing each potential hyperparameter combination in the search space [<xref ref-type="bibr" rid="ref-51">51</xref>].</p>
</sec>
<sec id="s2_3_2">
<label>2.3.2</label>
<title>Dropout Regularization</title>
<p>To prevent overfitting and increase generalization, dropout regularization with 20% probability was utilized. Dropout regularization involves dropping a fixed percentage of neurons to prevent the neural network from memorizing the learning patterns and to make it more robust, allowing it to generalize better to new, unseen data.</p>
</sec>
<sec id="s2_3_3">
<label>2.3.3</label>
<title>Early Stopping</title>
<p>Early stopping is an approach for reducing excessive learning that involves identifying the time at which overfitting begins during training of neural network model via cross-validation [<xref ref-type="bibr" rid="ref-52">52</xref>]. To reduce overfitting and improve training efficiency, early stopping criterion was used. Training was ended if the performance of model did not improve for 10 consecutive epochs.</p>
</sec>
<sec id="s2_3_4">
<label>2.3.4</label>
<title>Data Augmentation</title>
<p>Due to the limited dataset size and the desire to improve model generalization, data augmentation is usually performed by various researchers. Reference [<xref ref-type="bibr" rid="ref-18">18</xref>] used data augmentation technique to forecast CS of calcined clay cements using linear regression. Reference [<xref ref-type="bibr" rid="ref-53">53</xref>] used data enhancement technique in Convolutional Neural Networks (CNNs) training to proficiently develop a synthetic dataset of concrete cracks in concrete. Using this technique effectively increased the dataset size preserving the intrinsic associations between features and the target variable.</p>
<p>By using linear interpolation between preexisting data points, a data augmentation technique was used to expand the dataset size while maintaining its statistical characteristics. This produced new samples that stayed within the observed range of variables. Clipping with predetermined conditions was used to avoid generation of absurd data points. The replacement level (R.L%), was tightly limited to a range of 0%&#x2013;100%. Additionally, all characteristics of RCAs were set to zero when NCAs were utilized without any replacement. On the other hand, the NCA attributes were set to zero when RCAs were utilized exclusively. These boundary specifications preserved the overall distributions and correlations of the original dataset while ensuring that the augmented data remained logically meaningful and statistically representative.</p>
<p>Independent two-sample <italic>t</italic>-tests and Kolmogorov-Smirnov (KS) tests were performed on each numerical feature to ensure that the augmented dataset preserved the statistical consistency with the original dataset. The <inline-formula id="ieqn-86"><mml:math id="mml-ieqn-86"><mml:mi>t</mml:mi></mml:math></inline-formula>-tests assessed differences in feature means, while the KS tests evaluated the similarity of their overall distributions.</p>
<p>Both the <italic>t</italic>-tests and KS-tests as shown in <xref ref-type="table" rid="table-2">Table 2</xref> yielded non-significant results (<italic>p</italic> &#x003E; 0.05 for all variables), confirming that the augmented dataset is statistically indistinguishable from the original.</p>
<table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>Statistical comparison of original and augmented datasets using independent <italic>t</italic>-test and Kolmogorov-Smirnov (KS) test</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Variable</th>
<th><italic>t</italic>-stat</th>
<th><italic>p</italic>-value (<italic>t</italic>-test)</th>
<th>KS-stat</th>
<th><italic>p</italic>-value (KS-test)</th>
</tr>
</thead>
<tbody>
<tr>
<td>R.L (%)</td>
<td>&#x2212;0.227</td>
<td>0.821</td>
<td>0.036</td>
<td>0.633</td>
</tr>
<tr>
<td>w/c ratio</td>
<td>0.191</td>
<td>0.849</td>
<td>0.014</td>
<td>1.000</td>
</tr>
<tr>
<td>Water (kg/m<sup>3</sup>)</td>
<td>0.068</td>
<td>0.945</td>
<td>0.016</td>
<td>0.999</td>
</tr>
<tr>
<td>Cement (kg/m<sup>3</sup>)</td>
<td>&#x2212;0.161</td>
<td>0.872</td>
<td>0.016</td>
<td>0.999</td>
</tr>
<tr>
<td>Sand (kg/m<sup>3</sup>)</td>
<td>&#x2212;0.045</td>
<td>0.965</td>
<td>0.011</td>
<td>1.000</td>
</tr>
<tr>
<td>NCA (kg/m<sup>3</sup>)</td>
<td>0.230</td>
<td>0.819</td>
<td>0.031</td>
<td>0.802</td>
</tr>
<tr>
<td>RCA (kg/m<sup>3</sup>)</td>
<td>&#x2212;0.204</td>
<td>0.839</td>
<td>0.036</td>
<td>0.622</td>
</tr>
<tr>
<td>SP (kg/m<sup>3</sup>)</td>
<td>&#x2212;0.171</td>
<td>0.864</td>
<td>0.009</td>
<td>1.000</td>
</tr>
<tr>
<td>Dmax_RCA (mm)</td>
<td>&#x2212;0.070</td>
<td>0.944</td>
<td>0.015</td>
<td>1.000</td>
</tr>
<tr>
<td>Dmax_NCA (mm)</td>
<td>0.090</td>
<td>0.928</td>
<td>0.029</td>
<td>0.852</td>
</tr>
<tr>
<td>WRCA (%)</td>
<td>0.140</td>
<td>0.889</td>
<td>0.016</td>
<td>0.999</td>
</tr>
<tr>
<td>WNCA (%)</td>
<td>&#x2212;0.045</td>
<td>0.964</td>
<td>0.030</td>
<td>0.833</td>
</tr>
<tr>
<td><inline-formula id="ieqn-93"><mml:math id="mml-ieqn-93"><mml:mi>&#x03C1;</mml:mi></mml:math></inline-formula>NCA (kg/m<sup>3</sup>)</td>
<td>0.030</td>
<td>0.976</td>
<td>0.030</td>
<td>0.833</td>
</tr>
<tr>
<td><inline-formula id="ieqn-95"><mml:math id="mml-ieqn-95"><mml:mi>&#x03C1;</mml:mi></mml:math></inline-formula>RCA (kg/m<sup>3</sup>)</td>
<td>&#x2212;0.044</td>
<td>0.965</td>
<td>0.016</td>
<td>0.999</td>
</tr>
<tr>
<td>Compressive Strength (MPa)</td>
<td>&#x2212;0.064</td>
<td>0.949</td>
<td>0.011</td>
<td>1.000</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s2_4">
<label>2.4</label>
<title>Model Evaluation</title>
<p>Once the prediction model is developed, its performance must be assessed. The model evaluation indices used in this paper are MSE, RMSE, R, <inline-formula id="ieqn-97"><mml:math id="mml-ieqn-97"><mml:msup><mml:mi>R</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:math></inline-formula>, MAPE and MAE. MSE calculates the variation of anticipated and observed responses while <inline-formula id="ieqn-98"><mml:math id="mml-ieqn-98"><mml:msup><mml:mi>R</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:math></inline-formula> evaluates the feasibility, viability, and linearity of the model. MAPE measures the average absolute difference in percentage between forecasted values and actual values whereas MAE gauges the average absolute difference between anticipated and actual values.
<disp-formula id="eqn-29"><label>(29)</label><mml:math id="mml-eqn-29" display="block"><mml:mi>M</mml:mi><mml:mi>S</mml:mi><mml:msub><mml:mi>E</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:mrow><mml:mo>(</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:mrow><mml:mo>(</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>In <xref ref-type="disp-formula" rid="eqn-29">(29)</xref>, MSE is calculated where <inline-formula id="ieqn-99"><mml:math id="mml-ieqn-99"><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> stands for the predicted values, <inline-formula id="ieqn-100"><mml:math id="mml-ieqn-100"><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> stands for the actual values, and <inline-formula id="ieqn-101"><mml:math id="mml-ieqn-101"><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> stands for the residuals.
<disp-formula id="eqn-30"><label>(30)</label><mml:math id="mml-eqn-30" display="block"><mml:msub><mml:mrow><mml:mtext>RMSE</mml:mtext></mml:mrow><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msqrt><mml:mfrac><mml:mn>1</mml:mn><mml:mi>M</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mn>2</mml:mn></mml:msup></mml:msqrt></mml:math></disp-formula>
<disp-formula id="eqn-31"><label>(31)</label><mml:math id="mml-eqn-31" display="block"><mml:mi>R</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mover><mml:mi>X</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mover><mml:mi>Y</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:msqrt><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mover><mml:mi>X</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mn>2</mml:mn></mml:msup></mml:msqrt><mml:msqrt><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mover><mml:mi>Y</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mn>2</mml:mn></mml:msup></mml:msqrt></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p>In <xref ref-type="disp-formula" rid="eqn-31">(31)</xref>, <inline-formula id="ieqn-102"><mml:math id="mml-ieqn-102"><mml:msub><mml:mi>X</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-103"><mml:math id="mml-ieqn-103"><mml:msub><mml:mi>Y</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:math></inline-formula> represent paired observations, where <inline-formula id="ieqn-104"><mml:math id="mml-ieqn-104"><mml:msub><mml:mi>X</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:math></inline-formula> is the explanatory variable and <inline-formula id="ieqn-105"><mml:math id="mml-ieqn-105"><mml:msub><mml:mi>Y</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:math></inline-formula> is the response variable. The mean of the response variable <italic>Y</italic> is denoted by <inline-formula id="ieqn-106"><mml:math id="mml-ieqn-106"><mml:mrow><mml:mover><mml:mi>Y</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula> and average of explanatory variable <italic>X</italic> is <inline-formula id="ieqn-107"><mml:math id="mml-ieqn-107"><mml:mrow><mml:mover><mml:mi>X</mml:mi><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula>.</p>
<p><disp-formula id="eqn-32"><label>(32)</label><mml:math id="mml-eqn-32" display="block"><mml:msup><mml:mi>R</mml:mi><mml:mn>2</mml:mn></mml:msup><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mrow><mml:munder><mml:mo>&#x2211;</mml:mo><mml:mi>i</mml:mi></mml:munder><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mrow><mml:mover><mml:mi>y</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mrow><mml:munder><mml:mo>&#x2211;</mml:mo><mml:mi>i</mml:mi></mml:munder><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mover><mml:msub><mml:mi>Y</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo stretchy="false">&#x00AF;</mml:mo></mml:mover></mml:mrow><mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:math></disp-formula>
<disp-formula id="eqn-33"><label>(33)</label><mml:math id="mml-eqn-33" display="block"><mml:mrow><mml:mtext>MAE</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>M</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow></mml:math></disp-formula>
<disp-formula id="eqn-34"><label>(34)</label><mml:math id="mml-eqn-34" display="block"><mml:mrow><mml:mtext>MAPE</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>M</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mo>|</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>.</mml:mo><mml:mi>S</mml:mi></mml:mrow></mml:msub></mml:mfrac><mml:mo>|</mml:mo></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mn>100</mml:mn><mml:mi mathvariant="normal">&#x0025;</mml:mi></mml:math></disp-formula></p>
</sec>
<sec id="s2_5">
<label>2.5</label>
<title>Sensitivity Analysis</title>
<p>The sensitivity analysis of a AI based regression model is a strategy which is adopted to discover if the predicted value is altered by variations in the assumptions of the independents [<xref ref-type="bibr" rid="ref-54">54</xref>].</p>
<sec id="s2_5_1">
<label>2.5.1</label>
<title>Perturbation Sensitivity Analysis</title>
<p>Reference [<xref ref-type="bibr" rid="ref-55">55</xref>] scrutinize different perturbation based sensitivity analysis approaches on modern transformer model to gauge their performances. This method initially uses a generated model to make predictions, then introduce random noise to perturb it while preserving the integrity of all other aspects and finally inspects the predictions using the perturbed input features. Ultimately the difference is calculated using
<disp-formula id="eqn-35"><label>(35)</label><mml:math id="mml-eqn-35" display="block"><mml:mi>S</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>v</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:msub><mml:mi>y</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mtext>targetOutput</mml:mtext></mml:mrow><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mrow><mml:mtext>targetOutput</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mtext>perturbed</mml:mtext></mml:mrow></mml:mrow><mml:mi>i</mml:mi></mml:msubsup><mml:mo>|</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>In <xref ref-type="disp-formula" rid="eqn-35">(35)</xref>, <inline-formula id="ieqn-108"><mml:math id="mml-ieqn-108"><mml:mi>S</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>v</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:msub><mml:mi>y</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:math></inline-formula> is the sensitivity of the output of the model to the perturbation of the <inline-formula id="ieqn-109"><mml:math id="mml-ieqn-109"><mml:mi>j</mml:mi></mml:math></inline-formula>th input feature, <inline-formula id="ieqn-110"><mml:math id="mml-ieqn-110"><mml:mi>n</mml:mi></mml:math></inline-formula> is the number of samples in the dataset, while <inline-formula id="ieqn-111"><mml:math id="mml-ieqn-111"><mml:msub><mml:mtext>targetOutput</mml:mtext><mml:mi>i</mml:mi></mml:msub></mml:math></inline-formula> is the original predicted output for the <inline-formula id="ieqn-112"><mml:math id="mml-ieqn-112"><mml:mi>i</mml:mi></mml:math></inline-formula>th sample using the original input features and <inline-formula id="ieqn-113"><mml:math id="mml-ieqn-113"><mml:msub><mml:mtext>targetOutput</mml:mtext><mml:mrow><mml:mi>p</mml:mi><mml:mi>e</mml:mi><mml:mi>r</mml:mi><mml:mi>t</mml:mi><mml:mi>u</mml:mi><mml:mi>r</mml:mi><mml:mi>b</mml:mi><mml:mi>e</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the predicted output for the <inline-formula id="ieqn-114"><mml:math id="mml-ieqn-114"><mml:mi>i</mml:mi></mml:math></inline-formula>th sample using the perturbed input feature <inline-formula id="ieqn-115"><mml:math id="mml-ieqn-115"><mml:mi>j</mml:mi></mml:math></inline-formula>.</p>
</sec>
<sec id="s2_5_2">
<label>2.5.2</label>
<title>Weight-Partitioning Sensitivity Analysis</title>
<p>In context of ANNs, the weight matrix (W) is generated from the developed model and has dimensions of <inline-formula id="ieqn-116"><mml:math id="mml-ieqn-116"><mml:mi>i</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>j</mml:mi></mml:math></inline-formula> where <italic>i</italic> is the number of neurons in the first layer and <italic>j</italic> is the number of input features. Larger absolute weights associated with neurons indicates greater influence on output. From the absolute weights, total contribution for each input feature is calculated using <xref ref-type="disp-formula" rid="eqn-36">(36)</xref> where for each neuron m and input feature n;
<disp-formula id="eqn-36"><label>(36)</label><mml:math id="mml-eqn-36" display="block"><mml:msub><mml:mi>C</mml:mi><mml:mi>n</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>m</mml:mi></mml:munderover><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow></mml:math></disp-formula></p>
</sec>
</sec>
<sec id="s2_6">
<label>2.6</label>
<title>Partial Dependence Plot</title>
<p>Partial dependence plots (PDPs) display the slight impact of one or a small number of input features on an anticipated result of ML model. The key concept is that, while maintaining other characteristics constant, PDPs offer a transparent, graphical representation of how small modifications to the inputs impact the prediction [<xref ref-type="bibr" rid="ref-56">56</xref>]. This helps illustrate if a feature has a linear, monotonic, or complicated relationship with the target. However, PDPs presume feature independence, which can result in false interpretations if features are correlated [<xref ref-type="bibr" rid="ref-57">57</xref>]. Mathematically, for a feature <inline-formula id="ieqn-117"><mml:math id="mml-ieqn-117"><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:math></inline-formula>:
<disp-formula id="eqn-37"><label>(37)</label><mml:math id="mml-eqn-37" display="block"><mml:mrow><mml:msub><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>pd</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mi>i</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>Here, <inline-formula id="ieqn-118"><mml:math id="mml-ieqn-118"><mml:msubsup><mml:mi>x</mml:mi><mml:mi>i</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> represents all features except <inline-formula id="ieqn-119"><mml:math id="mml-ieqn-119"><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:math></inline-formula>, and <inline-formula id="ieqn-120"><mml:math id="mml-ieqn-120"><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula> is the predictive model.</p>
</sec>
<sec id="s2_7">
<label>2.7</label>
<title>Accumulated Local Effects</title>
<p>By calculating the local impact of a feature on the prediction of ML model, accumulated local effects (ALE) plots are intended to help interpret intricate machine learning models [<xref ref-type="bibr" rid="ref-58">58</xref>]. ALE avoids extrapolating into irrational areas of the feature space and takes feature correlations into consideration, in contrast to PDPs. Because of this, ALE works particularly well with correlated and high-dimensional datasets. The local effect of feature <inline-formula id="ieqn-121"><mml:math id="mml-ieqn-121"><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:math></inline-formula> at a point <inline-formula id="ieqn-122"><mml:math id="mml-ieqn-122"><mml:mi>x</mml:mi></mml:math></inline-formula> is the partial derivative of the model with respect to that feature:
<disp-formula id="eqn-38"><label>(38)</label><mml:math id="mml-eqn-38" display="block"><mml:mrow><mml:msub><mml:mi>&#x03D5;</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:mfrac></mml:mrow></mml:math></disp-formula></p>
<p>This measures how much the prediction changes locally if we change only <inline-formula id="ieqn-123"><mml:math id="mml-ieqn-123"><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:math></inline-formula>. Since derivatives are noisy and localized, ALE integrates them across the <inline-formula id="ieqn-124"><mml:math id="mml-ieqn-124"><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:math></inline-formula> range. Suppose the domain of <inline-formula id="ieqn-125"><mml:math id="mml-ieqn-125"><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:math></inline-formula> is split into intervals:
<disp-formula id="ueqn-39"><mml:math id="mml-ueqn-39" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:msub><mml:mi>z</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>&#x003C;</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x003C;</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>&#x003C;</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mi>K</mml:mi></mml:msub></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>with <italic>K</italic> bins. For a bin <inline-formula id="ieqn-126"><mml:math id="mml-ieqn-126"><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:mi>K</mml:mi></mml:math></inline-formula>, the ALE effect increment is:
<disp-formula id="eqn-39"><label>(39)</label><mml:math id="mml-eqn-39" display="block"><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>k</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mrow><mml:mi mathvariant="double-struck">E</mml:mi></mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:msubsup><mml:mo>&#x222B;</mml:mo><mml:mrow><mml:msub><mml:mi>z</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>z</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow></mml:msubsup><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:mfrac><mml:mspace width="thinmathspace" /><mml:mi>d</mml:mi><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mtext>&#x00A0;</mml:mtext><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo maxsize="2.470em" minsize="2.470em">|</mml:mo></mml:mrow></mml:mstyle><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo>&#x2208;</mml:mo><mml:mo stretchy="false">[</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">]</mml:mo><mml:mo>]</mml:mo></mml:mrow></mml:mrow></mml:math></disp-formula></p>
<p>Here, <inline-formula id="ieqn-127"><mml:math id="mml-ieqn-127"><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> are all features except <inline-formula id="ieqn-128"><mml:math id="mml-ieqn-128"><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:math></inline-formula>. The conditional expectation ensures we only integrate where data actually exists. The ALE function for feature <inline-formula id="ieqn-129"><mml:math id="mml-ieqn-129"><mml:mi>j</mml:mi></mml:math></inline-formula> at point <inline-formula id="ieqn-130"><mml:math id="mml-ieqn-130"><mml:msub><mml:mi>z</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:math></inline-formula> is the accumulated sum of increments up to that bin:
<disp-formula id="eqn-40"><label>(40)</label><mml:math id="mml-eqn-40" display="block"><mml:mrow><mml:msubsup><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>ALE</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mspace width="thinmathspace" /><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>To make ALE functions comparable and remove arbitrary offsets, they are centered:
<disp-formula id="eqn-41"><label>(41)</label><mml:math id="mml-eqn-41" display="block"><mml:mrow><mml:msubsup><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x007E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>ALE</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mspace width="thinmathspace" /><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>ALE</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mspace width="thinmathspace" /><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>K</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover><mml:msubsup><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>ALE</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mspace width="thinmathspace" /><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>This gives a function showing how predictions accumulate as <inline-formula id="ieqn-131"><mml:math id="mml-ieqn-131"><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:math></inline-formula> increases. This ensures:
<disp-formula id="eqn-42"><label>(42)</label><mml:math id="mml-eqn-42" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:mrow><mml:mrow><mml:mi mathvariant="double-struck">E</mml:mi></mml:mrow><mml:mspace width="negativethinmathspace" /><mml:mrow><mml:mo>[</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x007E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>ALE</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mspace width="thinmathspace" /><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>]</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="eqn-43"><label>(43)</label><mml:math id="mml-eqn-43" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:mrow><mml:msubsup><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x007E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>ALE</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mspace width="thinmathspace" /><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mi mathvariant="double-struck">E</mml:mi></mml:mrow><mml:mspace width="negativethinmathspace" /><mml:mrow><mml:mo>[</mml:mo><mml:msubsup><mml:mo>&#x222B;</mml:mo><mml:mrow><mml:msub><mml:mi>z</mml:mi><mml:mrow><mml:mi>l</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>z</mml:mi><mml:mi>l</mml:mi></mml:msub></mml:mrow></mml:msubsup><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mo>&#x2216;</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:mfrac><mml:mspace width="thinmathspace" /><mml:mi>d</mml:mi><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mtext>&#x00A0;</mml:mtext><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo maxsize="2.470em" minsize="2.470em">|</mml:mo></mml:mrow></mml:mstyle><mml:mtext>&#x00A0;</mml:mtext><mml:msub><mml:mi>x</mml:mi><mml:mi>j</mml:mi></mml:msub><mml:mo>&#x2208;</mml:mo><mml:mo stretchy="false">[</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mrow><mml:mi>l</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mi>l</mml:mi></mml:msub><mml:mo stretchy="false">]</mml:mo><mml:mo>]</mml:mo></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>K</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover><mml:msubsup><mml:mrow><mml:mover><mml:mi>f</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>ALE</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mspace width="thinmathspace" /><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>z</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Results</title>
<p>The first and the most critical step in developing an ANN model that is best suited to the given challenge is to select a neural network design which may include number of layers, number of neurons, architecture, training function, activation function, learning rate, number of iterations/epochs, etc. In the present study, five different optimizers/training functions (BR, LM, CGB, FP-CG, and PR-CG) have been used for model training and are compared based on MAE, MAPE, R, R<sup>2</sup>, MSE, and RMSE to assess the model performance. A total of 125 FFNN models have been developed based on different numbers of hidden layers, hidden layer neurons, and training functions with grid search. When grid search was conducted initially, best training function came out to be FP-CG with two hidden layers having 30 and 10 neurons in first and second layer consecutively as shown in <xref ref-type="table" rid="table-3">Table 3</xref>. The accuracy of the model developed in terms of coefficient of determination R<sup>2</sup> &#x003D; 0.7627 which is not acceptable by literature.</p>
<table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Performance metrics for different training functions and layer sizes</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th rowspan="2">Layer size</th>
<th rowspan="2">Training function</th>
<th rowspan="2">Number of neurons</th>
<th colspan="6">Performance metrics</th>
</tr>
<tr>
<th>RMSE</th>
<th>MSE</th>
<th>MAE</th>
<th>MAPE</th>
<th>R<sup>2</sup></th>
<th>R</th>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="5">3</td>
<td>BR</td>
<td rowspan="5">30, 20, 10</td>
<td>7.8089</td>
<td>60.9789</td>
<td>6.0075</td>
<td>15.1321%</td>
<td>0.6092</td>
<td>0.7810</td>
</tr>
<tr>
<td>LM</td>
<td>8.3340</td>
<td>69.4554</td>
<td>6.0474</td>
<td>14.6841%</td>
<td>0.5549</td>
<td>0.7571</td>
</tr>
<tr>
<td>CGB</td>
<td>7.8422</td>
<td>61.5005</td>
<td>6.0301</td>
<td>15.6791%</td>
<td>0.6059</td>
<td>0.7796</td>
</tr>
<tr>
<td>PR-CG</td>
<td>7.2939</td>
<td>53.2009</td>
<td>5.3444</td>
<td>13.5138%</td>
<td>0.6591</td>
<td>0.8123</td>
</tr>
<tr>
<td>FB-CG</td>
<td>6.4300</td>
<td>41.3447</td>
<td>3.6261</td>
<td>9.1366%</td>
<td>0.7351</td>
<td>0.8662</td>
</tr>
<tr>
<td rowspan="5">2</td>
<td>BR</td>
<td rowspan="5">30, 15</td>
<td>7.4556</td>
<td>55.5863</td>
<td>5.3942</td>
<td>13.6858%</td>
<td>0.6438</td>
<td>0.8031</td>
</tr>
<tr>
<td>LM</td>
<td>8.9057</td>
<td>79.3110</td>
<td>6.9398</td>
<td>17.7142%</td>
<td>0.4918</td>
<td>0.7034</td>
</tr>
<tr>
<td>CGB</td>
<td>6.7345</td>
<td>45.3537</td>
<td>4.3993</td>
<td>10.7081%</td>
<td>0.7094</td>
<td>0.8482</td>
</tr>
<tr>
<td>PR-CG</td>
<td>7.0250</td>
<td>49.3505</td>
<td>5.3479</td>
<td>13.4818%</td>
<td>0.6838</td>
<td>0.8296</td>
</tr>
<tr>
<td>FP-CG</td>
<td>7.6852</td>
<td>59.620</td>
<td>5.6290</td>
<td>14.16%</td>
<td>0.6215</td>
<td>0.7921</td>
</tr>
<tr>
<td rowspan="5"><bold>2</bold></td>
<td>BR</td>
<td rowspan="5">30, 10</td>
<td>6.7974</td>
<td>46.2050</td>
<td>3.5561</td>
<td>9.1256%</td>
<td>0.7039</td>
<td>0.8535</td>
</tr>
<tr>
<td>LM</td>
<td>9.3574</td>
<td>87.5611</td>
<td>6.9727</td>
<td>18.2124%</td>
<td>0.4389</td>
<td>0.6751</td>
</tr>
<tr>
<td>CGB</td>
<td>8.0653</td>
<td>65.0484</td>
<td>6.2292</td>
<td>15.8067%</td>
<td>0.5832</td>
<td>0.7646</td>
</tr>
<tr>
<td>PR-CG</td>
<td>6.2486</td>
<td>39.0449</td>
<td>4.3683</td>
<td>10.9004%</td>
<td>0.7498</td>
<td>0.8670</td>
</tr>
<tr>
<td>FP-CG</td>
<td>6.0859</td>
<td>37.0388</td>
<td>3.9782</td>
<td>10.1297%</td>
<td>0.7627</td>
<td>0.8760</td>
</tr>
<tr>
<td rowspan="5">2</td>
<td>BR</td>
<td rowspan="5">30, 5</td>
<td>8.4966</td>
<td>72.1921</td>
<td>6.4904</td>
<td>16.7663%</td>
<td>0.5374</td>
<td>0.7336</td>
</tr>
<tr>
<td>LM</td>
<td>8.6017</td>
<td>73.9899</td>
<td>6.4429</td>
<td>15.8589%</td>
<td>0.5259</td>
<td>0.7277</td>
</tr>
<tr>
<td>CGB</td>
<td>7.5806</td>
<td>57.4654</td>
<td>5.7978</td>
<td>14.9993%</td>
<td>0.6318</td>
<td>0.7950</td>
</tr>
<tr>
<td>PR-CG</td>
<td>9.3544</td>
<td>87.5053</td>
<td>7.4747</td>
<td>18.9419%</td>
<td>0.4393</td>
<td>0.6636</td>
</tr>
<tr>
<td>FP-CG</td>
<td>6.2532</td>
<td>39.1022</td>
<td>4.3911</td>
<td>11.1559%</td>
<td>0.7494</td>
<td>0.8675</td>
</tr>
<tr>
<td rowspan="5">2</td>
<td>BR</td>
<td rowspan="5">20, 10</td>
<td>8.0297</td>
<td>64.4755</td>
<td>6.1618</td>
<td>15.4525%</td>
<td>0.5868</td>
<td>0.7662</td>
</tr>
<tr>
<td>LM</td>
<td>7.5523</td>
<td>57.0379</td>
<td>5.6103</td>
<td>14.1980%</td>
<td>0.6345</td>
<td>0.7971</td>
</tr>
<tr>
<td>CGB</td>
<td>9.3268</td>
<td>86.9895</td>
<td>7.1941</td>
<td>18.0335%</td>
<td>0.4426</td>
<td>0.6665</td>
</tr>
<tr>
<td>PR-CG</td>
<td>7.1415</td>
<td>51.0004</td>
<td>5.2415</td>
<td>13.3471%</td>
<td>0.6732</td>
<td>0.8208</td>
</tr>
<tr>
<td>FP-CG</td>
<td>6.6890</td>
<td>44.7422</td>
<td>5.0034</td>
<td>12.7922%</td>
<td>0.7133</td>
<td>0.8449</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="table" rid="table-4">Table 4</xref> shows the results of the RF and XGBoost model trained using raw data collected from literature.</p>
<table-wrap id="table-4">
<label>Table 4</label>
<caption>
<title>Performance metrics comparison between RF and XGBoost</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th align="center" rowspan="2">Metric</th>
<th colspan="3">Random forest</th>
<th colspan="3">XGBoost</th>
</tr>
<tr>
<th>Training</th>
<th>Validation</th>
<th>Testing</th>
<th>Training</th>
<th>Validation</th>
<th>Testing</th>
</tr>
</thead>
<tbody>
<tr>
<td>MAE</td>
<td>2.0235</td>
<td>5.6086</td>
<td>4.3029</td>
<td>0.5960</td>
<td>4.7785</td>
<td>4.0405</td>
</tr>
<tr>
<td>MSE</td>
<td>7.6009</td>
<td>55.2304</td>
<td>38.9758</td>
<td>0.8068</td>
<td>46.4025</td>
<td>37.8390</td>
</tr>
<tr>
<td>RMSE</td>
<td>2.7570</td>
<td>7.4317</td>
<td>6.2431</td>
<td>0.8982</td>
<td>6.8119</td>
<td>6.1513</td>
</tr>
<tr>
<td><inline-formula id="ieqn-135"><mml:math id="mml-ieqn-135"><mml:msup><mml:mi>R</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:math></inline-formula> Score</td>
<td>0.9515</td>
<td>0.6849</td>
<td>0.7000</td>
<td>0.9949</td>
<td>0.7352</td>
<td>0.7088</td>
</tr>
<tr>
<td><italic>R</italic> (Correlation)</td>
<td>0.9821</td>
<td>0.8350</td>
<td>0.8435</td>
<td>0.9976</td>
<td>0.8619</td>
<td>0.8547</td>
</tr>
<tr>
<td>MAPE (%)</td>
<td>5.12</td>
<td>15.21</td>
<td>11.81</td>
<td>1.38</td>
<td>13.24</td>
<td>10.55</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="fig" rid="fig-6">Fig. 6</xref> shows the performance curve of ANN with FP-CG. The plot shows the MSE for training, validation, and testing datasets over 1000 epochs. The performance plot shows best performance at 411 epoch with least MSE, but as training is called for 1000 epochs, training continues resulting in decrease in accuracy. To avoid overfitting early stopping was applied, which did not show any significant improvement. With early stopping LM optimizer, with two hidden layers having 30 and 15 neurons respectively performed well. <xref ref-type="fig" rid="fig-7">Fig. 7</xref> reveals the performance of FFNN trained on literature-based data, comparatively lower predictive accuracy, as evidenced by greater dispersion from the ideal prediction line.</p>
<fig id="fig-6">
<label>Figure 6</label>
<caption>
<title>Performance of neural network training with FP-CG</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-6.tif"/>
</fig><fig id="fig-7">
<label>Figure 7</label>
<caption>
<title>Neural net training with FP-CG</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-7.tif"/>
</fig>
<p>It can be seen from <xref ref-type="table" rid="table-5">Table 5</xref> that the optimal FFNN structure was 30-20-10, with a R<sup>2</sup> of 0.9669; the MAE was 0.873, with the hyperbolic tangent sigmoid activation function for hidden layers and default activation function for the neurons in the output layer is the purelin (linear transfer function). It also demonstrates that performance remains consistent across the three datasets, clearly indicating that the FFNN utilizing the LM optimizer for training surpasses all other models.</p>
<table-wrap id="table-5">
<label>Table 5</label>
<caption>
<title>Performance metrics for different training functions and layer sizes for augmented data</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th align="center" rowspan="2">Layer size</th>
<th align="center" rowspan="2">Training function</th>
<th align="center" rowspan="2">Number of neurons</th>
<th colspan="6">Performance metrics</th>
</tr>
<tr>
<th>RMSE</th>
<th>MSE</th>
<th>MAE</th>
<th>MAPE</th>
<th>R<sup>2</sup></th>
<th>R</th>         
</tr>
</thead>
<tbody>
<tr>
<td rowspan="5">3</td>
<td>BR</td>
<td rowspan="5">30, 20, 10</td>
<td>3.2295</td>
<td>10.4297</td>
<td>1.9403</td>
<td>4.77%</td>
<td>0.9295</td>
<td>0.9642</td>
</tr>
<tr>
<td>LM</td>
<td>2.2127</td>
<td>4.8962</td>
<td>0.873</td>
<td><bold>2.09%</bold></td>
<td>0.9669</td>
<td>0.9835</td>
</tr>
<tr>
<td>CGB</td>
<td>3.2283</td>
<td>10.4222</td>
<td>1.7835</td>
<td>4.35%</td>
<td>0.9296</td>
<td>0.9643</td>
</tr>
<tr>
<td>PR-CG</td>
<td>3.4915</td>
<td>12.1904</td>
<td>2.144</td>
<td>5.36%</td>
<td>0.9176</td>
<td>0.9582</td>
</tr>
<tr>
<td>FP-CG</td>
<td>4.2248</td>
<td>17.849</td>
<td>3.0429</td>
<td>7.64%</td>
<td>0.8794</td>
<td>0.9379</td>
</tr>
<tr>
<td rowspan="5">2</td>
<td>BR</td>
<td rowspan="5">30, 15</td>
<td>3.347</td>
<td>11.2025</td>
<td>1.8733</td>
<td>4.64%</td>
<td>0.9243</td>
<td>0.9618</td>
</tr>
<tr>
<td>LM</td>
<td>10.384</td>
<td>107.8276</td>
<td>7.9436</td>
<td>20.98%</td>
<td>0.2715</td>
<td>0.5214</td>
</tr>
<tr>
<td>CGB</td>
<td>3.3772</td>
<td>11.4056</td>
<td>2.0563</td>
<td>5.04%</td>
<td>0.9229</td>
<td>0.9608</td>
</tr>
<tr>
<td>PR-CG</td>
<td>3.7539</td>
<td>14.0921</td>
<td>2.4469</td>
<td>6.10%</td>
<td>0.9048</td>
<td>0.9512</td>
</tr>
<tr>
<td>FP-CG</td>
<td>3.7983</td>
<td>14.4272</td>
<td>2.6253</td>
<td>6.48%</td>
<td>0.9025</td>
<td>0.95</td>
</tr>
<tr>
<td rowspan="5">2</td>
<td>BR</td>
<td rowspan="5">30, 10</td>
<td>2.5082</td>
<td>6.2912</td>
<td>1.096</td>
<td>2.64%</td>
<td>0.9575</td>
<td>0.9787</td>
</tr>
<tr>
<td>LM</td>
<td>3.5221</td>
<td>12.4055</td>
<td>2.0581</td>
<td>4.98%</td>
<td>0.9162</td>
<td>0.9575</td>
</tr>
<tr>
<td>CGB</td>
<td>3.2045</td>
<td>10.2685</td>
<td>2.0863</td>
<td>5.11%</td>
<td>0.9306</td>
<td>0.9647</td>
</tr>
<tr>
<td>PR-CG</td>
<td>4.025</td>
<td>16.2009</td>
<td>2.749</td>
<td>7.01%</td>
<td>0.8906</td>
<td>0.9438</td>
</tr>
<tr>
<td>FP-CG</td>
<td>3.4618</td>
<td>11.9841</td>
<td>2.2753</td>
<td>5.67%</td>
<td>0.919</td>
<td>0.9588</td>
</tr>
<tr>
<td rowspan="5">2</td>
<td>BR</td>
<td rowspan="5">30, 5</td>
<td>2.4445</td>
<td>5.9754</td>
<td>1.2841</td>
<td>3.10%</td>
<td>0.9596</td>
<td>0.9797</td>
</tr>
<tr>
<td>LM</td>
<td>5.7521</td>
<td>33.0864</td>
<td>0.7239</td>
<td>1.67%</td>
<td>0.7765</td>
<td>0.9035</td>
</tr>
<tr>
<td>CGB</td>
<td>3.6421</td>
<td>13.2646</td>
<td>2.3452</td>
<td>5.83%</td>
<td>0.9104</td>
<td>0.9541</td>
</tr>
<tr>
<td>PR-CG</td>
<td>4.3729</td>
<td>19.1221</td>
<td>2.9464</td>
<td>7.41%</td>
<td>0.8708</td>
<td>0.9332</td>
</tr>
<tr>
<td>FP-CG</td>
<td>4.808</td>
<td>23.1169</td>
<td>3.5008</td>
<td>8.69%</td>
<td>0.8438</td>
<td>0.9187</td>
</tr>
<tr>
<td rowspan="5">2</td>
<td>BR</td>
<td rowspan="5">20, 10</td>
<td>2.5756</td>
<td>6.6338</td>
<td>1.3392</td>
<td>3.18%</td>
<td>0.9552</td>
<td>0.9774</td>
</tr>
<tr>
<td>LM</td>
<td>5.5461</td>
<td>30.7597</td>
<td>3.1886</td>
<td>7.76%</td>
<td>0.7922</td>
<td>0.8901</td>
</tr>
<tr>
<td>CGB</td>
<td>3.8519</td>
<td>14.8373</td>
<td>2.6256</td>
<td>6.56%</td>
<td>0.8998</td>
<td>0.9486</td>
</tr>
<tr>
<td>PR-CG</td>
<td>4.1165</td>
<td>16.9456</td>
<td>2.9472</td>
<td>7.37%</td>
<td>0.8855</td>
<td>0.941</td>
</tr>
<tr>
<td>FP-CG</td>
<td>4.9154</td>
<td>24.1616</td>
<td>3.5171</td>
<td>8.80%</td>
<td>0.8368</td>
<td>0.9149</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="table" rid="table-6">Table 6</xref> presents the accuracy metrics results for the RF and XGBoost models. Both models exhibited strong performance on the training dataset; however, their effectiveness diminished when evaluated on the validation and testing datasets, indicating evident signs of overfitting. <xref ref-type="fig" rid="fig-8">Fig. 8</xref> illustrates that the Levenberg-Marquardt-trained neural network achieved its best validation performance (MSE &#x003D; 27.0455) at the final epoch, indicating stable and effective learning from augmented data.</p>
<table-wrap id="table-6">
<label>Table 6</label>
<caption>
<title>Performance metrics comparison between RF and XGBoost models for augmented data</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th align="center" rowspan="2">Metric</th>
<th colspan="3">Random Forest</th>
<th colspan="3">XGBoost</th>
</tr>
<tr>
<th>Training</th>
<th>Validation</th>
<th>Testing</th>
<th>Training</th>
<th>Validation</th>
<th>Testing</th>
</tr>
</thead>
<tbody>
<tr>
<td>MAE</td>
<td>1.5103</td>
<td>3.8592</td>
<td>4.1496</td>
<td>0.5307</td>
<td>3.6360</td>
<td>3.8290</td>
</tr>
<tr>
<td>MSE</td>
<td>4.2428</td>
<td>28.1425</td>
<td>31.9355</td>
<td>0.5482</td>
<td>25.6758</td>
<td>27.4165</td>
</tr>
<tr>
<td>RMSE</td>
<td>2.0598</td>
<td>5.3050</td>
<td>5.6512</td>
<td>0.7404</td>
<td>5.0671</td>
<td>5.2361</td>
</tr>
<tr>
<td>R<sup>2</sup> Score</td>
<td>0.9606</td>
<td>0.7117</td>
<td>0.6803</td>
<td>0.9949</td>
<td>0.7370</td>
<td>0.7255</td>
</tr>
<tr>
<td>R (Correlation)</td>
<td>0.9854</td>
<td>0.8464</td>
<td>0.8267</td>
<td>0.9977</td>
<td>0.8592</td>
<td>0.8523</td>
</tr>
<tr>
<td>MAPE (%)</td>
<td>3.62</td>
<td>9.20</td>
<td>10.05</td>
<td>1.26</td>
<td>8.61</td>
<td>9.21</td>
</tr>
</tbody>
</table>
</table-wrap><fig id="fig-8">
<label>Figure 8</label>
<caption>
<title>Performance of neural network training for augmented data with LM</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-8.tif"/>
</fig>
<p><xref ref-type="fig" rid="fig-9">Fig. 9</xref> demonstrates enhanced alignment and reduced prediction error when the model is trained on augmented data, thereby confirming the efficacy of data augmentation in improving model performance. The comparative scatter plots in <xref ref-type="fig" rid="fig-10">Fig. 10</xref> reveal that all three models&#x2014;FFNN, RF, and XGBoost&#x2014;exhibit strong predictive alignment between actual and predicted values across training, validation, and testing subsets. Notably, the FFNN model demonstrates superior consistency and accuracy, as evidenced by its tighter clustering along the ideal prediction line.</p>
<fig id="fig-9">
<label>Figure 9</label>
<caption>
<title>Neural net training for augmented data with LM</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-9.tif"/>
</fig><fig id="fig-10">
<label>Figure 10</label>
<caption>
<title>Comparison of Actual vs. Predicted values for different models</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-10.tif"/>
</fig>
<p><xref ref-type="fig" rid="fig-11">Fig. 11</xref> expresses that the model trained on collected and augmented data generates predictions of CS that closely approximate the true values. <xref ref-type="fig" rid="fig-12">Fig. 12</xref> is a 3D bar graph, highlights the importance of synergizing the right neural network architecture with an appropriate training function. It depicts six assessment parameters (RMSE, MSE, MAE, MAPE, R<sup>2</sup>, and R) for multiple layers and training function layouts in a neural network for augmented data. The <italic>x</italic>-axis illustrates distinct layer configurations, the <italic>y</italic>-axis shows the values of different assessment parameters, while the <italic>z</italic>-axis narrates different training functions used in grid search for hyperparameter tuning. It is apparent that the LM training function, paired with the 30-20-10 layer configuration, generates the most accurate results. This coalition depicts the lowest MSE, representing minimal error in forecasting, and the highest R<sup>2</sup> value, expressing a powerful alignment between the predicted and actual data.</p>
<fig id="fig-11">
<label>Figure 11</label>
<caption>
<title>Comparison of model performance for collected and augmented data</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-11a.tif"/>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-11b.tif"/>
</fig><fig id="fig-12">
<label>Figure 12</label>
<caption>
<title>3D bar chart for comparison of performance of different optimizers</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-12.tif"/>
</fig>
<sec id="s3_1">
<label>3.1</label>
<title>Sensitivity Analysis</title>
<p><xref ref-type="fig" rid="fig-13">Fig. 13</xref> highlights the influence of various input variables on the predicted CS of concrete containing RAs. The results of the perturbation analysis reveal that cement has the most significant effect, whereas the weight partitioning sensitivity analysis reveals that water has the utmost influence on strength. Furthermore, based on weight-partitioning sensitivity analysis, cement is the second most influential factor. This result is consistent with the widely approved fact that cement being main component of concrete has a significant impact on its strength. CS of concrete is eventually determined by the chemical reactions and hydration processes that are directly influenced by the quantity, quality, and composition of the cement. The presence of water is vital in concrete, as it initiates and sustains the hydration process, allowing cement to react with other components and form a strong bond [<xref ref-type="bibr" rid="ref-55">55</xref>]. But with too much amount of water, pores get saturated leaving gaps, which weakens the material. Other parameters that can influence, presence of admixtures, the temperature at the time of hydration, and aggregate and cement paste interface [<xref ref-type="bibr" rid="ref-59">59</xref>]. From 13, other than water and cement, superplasticizer and water absorption are the other important factors.</p>
<fig id="fig-13">
<label>Figure 13</label>
<caption>
<title>Sensitivity analysis of FFNN trained model using LM optimizer</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-13.tif"/>
</fig>
<p><xref ref-type="fig" rid="fig-14">Figs. 14</xref> and <xref ref-type="fig" rid="fig-15">15</xref> show consistent results that water to cement ratio is the most influential factor for both the RF and XGBoost models.</p>
<fig id="fig-14">
<label>Figure 14</label>
<caption>
<title>Sensitivity analysis of RF model</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-14.tif"/>
</fig><fig id="fig-15">
<label>Figure 15</label>
<caption>
<title>Sensitivity analysis of XGBoost model</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-15.tif"/>
</fig>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Partial Dependence Plots</title>
<p><xref ref-type="fig" rid="fig-16">Fig. 16</xref> shows the PDPs analyses of all the three models which clearly indicate that the presence of cement and superplasticizer contributes positively to the predicted strength. Conversely, increased water-to-cement ratios and the properties of RAs lead to a decrease in strength. While XGBoost is adept at identifying more pronounced nonlinear relationships, both models demonstrate trends that are consistent with domain knowledge, even though PDPs may have limitations in scenarios involving correlated features.</p>
<fig id="fig-16">
<label>Figure 16</label>
<caption>
<title>Comparison of PDPs for different models</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-16.tif"/>
</fig>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Accumulated Local Effects</title>
<p><xref ref-type="fig" rid="fig-17">Fig. 17</xref> illustrates the aggregated localized impacts of the input features on the outputs of the three models employed in this investigation. The RF and XGBoost models exhibit identical behavior. The enhancement of compressive strength is primarily influenced by the presence of cement, sand and SP. Conversely, higher water-to-cement (w/c) ratios result in reduction in matrix density, while excessive RCA content is linked to lower quality, increased porosity, and weaker interfacial zones. In contrast, NCA content has a positive impact, and other factors such as maximum aggregate size (Dmax) and absorption capacities exhibit relatively minor or negligible effects within the examined range. While the ALE analysis of FFNN indicates that the strength is most significantly affected by increased cement content and superplasticizer dosage, whereas higher water-to-cement ratios, porosity of RAs, and water absorption considerably diminish the predicted strength. Nonlinear effects associated with aggregate size and water content imply the existence of optimal ranges instead of straightforward trends. Characteristics such as NCAs and sand exhibit limited or context-sensitive effects. In summary, the model effectively encapsulates both anticipated physical relationships and intricate interactions, thereby enhancing its interpretability and alignment with the domain.</p>
<fig id="fig-17">
<label>Figure 17</label>
<caption>
<title>Comparison of ALE plots for different models</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMES_72200-fig-17.tif"/>
</fig>
<p>The addition of RAs has a nonlinear influence on CS, as shown by the PDP 16 and ALE 17 graphs. In particular, the model shows that when RAs are added up to a moderate replacement levels (RL%), CS increases, after which strength tends to decrease. This trend aligns with the experimental results presented by Khan et al. [<xref ref-type="bibr" rid="ref-60">60</xref>], who found that structural concrete applications can retain acceptable strength with 30%&#x2013;40% RAs replacement. In a similar domain, Ali et al. [<xref ref-type="bibr" rid="ref-61">61</xref>] illustrated that high-strength concrete preserves its mechanical properties with up to 25%&#x2013;30% RAs when enhanced with mineral admixtures. Furthermore, previous research conducted by Kumar and Rao [<xref ref-type="bibr" rid="ref-62">62</xref>] corroborated that replacing up to 20%&#x2013;30% with RAs results in CS that is comparable to traditional mixes. Significantly, substituting NAs with RAs at a level of 30% has demonstrated a reduction in <inline-formula id="ieqn-141"><mml:math id="mml-ieqn-141"><mml:mrow><mml:mi mathvariant="normal">C</mml:mi><mml:msub><mml:mi mathvariant="normal">O</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> emissions by roughly 12%&#x2013;20%, contingent upon regional sourcing and transportation considerations [<xref ref-type="bibr" rid="ref-61">61</xref>]. These findings indicate that the model successfully reflects behavior consistent with the domain, allowing engineers and material designers to estimate strength based on the properties of aggregates without the need for extensive laboratory evaluations. The capacity to predict performance while incorporating recycled materials enhances the significance of model for sustainable mix design, providing both mechanical dependability and measurable environmental advantages.</p>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Conclusion</title>
<p>In this study, synthetic data generation and FFNNs were utilized to forecast the CS of RAC. Data augmentation increased the dataset points, ensuring that the synthetic data maintained similar statistical characteristics to the original dataset collected from literature. Hyperparameters such as the number of layers, neurons, and five different optimizers were optimized using a grid search technique, which is time-saving and outperforms traditional trial-and-error techniques. Among these, the Levenberg-Marquardt optimizer emerged as the best, yielding faster convergence and high forecasting accuracy. The optimized FFNN model with a three-hidden-layer structure (30-20-10 neurons) achieved strong performance with R &#x003D; 0.9835, <inline-formula id="ieqn-142"><mml:math id="mml-ieqn-142"><mml:msup><mml:mi>R</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:math></inline-formula> &#x003D; 0.9669, RMSE &#x003D; 2.2127 MPa, and MAE &#x003D; 0.873 MPa. Sensitivity analysis further identified cement and water content as the most influential parameters in predicting RAC strength. The PDP and ALE plots revealed nuanced relationships between input variables and model predictions, identifying the RCA component as a key driver of model behavior. Notably, the analysis showed that increasing the RCA content enhances concrete strength up to a certain threshold, beyond which further addition leads to a decline in strength&#x2013;highlighting a non-linear effect that underscores the importance of optimal RCA dosage.</p>
<p>However, this study is limited by the simplified characterization of RAs, which excludes pretreatment effects, and does not account for the influence of supplementary cementitious materials (SCMs) or fine aggregate replacement factors that may significantly affect RAC performance in practical applications.</p>
<p>Subsequent research ought to broaden this methodology by integrating additional variables, including curing age, types of additives, and SCMs, which are essential for the performance of RAC. Given the environmental ramifications of cement, transfer learning with the same developed FFNN model can be implemented on SCM-based datasets, minimizing the need for collecting substantial new dataset with improved predictive accuracy and promoting sustainability through reduced <inline-formula id="ieqn-143"><mml:math id="mml-ieqn-143"><mml:mi>C</mml:mi><mml:msub><mml:mi>O</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula> emissions. This study demonstrates that synthetic data augmentation, when paired with optimized neural networks, offers an effective and flexible framework for predicting RAC compressive strength.</p>
</sec>
</body>
<back>
<ack>
<p>This work was supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-DDRSP2503).</p>
</ack>
<sec><title>Funding Statement</title>
<p>This work was supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-DDRSP2503).</p></sec>
<sec><title>Author Contributions</title>
<p>Sandeerah Choudhary: Conceptualization, methodology, original draft, simulation; Tallha Akram: Methodology, simulation, review, data curability; Qaisar Abbas: Review &#x0026; editing, funding; Irshad Qureshi: Review &#x0026; editing, supervision; Mutlaq B. Aldajani: Review &#x0026; editing, funding; Hammad Salahuddin: Simulation, review &#x0026; editing. All authors reviewed the results and approved the final version of the manuscript.</p></sec>
<sec sec-type="data-availability">
<title>Availability of Data and Materials</title>
<p>The datasets generated or analyzed during the current study are available in the Google Drive repository at <ext-link ext-link-type="uri" xlink:href="https://drive.google.com/drive/folders/1sBnxUfpPGW0mwIDbXqvrcwSwQ4ihAOhT">https://drive.google.com/drive/folders/1sBnxUfpPGW0mwIDbXqvrcwSwQ4ihAOhT</ext-link>, accessed on 12 October 2025.</p>
</sec>
<sec>
<title>Ethics Approval</title>
<p>This research did not involve human participants or animals. Ethical approval was therefore not required.</p>
</sec>
<sec sec-type="COI-statement">
<title>Conflicts of Interest</title>
<p>The authors declare no conflicts of interest to report regarding the present study.</p>
</sec>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Mohamed</surname> <given-names>OA</given-names></string-name>, <string-name><surname>Zuaiter</surname> <given-names>HA</given-names></string-name>, <string-name><surname>Jawa</surname> <given-names>MM</given-names></string-name></person-group>. <article-title>Carbonation and chloride penetration resistance of sustainable structural concrete with alkali-activated and ordinary Portland cement binders: a critical review</article-title>. <source>Sustain Struct</source>. <year>2025</year>;<volume>5</volume>(<issue>2</issue>):<fpage>000075</fpage>. doi:<pub-id pub-id-type="doi">10.54113/j.sust.2025.000075</pub-id>.</mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Chowdhury</surname> <given-names>JA</given-names></string-name>, <string-name><surname>Islam</surname> <given-names>MS</given-names></string-name>, <string-name><surname>Islam</surname> <given-names>MA</given-names></string-name>, <string-name><surname>Al Bari</surname> <given-names>MA</given-names></string-name>, <string-name><surname>Debnath</surname> <given-names>AK</given-names></string-name></person-group>. <article-title>Analysis of mechanical properties of fly ash and boiler slag integrated geopolymer composites</article-title>. <source>Sustain Struct</source>. <year>2025</year>;<volume>5</volume>(<issue>2</issue>):<fpage>000073</fpage>. doi:<pub-id pub-id-type="doi">10.54113/j.sust.2025.000073</pub-id>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab>U S Geological Survey</collab></person-group>. <article-title>Natural Aggregates Statistics and Information</article-title>. <comment>[cited 2025 Oct 13]</comment>. Available from: <ext-link ext-link-type="uri" xlink:href="https://www.usgs.gov/centers/national-minerals-information-center/natural-aggregates-statistics-and-information">https://www.usgs.gov/centers/national-minerals-information-center/natural-aggregates-statistics-and-information</ext-link>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ajayi</surname> <given-names>S</given-names></string-name>, <string-name><surname>Oyedele</surname> <given-names>L</given-names></string-name>, <string-name><surname>Akinade</surname> <given-names>O</given-names></string-name>, <string-name><surname>Bilal</surname> <given-names>M</given-names></string-name>, <string-name><surname>Owolabi</surname> <given-names>H</given-names></string-name>, <string-name><surname>Alaka</surname> <given-names>H</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Reducing waste to landfill: a need for cultural change in the UK construction industry</article-title>. <source>J Build Eng (JOBE)</source>. <year>2016</year>;<volume>5</volume>(<issue>1</issue>):<fpage>185</fpage>&#x2013;<lpage>93</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.jobe.2015.12.007</pub-id>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Hu</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>R</given-names></string-name>, <string-name><surname>Su</surname> <given-names>P</given-names></string-name>, <string-name><surname>Huang</surname> <given-names>J</given-names></string-name>, <string-name><surname>Peng</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>Construction and demolition waste generation prediction and spatiotemporal analysis: a case study in Sichuan, China</article-title>. <source>Environ Sci Pollut Res</source>. <year>2023</year>;<volume>30</volume>(<issue>14</issue>):<fpage>41623</fpage>&#x2013;<lpage>43</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s11356-022-25062-6</pub-id>; <pub-id pub-id-type="pmid">36635474</pub-id></mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Naderpour</surname> <given-names>H</given-names></string-name>, <string-name><surname>Rafiean</surname> <given-names>AH</given-names></string-name>, <string-name><surname>Fakharian</surname> <given-names>P</given-names></string-name></person-group>. <article-title>Compressive strength prediction of environmentally friendly concrete using artificial neural networks</article-title>. <source>J Build Eng</source>. <year>2018</year>;<volume>16</volume>:<fpage>213</fpage>&#x2013;<lpage>9</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.jobe.2018.01.007</pub-id>.</mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Tam</surname> <given-names>VWY</given-names></string-name>, <string-name><surname>Soomro</surname> <given-names>M</given-names></string-name>, <string-name><surname>Evangelista</surname> <given-names>ACJ</given-names></string-name></person-group>. <article-title>A review of recycled aggregate in concrete applications (2000&#x2013;2017)</article-title>. <source>Construct Build Mat</source>. <year>2018</year>;<volume>172</volume>:<fpage>272</fpage>&#x2013;<lpage>92</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.conbuildmat.2018.03.240</pub-id>.</mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Abdelfatah</surname> <given-names>A</given-names></string-name>, <string-name><surname>Tabsh</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Review of research on and implementation of recycled concrete aggregate in the GCC</article-title>. <source>Adv Civil Eng</source>. <year>2011</year>;<volume>2011</volume>:<fpage>567924</fpage>. doi:<pub-id pub-id-type="doi">10.1155/2011/567924</pub-id>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Rodr&#x00ED;guez</surname> <given-names>C</given-names></string-name>, <string-name><surname>Parra</surname> <given-names>C</given-names></string-name>, <string-name><surname>Casado</surname> <given-names>G</given-names></string-name>, <string-name><surname>Mi&#x00F1;ano</surname> <given-names>I</given-names></string-name>, <string-name><surname>Albaladejo</surname> <given-names>F</given-names></string-name>, <string-name><surname>Benito</surname> <given-names>F</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>The incorporation of construction and demolition wastes as recycled mixed aggregates in non-structural concrete precast pieces</article-title>. <source>J Cleaner Product</source>. <year>2016</year>;<volume>127</volume>:<fpage>152</fpage>&#x2013;<lpage>61</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.jclepro.2016.03.137</pub-id>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Gulati</surname> <given-names>R</given-names></string-name>, <string-name><surname>Bano</surname> <given-names>S</given-names></string-name>, <string-name><surname>Bano</surname> <given-names>F</given-names></string-name>, <string-name><surname>Singh</surname> <given-names>S</given-names></string-name>, <string-name><surname>Singh</surname> <given-names>V</given-names></string-name></person-group>. <article-title>Compressive strength of concrete formulated with waste materials using neural networks</article-title>. <source>Asian J Civil Eng</source>. <year>2024</year>;<volume>25</volume>(<issue>6</issue>):<fpage>4657</fpage>&#x2013;<lpage>72</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s42107-024-01071-3</pub-id>.</mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Sathiparan</surname> <given-names>N</given-names></string-name>, <string-name><surname>Subramaniam</surname> <given-names>DN</given-names></string-name></person-group>. <article-title>Optimizing fly ash and rice husk ash as cement replacements on the mechanical characteristics of pervious concrete</article-title>. <source>Sustain Struct</source>. <year>2025</year>;<volume>5</volume>(<issue>1</issue>):<fpage>000065</fpage>. doi:<pub-id pub-id-type="doi">10.1080/10298436.2022.2075867</pub-id>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zhang</surname> <given-names>X</given-names></string-name>, <string-name><surname>Akber</surname> <given-names>MZ</given-names></string-name>, <string-name><surname>Zheng</surname> <given-names>W</given-names></string-name></person-group>. <article-title>Prediction of seven-day compressive strength of field concrete</article-title>. <source>Constr Build Mater</source>. <year>2021</year>;<volume>305</volume>(<issue>5</issue>):<fpage>124604</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.conbuildmat.2021.124604</pub-id>.</mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Liu</surname> <given-names>K</given-names></string-name>, <string-name><surname>Zou</surname> <given-names>C</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>X</given-names></string-name>, <string-name><surname>Yan</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Innovative prediction models for the frost durability of recycled aggregate concrete using soft computing methods</article-title>. <source>J Build Eng</source>. <year>2021</year>;<volume>34</volume>(<issue>1</issue>):<fpage>101822</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.jobe.2020.101822</pub-id>.</mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Bui</surname> <given-names>Q-AT</given-names></string-name>, <string-name><surname>Nguyen</surname> <given-names>DD</given-names></string-name>, <string-name><surname>Le</surname> <given-names>HV</given-names></string-name>, <string-name><surname>Pham</surname> <given-names>BT</given-names></string-name>, <string-name><surname>Prakash</surname> <given-names>I</given-names></string-name></person-group>. <article-title>Prediction of shear bond strength of asphalt concrete pavement using machine learning models and grid search optimization technique</article-title>. <source>Comput Model Eng Sci</source>. <year>2025</year>;<volume>142</volume>(<issue>1</issue>):<fpage>691</fpage>&#x2013;<lpage>712</lpage>. doi:<pub-id pub-id-type="doi">10.32604/cmes.2024.054766</pub-id>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Tariq</surname> <given-names>J</given-names></string-name>, <string-name><surname>Hu</surname> <given-names>K</given-names></string-name>, <string-name><surname>Gillani</surname> <given-names>STA</given-names></string-name>, <string-name><surname>Chang</surname> <given-names>H</given-names></string-name>, <string-name><surname>Ashraf</surname> <given-names>MW</given-names></string-name>, <string-name><surname>Khan</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Enhancing the predictive accuracy of recycled aggregate concrete&#x2019;s strength using machine learning and statistical approaches: a review</article-title>. <source>Asian J Civil Eng</source>. <year>2025</year>;<volume>26</volume>(<issue>1</issue>):<fpage>21</fpage>&#x2013;<lpage>46</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s42107-024-01192-9</pub-id>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zhang</surname> <given-names>J</given-names></string-name>, <string-name><surname>Huang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Aslani</surname> <given-names>F</given-names></string-name>, <string-name><surname>Ma</surname> <given-names>G</given-names></string-name>, <string-name><surname>Nener</surname> <given-names>B</given-names></string-name></person-group>. <article-title>A hybrid intelligent system for designing optimal proportions of recycled aggregate concrete</article-title>. <source>J Clean Prod</source>. <year>2020</year>;<volume>273</volume>(<issue>3</issue>):<fpage>122922</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.jclepro.2020.122922</pub-id>.</mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Munir</surname> <given-names>MJ</given-names></string-name>, <string-name><surname>Kazmi</surname> <given-names>SMS</given-names></string-name>, <string-name><surname>Wu</surname> <given-names>YF</given-names></string-name>, <string-name><surname>Lin</surname> <given-names>X</given-names></string-name>, <string-name><surname>Ahmad</surname> <given-names>MR</given-names></string-name></person-group>. <article-title>Development of novel design strength model for sustainable concrete columns: a new machine learning-based approach</article-title>. <source>J Clean Prod</source>. <year>2022</year>;<volume>357</volume>(<issue>8</issue>):<fpage>131988</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.jclepro.2022.131988</pub-id>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Munir</surname> <given-names>MJ</given-names></string-name>, <string-name><surname>Kazmi</surname> <given-names>SMS</given-names></string-name>, <string-name><surname>Wu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Patnaikuni</surname> <given-names>I</given-names></string-name></person-group>. <article-title>Influence of concrete strength on the stress-strain behavior of spirally confined recycled aggregate concrete</article-title>. <source>IOP Conf Ser Mat Sci Eng</source>. <year>2020</year>;<volume>829</volume>(<issue>1</issue>):<fpage>012004</fpage>. doi:<pub-id pub-id-type="doi">10.1088/1757-899x/829/1/012004</pub-id>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ly</surname> <given-names>HB</given-names></string-name>, <string-name><surname>Pham</surname> <given-names>BT</given-names></string-name>, <string-name><surname>Dao</surname> <given-names>DV</given-names></string-name>, <string-name><surname>Le</surname> <given-names>VM</given-names></string-name>, <string-name><surname>Le</surname> <given-names>LM</given-names></string-name>, <string-name><surname>Le</surname> <given-names>TT</given-names></string-name></person-group>. <article-title>Improvement of ANFIS model for prediction of compressive strength of manufactured sand concrete</article-title>. <source>Appl Sci</source>. <year>2019</year>;<volume>9</volume>(<issue>18</issue>):<fpage>3841</fpage>. doi:<pub-id pub-id-type="doi">10.3390/app9183841</pub-id>.</mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Cheng</surname> <given-names>MY</given-names></string-name>, <string-name><surname>Chou</surname> <given-names>JS</given-names></string-name>, <string-name><surname>Roy</surname> <given-names>AFV</given-names></string-name>, <string-name><surname>Wu</surname> <given-names>YW</given-names></string-name></person-group>. <article-title>High-performance concrete compressive strength prediction using time-weighted evolutionary fuzzy support vector machines inference model</article-title>. <source>Automat Construct</source>. <year>2012</year>;<volume>28</volume>(<issue>2</issue>):<fpage>106</fpage>&#x2013;<lpage>15</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.autcon.2012.07.004</pub-id>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Dantas</surname> <given-names>ATA</given-names></string-name>, <string-name><surname>Leite</surname> <given-names>MB</given-names></string-name>, <string-name><surname>de Jesus Nagahama</surname> <given-names>K</given-names></string-name></person-group>. <article-title>Prediction of compressive strength of concrete containing construction and demolition waste using artificial neural networks</article-title>. <source>Construct Build Mat</source>. <year>2013</year>;<volume>38</volume>:<fpage>717</fpage>&#x2013;<lpage>22</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.conbuildmat.2012.09.026</pub-id>.</mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Nguyen</surname> <given-names>TA</given-names></string-name>, <string-name><surname>Ly</surname> <given-names>HB</given-names></string-name>, <string-name><surname>Mai</surname> <given-names>HVT</given-names></string-name>, <string-name><surname>Tran</surname> <given-names>VQ</given-names></string-name></person-group>. <article-title>Prediction of later-age concrete compressive strength using feedforward neural network</article-title>. <source>Adv Mater Sci Eng</source>. <year>2020</year>;<volume>2020</volume>(<issue>1</issue>):<fpage>762</fpage>. doi:<pub-id pub-id-type="doi">10.1155/2020/9682740</pub-id>.</mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Murphy</surname> <given-names>KP</given-names></string-name></person-group>. <source>Machine learning: a probabilistic perspective</source>. <publisher-loc>Cambridge, MA, USA</publisher-loc>: <publisher-name>The MIT Press</publisher-name>; <year>2012</year>.</mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Aghabalaei Baghaei</surname> <given-names>K</given-names></string-name>, <string-name><surname>Hadigheh</surname> <given-names>SA</given-names></string-name></person-group>. <article-title>FRP bar-to-concrete connection durability in diverse environmental exposures: an optimal machine learning approach to predicting bond strength</article-title>. <source>Structures</source>. <year>2025</year>;<volume>76</volume>:<fpage>108988</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.istruc.2025.108988</pub-id>.</mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Nguyen</surname> <given-names>TD</given-names></string-name>, <string-name><surname>Cherif</surname> <given-names>R</given-names></string-name>, <string-name><surname>Mahieux</surname> <given-names>PY</given-names></string-name>, <string-name><surname>Lux</surname> <given-names>J</given-names></string-name>, <string-name><surname>A&#x00EF;t-Mokhtar</surname> <given-names>A</given-names></string-name>, <string-name><surname>Bastidas-Arteaga</surname> <given-names>E</given-names></string-name></person-group>. <article-title>Artificial intelligence algorithms for prediction and sensitivity analysis of mechanical properties of recycled aggregate concrete: a review</article-title>. <source>J Build Eng</source>. <year>2023</year>;<volume>66</volume>:<fpage>105929</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.jobe.2023.105929</pub-id>.</mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Xu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Zhao</surname> <given-names>X</given-names></string-name>, <string-name><surname>Yu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Xie</surname> <given-names>T</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>G</given-names></string-name>, <string-name><surname>Xue</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Parametric sensitivity analysis and modelling of mechanical properties of normal- and high-strength recycled aggregate concrete using grey theory, multiple nonlinear regression and artificial neural networks</article-title>. <source>Construct Build Mat</source>. <year>2019</year>;<volume>211</volume>:<fpage>479</fpage>&#x2013;<lpage>91</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.conbuildmat.2019.03.234</pub-id>.</mixed-citation></ref>
<ref id="ref-27"><label>[27]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Haider</surname> <given-names>SA</given-names></string-name>, <string-name><surname>Naqvi</surname> <given-names>SR</given-names></string-name>, <string-name><surname>Akram</surname> <given-names>T</given-names></string-name>, <string-name><surname>Kamran</surname> <given-names>M</given-names></string-name>, <string-name><surname>Qadri</surname> <given-names>NN</given-names></string-name></person-group>. <article-title>Modeling electrical properties for various geometries of antidots on a superconducting film</article-title>. <source>Appl Nanosci</source>. <year>2017</year>;<volume>7</volume>:<fpage>933</fpage>&#x2013;<lpage>45</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s13204-017-0633-4</pub-id>.</mixed-citation></ref>
<ref id="ref-28"><label>[28]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Naqvi</surname> <given-names>SR</given-names></string-name>, <string-name><surname>Akram</surname> <given-names>T</given-names></string-name>, <string-name><surname>Haider</surname> <given-names>SA</given-names></string-name>, <string-name><surname>Kamran</surname> <given-names>M</given-names></string-name>, <string-name><surname>Shahzad</surname> <given-names>A</given-names></string-name>, <string-name><surname>Khan</surname> <given-names>W</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Precision modeling: application of metaheuristics on current-voltage curves of superconducting films</article-title>. <source>Electronics</source>. <year>2018</year>;<volume>7</volume>(<issue>8</issue>):<fpage>138</fpage>. doi:<pub-id pub-id-type="doi">10.3390/electronics7080138</pub-id>.</mixed-citation></ref>
<ref id="ref-29"><label>[29]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Chopra</surname> <given-names>P</given-names></string-name>, <string-name><surname>Sharma</surname> <given-names>RK</given-names></string-name>, <string-name><surname>Kumar</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Artificial neural networks for the prediction of compressive strength of concrete</article-title>. <source>Int J Appl Sci Eng</source>. <year>2015</year>;<volume>13</volume>:<fpage>187</fpage>&#x2013;<lpage>204</lpage>.</mixed-citation></ref>
<ref id="ref-30"><label>[30]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Nunez</surname> <given-names>I</given-names></string-name>, <string-name><surname>Marani</surname> <given-names>A</given-names></string-name>, <string-name><surname>Flah</surname> <given-names>M</given-names></string-name>, <string-name><surname>Nehdi</surname> <given-names>ML</given-names></string-name></person-group>. <article-title>Estimating compressive strength of modern concrete mixtures using computational intelligence: a systematic review</article-title>. <source>Constr Build Mater</source>. <year>2021</year>;<volume>310</volume>(<issue>1</issue>):<fpage>125279</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.conbuildmat.2021.125279</pub-id>.</mixed-citation></ref>
<ref id="ref-31"><label>[31]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Duan</surname> <given-names>ZH</given-names></string-name>, <string-name><surname>Kou</surname> <given-names>SC</given-names></string-name>, <string-name><surname>Poon</surname> <given-names>CS</given-names></string-name></person-group>. <article-title>Prediction of compressive strength of recycled aggregate concrete using artificial neural networks</article-title>. <source>Constr Build Mater</source>. <year>2013</year>;<volume>40</volume>(<issue>1</issue>):<fpage>1200</fpage>&#x2013;<lpage>6</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.conbuildmat.2012.04.063</pub-id>.</mixed-citation></ref>
<ref id="ref-32"><label>[32]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Duan</surname> <given-names>ZH</given-names></string-name>, <string-name><surname>Kou</surname> <given-names>SC</given-names></string-name>, <string-name><surname>Poon</surname> <given-names>CS</given-names></string-name></person-group>. <article-title>Using artificial neural networks for predicting the elastic modulus of recycled aggregate concrete</article-title>. <source>Constr Build Mater</source>. <year>2013</year>;<volume>44</volume>(<issue>7</issue>):<fpage>524</fpage>&#x2013;<lpage>32</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.conbuildmat.2013.02.064</pub-id>.</mixed-citation></ref>
<ref id="ref-33"><label>[33]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Gholampour</surname> <given-names>A</given-names></string-name>, <string-name><surname>Mansouri</surname> <given-names>I</given-names></string-name>, <string-name><surname>Kisi</surname> <given-names>O</given-names></string-name>, <string-name><surname>Ozbakkaloglu</surname> <given-names>T</given-names></string-name></person-group>. <article-title>Evaluation of mechanical properties of concretes containing coarse recycled concrete aggregates using multivariate adaptive regression splines (MARS), M5 model tree (M5Tree), and least squares support vector regression (LSSVR) models</article-title>. <source>Neural Comput Appl</source>. <year>2020</year>;<volume>32</volume>(<issue>1</issue>):<fpage>295</fpage>&#x2013;<lpage>308</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s00521-018-3630-y</pub-id>.</mixed-citation></ref>
<ref id="ref-34"><label>[34]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Deng</surname> <given-names>F</given-names></string-name>, <string-name><surname>He</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>S</given-names></string-name>, <string-name><surname>Yu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Cheng</surname> <given-names>H</given-names></string-name>, <string-name><surname>Wu</surname> <given-names>X</given-names></string-name></person-group>. <article-title>Compressive strength prediction of recycled concrete based on deep learning</article-title>. <source>Constr Build Mater</source>. <year>2018</year>;<volume>175</volume>(<issue>7</issue>):<fpage>562</fpage>&#x2013;<lpage>9</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.conbuildmat.2018.04.169</pub-id>.</mixed-citation></ref>
<ref id="ref-35"><label>[35]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Hammoudi</surname> <given-names>A</given-names></string-name>, <string-name><surname>Moussaceb</surname> <given-names>K</given-names></string-name>, <string-name><surname>Belebchouche</surname> <given-names>C</given-names></string-name>, <string-name><surname>Dahmoune</surname> <given-names>F</given-names></string-name></person-group>. <article-title>Comparison of artificial neural network (ANN) and response surface methodology (RSM) prediction in compressive strength of recycled concrete aggregates</article-title>. <source>Constr Build Mater</source>. <year>2019</year>;<volume>209</volume>:<fpage>425</fpage>&#x2013;<lpage>36</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.conbuildmat.2019.03.119</pub-id>.</mixed-citation></ref>
<ref id="ref-36"><label>[36]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Khan</surname> <given-names>K</given-names></string-name>, <string-name><surname>Ahmad</surname> <given-names>W</given-names></string-name>, <string-name><surname>Amin</surname> <given-names>MN</given-names></string-name>, <string-name><surname>Aslam</surname> <given-names>F</given-names></string-name>, <string-name><surname>Ahmad</surname> <given-names>A</given-names></string-name>, <string-name><surname>Al-Faiad</surname> <given-names>MA</given-names></string-name></person-group>. <article-title>Comparison of prediction models based on machine learning for the compressive strength estimation of recycled aggregate concrete</article-title>. <source>Materials</source>. <year>2022</year>;<volume>15</volume>(<issue>10</issue>):<fpage>3430</fpage>. doi:<pub-id pub-id-type="doi">10.3390/ma15103430</pub-id>; <pub-id pub-id-type="pmid">35629456</pub-id></mixed-citation></ref>
<ref id="ref-37"><label>[37]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Li</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Yoon</surname> <given-names>J</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>R</given-names></string-name>, <string-name><surname>Rajabipour</surname> <given-names>F</given-names></string-name>, <string-name><surname>Srubar</surname> <given-names>WV</given-names>
<suffix>III</suffix></string-name>, <string-name><surname>Dabo</surname> <given-names>I</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Machine learning in concrete science: applications, challenges, and best practices</article-title>. <source>npj Comp Mater</source>. <year>2022</year>;<volume>8</volume>(<issue>1</issue>):<fpage>127</fpage>. doi:<pub-id pub-id-type="doi">10.1038/s41524-022-00810-x</pub-id>.</mixed-citation></ref>
<ref id="ref-38"><label>[38]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Nguyen</surname> <given-names>HAT</given-names></string-name>, <string-name><surname>Pham</surname> <given-names>DH</given-names></string-name>, <string-name><surname>Ahn</surname> <given-names>Y</given-names></string-name></person-group>. <article-title>Effect of data augmentation using deep learning on predictive models for geopolymer compressive strength</article-title>. <source>Appl Sci</source>. <year>2024</year>;<volume>14</volume>(<issue>9</issue>):<fpage>3601</fpage>. doi:<pub-id pub-id-type="doi">10.3390/app14093601</pub-id>.</mixed-citation></ref>
<ref id="ref-39"><label>[39]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Yasuno</surname> <given-names>T</given-names></string-name>, <string-name><surname>Nakajima</surname> <given-names>M</given-names></string-name>, <string-name><surname>Sekiguchi</surname> <given-names>T</given-names></string-name>, <string-name><surname>Noda</surname> <given-names>K</given-names></string-name>, <string-name><surname>Aoyanagi</surname> <given-names>K</given-names></string-name>, <string-name><surname>Kato</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Synthetic image augmentation for damage region segmentation using conditional GAN with structure edge</article-title>. <comment>arXiv:2005.08628. 2020</comment>.</mixed-citation></ref>
<ref id="ref-40"><label>[40]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Chen</surname> <given-names>N</given-names></string-name>, <string-name><surname>Zhao</surname> <given-names>S</given-names></string-name>, <string-name><surname>Gao</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>D</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>P</given-names></string-name>, <string-name><surname>Oeser</surname> <given-names>M</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Virtual mix design: prediction of compressive strength of concrete with industrial wastes using deep data augmentation</article-title>. <source>Constr Build Mater</source>. <year>2022</year>;<volume>323</volume>(<issue>5</issue>):<fpage>126580</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.conbuildmat.2022.126580</pub-id>.</mixed-citation></ref>
<ref id="ref-41"><label>[41]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Liu</surname> <given-names>KH</given-names></string-name>, <string-name><surname>Xie</surname> <given-names>TY</given-names></string-name>, <string-name><surname>Cai</surname> <given-names>ZK</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>GM</given-names></string-name>, <string-name><surname>Zhao</surname> <given-names>XY</given-names></string-name></person-group>. <article-title>Data-driven prediction and optimization of axial compressive strength for FRP-reinforced CFST columns using synthetic data augmentation</article-title>. <source>Eng Struct</source>. <year>2024</year>;<volume>300</volume>(<issue>5</issue>):<fpage>117225</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.engstruct.2023.117225</pub-id>.</mixed-citation></ref>
<ref id="ref-42"><label>[42]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Marani</surname> <given-names>A</given-names></string-name>, <string-name><surname>Jamali</surname> <given-names>A</given-names></string-name>, <string-name><surname>Nehdi</surname> <given-names>ML</given-names></string-name></person-group>. <article-title>Predicting ultra-high-performance concrete compressive strength using tabular generative adversarial networks</article-title>. <source>Materials</source>. <year>2020</year>;<volume>13</volume>(<issue>21</issue>):<fpage>4757</fpage>. doi:<pub-id pub-id-type="doi">10.3390/ma13214757</pub-id>; <pub-id pub-id-type="pmid">33114394</pub-id></mixed-citation></ref>
<ref id="ref-43"><label>[43]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Abunassar</surname> <given-names>N</given-names></string-name>, <string-name><surname>Alas</surname> <given-names>M</given-names></string-name>, <string-name><surname>Ali</surname> <given-names>SIA</given-names></string-name></person-group>. <article-title>Prediction of compressive strength in self-compacting concrete containing fly ash and silica fume using ANN and SVM</article-title>. <source>Arabian J Sci Eng</source>. <year>2023</year>;<volume>48</volume>(<issue>4</issue>):<fpage>5171</fpage>&#x2013;<lpage>84</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s13369-022-07359-3</pub-id>.</mixed-citation></ref>
<ref id="ref-44"><label>[44]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>G&#x00FC;&#x00E7;l&#x00FC;</surname> <given-names>U</given-names></string-name>, <string-name><surname>van Gerven</surname> <given-names>MA</given-names></string-name></person-group>. <article-title>Deep neural networks reveal a gradient in the complexity of neural representations across the ventral stream</article-title>. <source>J Neurosci</source>. <year>2015</year>;<volume>35</volume>(<issue>27</issue>):<fpage>10005</fpage>&#x2013;<lpage>14</lpage>. doi:<pub-id pub-id-type="doi">10.1523/jneurosci.5023-14.2015</pub-id>; <pub-id pub-id-type="pmid">26157000</pub-id></mixed-citation></ref>
<ref id="ref-45"><label>[45]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Nguyen</surname> <given-names>TH</given-names></string-name>, <string-name><surname>Nguyen</surname> <given-names>T</given-names></string-name>, <string-name><surname>Truong</surname> <given-names>TT</given-names></string-name>, <string-name><surname>Doan</surname> <given-names>DTV</given-names></string-name>, <string-name><surname>Tran</surname> <given-names>DH</given-names></string-name></person-group>. <article-title>Corrosion effect on bond behavior between rebar and concrete using Bayesian regularized feed-forward neural network</article-title>. <source>Structures</source>. <year>2023</year>;<volume>51</volume>(<issue>3</issue>):<fpage>1525</fpage>&#x2013;<lpage>38</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.istruc.2023.03.128</pub-id>.</mixed-citation></ref>
<ref id="ref-46"><label>[46]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kamran</surname> <given-names>M</given-names></string-name>, <string-name><surname>Haider</surname> <given-names>SA</given-names></string-name>, <string-name><surname>Akram</surname> <given-names>T</given-names></string-name>, <string-name><surname>Naqvi</surname> <given-names>SR</given-names></string-name>, <string-name><surname>He</surname> <given-names>SK</given-names></string-name></person-group>. <article-title>Prediction of IV curves for a superconducting thin film using artificial neural networks</article-title>. <source>Superlatt Microstruct</source>. <year>2016</year>;<volume>95</volume>(<issue>1</issue>):<fpage>88</fpage>&#x2013;<lpage>94</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.spmi.2016.04.018</pub-id>.</mixed-citation></ref>
<ref id="ref-47"><label>[47]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Demirt&#x00FC;rk</surname> <given-names>D</given-names></string-name>, <string-name><surname>Mintemur</surname> <given-names>&#x00D6;</given-names></string-name>, <string-name><surname>Arslan</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Optimizing LightGBM and XGBoost algorithms for estimating compressive strength in high-performance concrete</article-title>. <source>Arab J Sci Eng</source>. <year>2025</year>;<volume>52</volume>(<issue>7</issue>):<fpage>129657</fpage>. doi:<pub-id pub-id-type="doi">10.1007/s13369-025-10217-7</pub-id>.</mixed-citation></ref>
<ref id="ref-48"><label>[48]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Mai</surname> <given-names>HVT</given-names></string-name>, <string-name><surname>Nguyen</surname> <given-names>TA</given-names></string-name>, <string-name><surname>Ly</surname> <given-names>HB</given-names></string-name>, <string-name><surname>Tran</surname> <given-names>VQ</given-names></string-name></person-group>. <article-title>Prediction compressive strength of concrete containing GGBFS using random forest model</article-title>. <source>Adv Civil Eng</source>. <year>2021</year>;<volume>2021</volume>(<issue>1</issue>):<fpage>6671448</fpage>. doi:<pub-id pub-id-type="doi">10.1155/2021/6671448</pub-id>.</mixed-citation></ref>
<ref id="ref-49"><label>[49]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Paranhos</surname> <given-names>RS</given-names></string-name>, <string-name><surname>Cazacliu</surname> <given-names>BG</given-names></string-name>, <string-name><surname>Sampaio</surname> <given-names>CH</given-names></string-name>, <string-name><surname>Petter</surname> <given-names>CO</given-names></string-name>, <string-name><surname>Neto</surname> <given-names>RO</given-names></string-name>, <string-name><surname>Huchet</surname> <given-names>F</given-names></string-name></person-group>. <article-title>A sorting method to value recycled concrete</article-title>. <source>J Cleaner Product</source>. <year>2016</year>;<volume>112</volume>:<fpage>2249</fpage>&#x2013;<lpage>58</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.jclepro.2015.10.021</pub-id>.</mixed-citation></ref>
<ref id="ref-50"><label>[50]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Van Tran</surname> <given-names>M</given-names></string-name>, <string-name><surname>La</surname> <given-names>H</given-names></string-name>, <string-name><surname>Nguyen</surname> <given-names>T</given-names></string-name></person-group>. <article-title>Hybrid machine learning for predicting hydration heat in pipe-cooled mass concrete structures</article-title>. <source>Constr Build Mater</source>. <year>2025</year>;<volume>481</volume>:<fpage>141558</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.conbuildmat.2025.141558</pub-id>.</mixed-citation></ref>
<ref id="ref-51"><label>[51]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ogunsanya</surname> <given-names>M</given-names></string-name>, <string-name><surname>Isichei</surname> <given-names>J</given-names></string-name>, <string-name><surname>Desai</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Grid search hyperparameter tuning in additive manufacturing processes</article-title>. <source>Manufact Lett</source>. <year>2023</year>;<volume>35</volume>:<fpage>1031</fpage>&#x2013;<lpage>42</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.mfglet.2023.08.056</pub-id>.</mixed-citation></ref>
<ref id="ref-52"><label>[52]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Prechelt</surname> <given-names>L</given-names></string-name></person-group>. <article-title>Automatic early stopping using cross validation: quantifying the criteria</article-title>. <source>Neural Netw</source>. <year>1998</year>;<volume>11</volume>(<issue>4</issue>):<fpage>761</fpage>&#x2013;<lpage>7</lpage>. doi:<pub-id pub-id-type="doi">10.1016/s0893-6080(98)00010-0</pub-id>; <pub-id pub-id-type="pmid">12662814</pub-id></mixed-citation></ref>
<ref id="ref-53"><label>[53]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kim</surname> <given-names>JH</given-names></string-name>, <string-name><surname>Lee</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Efficient dataset collection for concrete crack detection with spatial-adaptive data augmentation</article-title>. <source>IEEE Access</source>. <year>2023</year>;<volume>11</volume>:<fpage>121902</fpage>&#x2013;<lpage>13</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2023.3328243</pub-id>.</mixed-citation></ref>
<ref id="ref-54"><label>[54]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Asteris</surname> <given-names>P</given-names></string-name>, <string-name><surname>Kolovos</surname> <given-names>K</given-names></string-name></person-group>. <article-title>Self-compacting concrete strength prediction using surrogate models</article-title>. <source>Neural Comput Appl</source>. <year>2019</year>;<volume>31</volume>(<issue>Suppl 1</issue>):<fpage>409</fpage>&#x2013;<lpage>24</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s00521-017-3007-7</pub-id>.</mixed-citation></ref>
<ref id="ref-55"><label>[55]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>Z</given-names></string-name></person-group>. <article-title>Validation, robustness, and accuracy of perturbation-based sensitivity analysis methods for time-series deep learning models</article-title>. <comment>arXiv:2401.16521. 2024</comment>.</mixed-citation></ref>
<ref id="ref-56"><label>[56]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Lexicon</surname> <given-names>SP</given-names></string-name></person-group>. <article-title>Partial dependence plots</article-title>; <year>2025 [Internet]</year>. <comment>[cited 2025 Sep 24]</comment>. Available from: <ext-link ext-link-type="uri" xlink:href="https://softwarepatternslexicon.com/machine-learning/model-validation-and-evaluation-patterns/advanced-evaluation-techniques/partial-dependence-plots/">https://softwarepatternslexicon.com/machine-learning/model-validation-and-evaluation-patterns/advanced-evaluation-techniques/partial-dependence-plots/</ext-link>.</mixed-citation></ref>
<ref id="ref-57"><label>[57]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Molnar</surname> <given-names>C</given-names></string-name>, <string-name><surname>Freiesleben</surname> <given-names>T</given-names></string-name>, <string-name><surname>K&#x00F6;nig</surname> <given-names>G</given-names></string-name>, <string-name><surname>Herbinger</surname> <given-names>J</given-names></string-name>, <string-name><surname>Reisinger</surname> <given-names>T</given-names></string-name>, <string-name><surname>Casalicchio</surname> <given-names>G</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Relating the partial dependence plot and permutation feature importance to the data generating process</article-title>. In: <conf-name>World Conference on Explainable Artificial Intelligence</conf-name>. <publisher-loc>Cham, Switzerland</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2023</year>. p. <fpage>456</fpage>&#x2013;<lpage>79</lpage>.</mixed-citation></ref>
<ref id="ref-58"><label>[58]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Apley</surname> <given-names>DW</given-names></string-name>, <string-name><surname>Zhu</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Visualizing the effects of predictor variables in black box supervised learning models</article-title>. <source>J Royal Statist Soc Series B Statist Method</source>. <year>2020</year>;<volume>82</volume>(<issue>4</issue>):<fpage>1059</fpage>&#x2013;<lpage>86</lpage>. doi:<pub-id pub-id-type="doi">10.1111/rssb.12377</pub-id>.</mixed-citation></ref>
<ref id="ref-59"><label>[59]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Hover</surname> <given-names>KC</given-names></string-name></person-group>. <article-title>The influence of water on the performance of concrete</article-title>. <source>Constr Build Mater</source>. <year>2011</year>;<volume>25</volume>(<issue>7</issue>):<fpage>3003</fpage>&#x2013;<lpage>13</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.conbuildmat.2011.01.010</pub-id>.</mixed-citation></ref>
<ref id="ref-60"><label>[60]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Khan</surname> <given-names>M</given-names></string-name>, <string-name><surname>Hussain</surname> <given-names>A</given-names></string-name>, <string-name><surname>Raza</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Studying the usability of recycled aggregate to produce new concrete</article-title>. <source>J Eng Appl Sci</source>. <year>2024</year>;<volume>74</volume>(<issue>1</issue>):<fpage>1</fpage>&#x2013;<lpage>12</lpage>.</mixed-citation></ref>
<ref id="ref-61"><label>[61]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ali</surname> <given-names>S</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Mehmood</surname> <given-names>T</given-names></string-name></person-group>. <article-title>Examining the influence of recycled aggregates on the fresh and mechanical characteristics of high-strength concrete</article-title>. <source>Sustainability</source>. <year>2024</year>;<volume>16</volume>(<issue>20</issue>):<fpage>9052</fpage>. doi:<pub-id pub-id-type="doi">10.3390/su16209052</pub-id>.</mixed-citation></ref>
<ref id="ref-62"><label>[62]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kumar</surname> <given-names>R</given-names></string-name>, <string-name><surname>Rao</surname> <given-names>P</given-names></string-name></person-group>. <article-title>Experimental study on strength behaviour of recycled aggregate concrete</article-title>. <source>Int J Eng Res Technol (IJERT)</source>. <year>2013</year>;<volume>2</volume>(<issue>5</issue>):<fpage>2278</fpage>&#x2013;<lpage>0181</lpage>. doi:<pub-id pub-id-type="doi">10.17577/IJERTV2IS100045</pub-id>.</mixed-citation></ref>
</ref-list>
</back></article>













