Next Article in Journal
Machine-Learning-Based Coefficient of Performance Prediction Model for Heat Pump Systems
Previous Article in Journal
Antimicrobial Activity and In Silico Molecular Docking Studies of Pentacyclic Spiro[oxindole-2,3′-pyrrolidines] Tethered with Succinimide Scaffolds
Previous Article in Special Issue
Sustainability Evaluation Using a Life Cycle and Circular Economy Approach in Precast Concrete with Waste Incorporation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Compressive Strength of Fly-Ash-Based Concrete Using Ensemble and Non-Ensemble Supervised Machine-Learning Approaches

1
School of Architectural Engineering, The Open University of Henan, Zhengzhou 450008, China
2
Anyang Institute of Technology, School of Civil and Architectural Engineering, Anyang 455000, China
3
Faculty of Civil Engineering, Cracow University of Technology, 24 Warszawska Str., 31-155 Cracow, Poland
4
Department of Civil Engineering, Abbottabad Campus, COMSATS University Islamabad, Abbottabad 22060, Pakistan
5
Department of Mathematics and Statistics, Riphah International University I-14, Islamabad 44000, Pakistan
6
Department of Civil Engineering, College of Engineering in Al-Kharj, Prince Sattam Bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
7
Faculty of Civil Engineering and Resource Management, AGH University of Science and Technology, 30-059 Cracow, Poland
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2022, 12(1), 361; https://doi.org/10.3390/app12010361
Submission received: 2 November 2021 / Revised: 20 December 2021 / Accepted: 22 December 2021 / Published: 30 December 2021
(This article belongs to the Special Issue Sustainability and Performance of Advanced Construction Materials)

Abstract

:
The utilization of waste material, such as fly ash, in the concrete industry will provide a valuable alternative solution for creating an eco-friendly environment. However, experimental work is time-consuming; employing soft machine learning techniques can accelerate the process of forecasting the strength properties of concrete. Ensemble machine learning modeling using Python Jupyter Notebook was employed in the forecasting of compressive strength (CS) of high-performance concrete. Multilayer perceptron neuron network (MLPNN) and decision tree (DT) were used as individual learning which then ensembled with bagging and boosting to provide strong correlations. Random forest (RF) and gradient boosting regression (GBR) were also used for prediction. A total of 471 data points with input parameters (e.g., cement, fine aggregate, coarse aggregate, superplasticizer, water, days, and fly ash), and an output parameter of compressive strength (CS), were retrieved to train and test the individual learners. Cross-validation with K-fold and statistical error (i.e., MAE, MSE, RMSE, and RMSLE) analysis was applied to check the accuracy of all models. All models showed the best correlation with an ensemble model rather than an individual one. DT with AdaBoost and random forest gave a strong correlation of R2 = 0.89 with fewer errors. Cross-validation results revealed a good response with an error of less than 10 MPa. Thus, ensemble modeling not only trains the data by employing several weak learners but also produces a robust correlation that can then be used to model and predict the mechanical performance of concrete.

1. Introduction

Cement is an important material in the manufacture of concrete [1]. Its production in industry has a significant influence on global warming. Climate change and the release of gases during the manufacture of cement has a harmful effect on the Earth [2,3,4,5,6]. Nearly a billion tons of gases are emitted during its production due to the burning of natural resources and fossil fuels. Thus, apart from its benefits in the construction industry the manufacture of concrete has negative effects on the environment [7]. Efforts have been made to make new binders that will have substantial effects on the properties of concrete but have minimum effect on the environment [8]. One approach is to use waste-hazard material in making high-performance concrete (HPC). Its use in making an alternative binder not only minimizes disposal on the land but can produce significant benefits in making better, stronger concrete [9]. HPC has many advantages in terms of workability, durability, and strength as compared to traditional cement [10]. However, supplementary materials, such as silica fume, fly ash, and blast furnace slag, and chemical admixture are normally required to make HPC.
Concrete is classified as the most used material after water [11]. This is due to its several advantages in the domain of civil engineering [12]. However, making concrete requires deep study of the constituents used and often requires experimental work to achieve the target compressive strength. This creates uncertainties in making concrete as it is a mixture of fine aggregate, coarse aggregate, cement, admixtures and supplementary raw materials (SRMs) [13]. These constituents are thus randomly distributed in the concrete matrix. This uncertainty in making HPC can result in different strengths and consume much time to achieve a precise strength [14]. HPC is dependent on many factors, such as the size of particles, the water-cement ratio, and aggregate proportions. Generally, physical measures are carried out to obtain compressive strength. This is achieved by testing various cubes and cylinders with mixed design ratios in the laboratory [15]. However, the process of laboratory work to ensure strength is quite uneconomical and time-consuming. Thus, efforts have been made to predict the variables using regression and machine learning models [16,17,18,19]. This can not only reduce uncertainty in prediction but can ensure the required quantities in mix design to give strength. These supervised machine-learning algorithms identify robust relationships and predict the best model using a set of input variables.
Numerous algorithms, classified as supervised and unsupervised learning, have been utilized in the past to forecast the CS of concrete. Shahmansouri et al. [20] used gene expression programming (GEP) by utilizing waste material granulated blast furnace slag (GGBS) with 351 data points in the prediction of geopolymer concrete properties (GPC). The author used sodium hydroxide (NaOH), days, natural zeolite (NZ), silica fume (SF), and granulated blast-furnace slag (GBFS) as input variables, and compressive strength as the output variable. The results revealed that using a GEP model, the training set and validation set gave strong correlations of R2 = 0.918 and 0.940, respectively. Awoyera et al. [21] utilized both (GEP) and artificial neuron network (ANN) machine learning algorithms in the prediction of GPC strength using FA, GGBS, SF, slump flow and several other input parameters to predict the outcomes of compressive, tensile and flexural strength. The results revealed that the GEP model excelled as it described the empirical relationships more effectively than ANN. Dantas et al. [22] and Duan et al. [23] used ANN applying it to the use of recycled aggregate and waste material from construction with 1178 data points in the forecasting of concrete properties. Nguyen et al. [24] used a hybrid algorithm using least square support vector machine (LSSVM) and particle swarm optimization (PSO) for the prediction of properties of fresh concrete with 142 data points. Zhang et al. [25] used such models for the prediction of sand concrete properties. Siddique et al. [26] used bottom ash (BA) in self-compacting concrete (SCC) and predicted its properties using the ANN technique with 80 mixes from literature studies and 31 samples from experimental work. Tran et al. [27] used the ANN model for the prediction of the strength of concrete-filled steel tube (CSFT) with ultra-high-strength concrete (768 data sets) and reported a strong correlation. Iqbal et al. [28] employed the GEP technique for prediction of green concrete properties utilizing waste foundry sand as a constituent with a 234-set database for compressive strength and 163 sets for splitting test. Congro et al. [29] used an ANN algorithm for the prediction of the flexural strength of fiber-reinforced concrete (FRC) with 400 data samples constructed from literature review. Machine learning is a useful tool in civil engineering and many other domains in the prediction of properties. Table 1 illustrates the various machine learning algorithms used by researchers.
Two machine learning (ML) approaches were used for modeling and predictions for making a model. One was the standard technique, usually maintained on a separate model, while another was the ensemble approach, such as bagging, boosting, and AdaBoost [45]. The outcomes of research on these ensemble models has demonstrated that these newly developed models are more effective than an individual ML model [46]. The ensemble ML approaches have a different way of operating; these techniques normally initially train multiple weak learners employing training data and integrate the weaker ones into a strong learner. Weak learners are the individual techniques such as DT, ANN, or SVM. The prediction of the properties of objects/materials, such as concrete, using ensemble ML algorithms can be effectively carried out.
Recently, the application of ensemble ML algorithms to forecast the performance of materials with high accuracy has attracted great interest from researchers. The popularity of ensemble ML techniques has increased and several studies have found that these types of algorithms can more accurately predict outcomes compared to individual ML approaches [47]. This research focused on the application of various machine learning algorithms, including bagging, multi-boosting, rotation boosting, and random sub-spaces, to determine landslide vulnerability and comparison between results. The accuracy of the technique (MLPNN) was evaluated using an ensemble-learning approach. It was found that the multi-boosting technique was more effective than the other approaches. The application of ensemble ML approaches is listed in Table 2.
This research described the application of various ML approaches to forecasting the CS of high-performance concrete containing waste material (fly ash). The novelty, as well as the significance of this study, is twofold. The aim was based on challenging the execution of network-based DT, MLPNN models considering boosting and bagging with AdaBoost, and random forest, gradient boosting as an ensemble technique for forecasting. Additionally, this research considered the performance of tree-based and network-based models, as well as the correlation between the findings from both tree-based and network-based and learning, using several ML basic learners. More precisely, this study includes a description of the correct parameters for collecting appropriate component models and can be considered as a critical metric for model design. Second, this research compared and implemented ensemble learning methodologies to other learning models (individual) to forecast the strength of HPC using waste material fly ash as a crucial ingredient. To the authors’ knowledge, no equivalent study utilizing ensemble machine learning modeling for HPC exists in the literature. These learner programs were written in Anaconda Navigator software version 1.9.12 using Jupyter Notebook and Python 3.7.

2. Database Description

The database required to estimate concrete strength utilizing waste materials was gathered from the literature [54] and is summarized in Appendix A. Seven characteristics were utilized to forecast the CS of the final product, including days, cement, fine aggregate, binder water, and superplasticizer, coarse aggregate, as well as waste items, such as fly ash. Their distribution correlations with the compressive strength’s input parameters are shown in Figure 1.
The input variables had a significant effect on the model’s outcomes. The employed parameters used to run the models for the prediction of CS of high-performance concrete also played a positive role in the prediction. The Spyder environment from Anaconda software (version 3.7) was used to run the various models with the help of Python coding. For input parameters, the distributions are shown in Figure 2. The obtained data (results), in the form of their performance of the employed technique, was severely affected by the parameters. In an artificial approach, ML is effective in neuron-based artificial techniques to forecast the number of properties by analyzing the important concentration of parameters. Python applies the ML approaches and plots the correlation with the help of the Seaborn command. Figure 3 represents the maximum values of parameters in predicting its outcome CS. It was found and can be seen that the contours indicate individual influences on strength. Moreover, applying these concentrations in performing experimental work yields the maximum outcome. The ranges, as well as the data variables employed to run the model, can be seen in Table 3 and Table 4, respectively.

3. Machine Learning Methods

Various researchers, when forecasting and evaluating the behavior of materials, commonly use ML algorithms. In this study, the prediction component of high-performance concrete’s CS was addressed by the deployment of machine learning-based methodologies including artificial neural network, random forest, decision tree, and gradient-boosting network. The selection of these approaches was based on their popularity and high accuracy in prediction. However, another learner was then applied for modeling the concrete’s strength. The flow diagram indicating the process of ensemble learning is shown in Figure 4.

3.1. Overview of Machine Learning Algorithms

3.1.1. DT Algorithm

A decision tree is a subset of machine learning that has a tree-like structure composed of branches and nodes. The nodes with outgoing corners are referred to as inner nodes, whereas the nodes without outgoing corners are referred to as leaves. Through an inner node corresponding to a given function, the instance used for regression or classification is separated into numerous classes. The input parameters are treated as a special function during the training step. The stimulant for the decision tree is a method that is seen as a decision-maker tree from supplied instances. By reducing the fitness function, the conducted technique determines the best decision tree. Due to the absence of classes in the given datasets, the regressor model was used in this study for the target parameter using the independent variable. For each parameter, the dataset is segmented at multiple places. The performed technique determines the difference between the expected and actual values for the pre-specified fitness function at each division point. The mistakes in the divided region are equal across the variables, and the parameters with the lowest fitness function values are chosen as the split point. This practice is repeated indefinitely to improve the model.

3.1.2. Random Forest (RF) Regressor

The random forest model is a classification and regression technique that has been the focus of numerous studies [32,53]. Shaqadan [55] employed the RF model to forecast the concrete’s CS. The primary distinction between RF and DT is that in DT a single tree is produced but, in RF, a forest of trees is generated and non-similar data are arbitrarily allocated to all the trees. For each tree, the database is assigned in the form of rows and columns, with distinct column and row dimensions determined [56]. The growth of every tree is based on the multiple factors mentioned below:
  • The frame of the given data is two-thirds of the total data collected randomly for each tree. This is referred to as bagging. Forecasted parameters are chosen freely and the node-splitting algorithm uses the finest split on these parameters.
  • For each tree, the out-of-bag error is determined using the remaining data. Additionally, errors from each tree are accumulated to determine the ultimate out-of-bag error rate.
  • Each tree displays a regression, and the model chooses the prediction with the most votes out of all the trees in the forest. These can be zeroes or ones. Forecasting probability is defined as the proportion of 1’s obtained.
Random forest (RF) is considered one of the most sophisticated ensemble approaches. It combines the advantages of variable importance measures (VIMs) with a small number of model parameters, as well as robust overfitting resistance [57,58]. It uses a decision tree as its base predictor to the random forest. Models constructed using RF can produce acceptable results along with the default setting of the parameter [59]. By utilizing RF, it is possible to minimize the number of possible combinations of base predictors and parameter values to one. Many applications for utilizing RF can be found in different domains, such as ecology [60,61,62] and bioinformatics [63,64,65], but it has seldom been applied to concrete in civil engineering [66].

3.1.3. ANN Approach

ANN approaches simulate the microstructure (neurons) of a biological nervous system [67]. ANNs are computational processing information networks emulating the biological connection between neurons. In artificial neural networks, the models perform a major role in performing neural activities [68]. The motivation in developing artificial neural networks (ANNs) originated from emulation of the human brain to implement neural intelligent tasks analogous to that of the human brain [69]. Neurons are the human brain’s basic unit, which underpins the capacity to memorize and think based on experience. These cells are present in the form of neuron networks. Generally, these cells receive input from other sources, perform non-linear neural activities and generate finalized output. In a similar way to other fields, ANNs are also utilized in engineering applications [70]. Recently, these neural activities have been found to be successful in civil engineering structural detection applications, identification behavior of the structural system, observation of groundwater, and prediction of mechanical properties.
ANNs are comprised of a large number of dense parallel linkages. These cells receive data in the form of weighted inputs and then employ an activation function to communicate the weighted output to other neurons, resulting in the activation of a cascade of neurons. These neural activities are made up of one or more layers of activity. The multilayer perceptron network is commonly employed in brain processes such as memory and learning. These input parameters, which have been established based on previous experience, are processed using neural layer processing systems. The perception reaction is generated based on the number of input variables, the number of input layers, and the result. The input, hidden, and output layers are the three layers of a network with the addition of three layers. The hidden layer can be located between the input layer and the output layer, and may be part of a larger network of hidden layers than the input layer. A single hidden layer can be used to solve some of the problems with perceptrons; nevertheless, it is more important and more successful to use more than one hidden layer [71]. Figure 5 is a representation of a neural network architecture, which includes one input layer, two hidden layers, and one output layer, among other elements. Every neuron in the layers, except the input layer, begins by analyzing a linear combination in conjunction with a bias. The neurons then compute nonlinear functions in the buried layers of their input, which they then feedback to the computer. It is necessary to point out that a sigmoid function is a non-linear function [72].
ANNs mainly consists of five parts, including input, activation functions, sum function, weights, and output. ANNs can train their model by adapting its weight factor from outside the network or by adjusting within the layer in response to input [74]. The external source is used to provide input parameter information while the input’s outcome set can be described by the weight of the values. The submission of inputs and weights with bias is characterized as a sum function. The mathematical expression of the weighted sum function is calculated using Equation (1).
n e t j = i = 1 n w i j x i + b
The activation function processes the net input which is found after submission from the sum function and determines the output of the cell. Normally sigmoid, ramp and Gaussian functions are the most popular activation functions. A schematic diagram of a multilayer perceptron, showing input, sum function with sigmoid activation, and output, is shown in Figure 6 [75]. The output of j neuron based on input variables is computed by employing sigmoid functions as follows. The sigmoid function is a non-linear activation function, except the input layer activates in every layer. The overall net out-based output of j neuron is obtained using Equation (2).
o u t j = f n e t j = 1 1 + e α n e t j
The MLPNN feed-forward algorithm is used to execute the ANNs model in this study. Both the hidden layers and the changing neuron with two in number are taken into consideration while evaluating the superior performance of MLP [76]. In addition, a linearly activating function, along with a non-linear transit function, is used between the hidden layer and the input layer to verify the relationship between the input and output parameters. Furthermore, the database derived from the literature was subdivided into two sets, namely the training set and the testing set. Because the effect of overfitting data is a significant concern in machine learning, this was done to mitigate the effect of overfitting. It was proposed in a prior study that 80 percent of the randomly generated data should be used for training while only 20% of it should be used for testing [77].

3.1.4. Gradient Boosting Algorithm

Boosting is an effective approach that combines many simple base classifiers to create a committee whose output is substantially better than that of any base classification [9]. The core concept of boosting is to sequentially introduce new models to the ensemble model. A new, poor basic learner model is trained on each particular iteration about the failure of the entire ensemble that has so far been learned [78]. Gradient boosting constructs the model stage-wise, as do other boosting approaches, and generalizes it by optimizing an arbitrarily variational loss function [79]. Using the gradient descent approach, the problem of minimization is solved and a prediction model is generated in the form of an ensemble of weak prediction models. Thus, it can be used for classification or regression-based problems [79].
GBR is a simplification of gradient booster and consists of three elements: a loss function (to be optimized), a weak learner (to predict), and an additive model (to incorporate weak learners to minimize the loss function) [80]. The key idea behind this algorithm is to create a maximum correlation between the new basic teachers and the negative gradient of loss function associated with the entire ensemble [80]. The loss function may be arbitrary, but to provide a better intuition, the learning method can lead to consecutive error-fitting when the error function is the classic square-error loss. In general, the researchers are responsible for selecting the loss function; they have a rich range of loss functions and the possibility of applying their task-specific losses. The mathematical descriptions of GBR are as follows (see Equation (3)). Consider an advanced type model [79]:
F x = m = 1 M γ m h m x
where ℎm (x) are the root functions, which are typically called poor students in the boosting aspect. GBR uses fixed-size decision trees as slow learners. Decision trees have several capabilities which enhance their ability to handle mixed-type data and to model complex functions. Like other boosting algorithms, GBR develops the additive model in a forward stage-wise fashion, as represented in Equation (4) [79]:
F m x = F m 1 x + γ m h m x
The decision tree in that stage ℎm (x) is taken to reduce the loss function L, given the model Fm-1 and its fit Fm−1(xi) (see Equation (5)).
F m x = F m 1 x + a r g m i n   i = 1 n L y i , F m 1 x i h x
GBR tries numerically to solve this error function problem via gradient descent. The GBR gradient is the negative gradient of the loss function evaluated in the existing model Fm−1 that can be determined for any differentiable loss function, as indicated in Equation (6). The main gains of GBR are:
  • normal management of mixed form data,
  • high predictive control,
  • output space robustness (via robust loss functions)
  • supports various loss functions.
F m x = F m 1 x + γ m i = 1 n F L y i , F m 1 x i

3.2. Bagging and Bossing Approaches

Ensemble approaches are used to optimize data mining and machine learning operations. Additionally, these strategies assist in mitigating unneeded training data concerns by merging and pooling numerous suboptimal prediction models (i.e., sub-models of components). By carefully changing training data, it is possible to generate a large number of sub-models/classifier components (1, 2, …, M). This could lead to a more productive learner. The optimal predictive model, on the other hand, can be built by averaging combine procedures on verified sub-models. Bagging, along with the bootstrap resampling strategy and benefit collation, is recognized as one of the most prevalent ensemble modeling strategies. The component models are replaced by the first training set (bootstrap tests up to the training set size). Many data points may appear several times in product models, while others may not exist at all. Finally, the outcome is obtained by averaging the component model outputs.
The boosting technique generates a cumulative model, which results in more accurate component development than an individual model. Meanwhile, boosting positions dependent sub-models in the final model using weighted averages of dependent sub-models. The present work used MLPNN, decision trees, and random forests with gradient boosting, as well as boosting and bagging, to predict the mechanical properties of commonly used concrete.

Parameter Tuning for Ensemble Learner

The tuning parameters for the model employed in ensemble procedures can be (i) variables correlated with the ideal number of sample learners and (ii) learning rates and other characteristics that specifically impact ensemble techniques.
In this study, the optimal range of sub-models was determined, bagging and boosting ensemble models (20 each) with 10, 20, 30, ..., 200 component sub-models were constructed for base learners, and correlation with high coefficient values was used to choose the best model. Figure 7 illustrates the relationship between ensemble model accuracy and the number of component sub-models. According to the report, the ensemble model with boosting exhibited a high correlation coefficient with the expected aspect. Additionally, decision trees with bagging and boosting as ensemble models of 20 and 130 sub-models had a substantial relationship, as illustrated in Figure 7a,b. Similarly, ensemble machine learning with 20 and 40 ensemble models demonstrated a significant effect, as illustrated in Figure 7c,d.
Initially, the research showed that using ensemble modeling improved the accuracy of both models. However, employing 200 or more component sub-models for the bagging ML technique and 10 or more for the boosting algorithms yielded more accurate results. Table 5 describes the architectures used in the ensemble models. It is worth noting that the learning rate for the study was 0.9 when both ML ensemble ML methods, bagging, and boosting algorithms, were used.

3.3. 10 K-Fold Cross-Validation Using 10 K-Fold Method

Typically, k-fold cross-validation is used to reduce the random sampling bias associated with training and to hold out data samples. According to Kohavi’s assessment, a ten-fold validation strategy produces reliable variance while consuming the least computing time [81]. The performance of a model that classifies a fixed amount of data into 10 subsets is evaluated using a ten-fold cross-validation procedure. It selects a distinct data subset for testing purposes and trains the selected model using the remaining data subsets in each of the 10 rotations of model development and validation. The test subset is processed to ensure that the algorithm, as displayed in Figure 8, is accurate. The algorithm’s accuracy is then reported as the average of the accuracy gained from ten models across ten validation cycles.

3.4. Evaluation of Models Using a Statistical Measure

To better assess the effectiveness of the individual and ensemble learner prediction models, three types of indicators are available, which are as follows:
MAE = 1 n i = 1 n x i x
MSE = 1 n i = 1 n y p r e d y r e f 2
RMSE = y p r e d y r e f 2 N
RMSLE = 1 N \ i = 1 N ( log y i + 1 log y i + 1 2    

4. Model Result

4.1. The Outcome of the DT Model

The forecasted outcome of compressive strength of concrete from a supervised machine-learning algorithm produced an effective result, as shown in Figure 9. Moreover, the DT algorithm was modeled with ensemble algorithms. The anticipated result from the DT approach, without the use of ensemble approaches, demonstrated a strong relationship with an outcome of R2 = 0.803 with its error distribution, as shown in Figure 9a,b. The error distribution in Figure 9b shows that the average error of the testing set was approximately 4.41 MPa. In addition, the data shows that 89 percent of the data of the test set had an average error less than 10 MPa. The distribution of error data also shows that around 9 percent of the error lay between 10 to 15 MPa. Furthermore, as shown in Figure 9b, the distribution exhibited a maximum error of 22.84 MPa and a minimum error of 0.036 MPa. Individually, DT was anticipated to be a favorable model; however, when DT was combined with ensemble techniques, it yielded higher accuracy with maximum R2, as indicated in Figure 9c–f. Figure 9c,d show the DT with bagging ML approach, together with a regression model and a distribution of errors, respectively. The ensemble-bagging model predicted a positive outcome with a coefficient of correlation (R2) equal to 0.87 and a lower error in the testing data set than the other models, as illustrated in Figure 9c,d. The bagging ensemble algorithm had higher accuracy, showing 95.74 percent error below 10 Mpa, as indicated in Figure 9d. Similarly, the ensemble model with boosting algorithm showed a significant response with R2 = 0.89 (Figure 9e), compared to ensemble bagging with R2 = 0.87 with less error, as illustrated in Figure 9f. Both the models, after ensembling, produced a positive response showing less error, with an average error of approximately 4.41 MPa for the decision tree model, 3.54 MPa for the decision tree with bagging algorithm, and 2.96 MPa for the Adaboost algorithm. Overall, the bagging algorithm showed 19.72 MPa and the boosting algorithm showed 32.87 MPa model enhancement compared to the individual model. The trend analysis of the models is shown in Table 6.

4.2. MLPNN Model Outcomes

MLPNN networks belong to a class of supervised learning in artificial intelligence approaches, and their applications in concrete demonstrate an appreciable relationship between forecasting strength with actual strength. The outcome of the concrete prediction via MLPNN with its error distribution of the testing data with a coefficient of correlation R2 = 0.77 is demonstrated in Figure 10a,b. The MLPNN distribution yielded an average error value of 5.24 MPa for the test set, with a minimum error value of 0.09 MPa and a maximum error value of 19.57 MPa. However, a significant enhancement of R2 was observed using ensemble bagging and boosting techniques. The MLPNN-bagging algorithm showed R2 = 0.833 with an average error of approximately 3.94 Mpa, as shown in Figure 10c,d. Whereas, MLPNN with boosting techniques produced a higher correlation of R2 = 0.80 with an average error of approximately 4.42 Mpa, presented in Figure 10e,f. Even though the ensemble model trains itself on the weak learner to obtain a strong learner this ultimately produces a better model [82]. The data analysis of the bagging and boosting algorithm models in terms of errors shows that most of the data points of the predicted model had errors below 10 MPa. The bagging and boosting algorithms demonstrated a 92.55 and 94.68 percent enhancement of the model compared to the individual model without ensemble technique. Table 7 presents the results of the statistical analysis of testing data performed using MLPNN ensemble modeling.

4.3. RF Model Outcome

The RF algorithm is a combination of bagging and random feature target in ensemble machine learning techniques, which is regarded as an efficient and user-friendly approach for producing forecasting models [83]. Figure 11 illustrates the anticipated accuracy level for the supervised learning technique in HPC. It can be seen in Figure 11a that the RF-ensemble model shows a strong correlation of R2 = 0.89 with a robust predicted outcome. Additionally, the anticipated outcome of the RF model can also be analyzed using the error distribution, as illustrated in Figure 11b. The error distribution showed an average error value of 3.17 MPa of test data. Additionally, the 96.80 MPa result of the test set indicates that the error was less than 10 MPa, suggesting that it is accurate for nonlinear forecasting of strength for normal concrete. Table 8 contains statistical information on the data used in the random forest regression.

4.4. GB Model Outcome

The response of prediction by GBR is illustrated in Figure 12. The ensemble-based algorithms produced a strong correlation of R2 = 0.84 with less error compared to individual learners, as shown in Figure 12b. In addition, the statistical analysis revealed that 88 data points of the testing set showed lower error than 10 Mpa, with a maximum and minimum error of approximately 23.97 and 0 MPa. Moreover, if evaluated in terms of percentage, then 93.61 MPa represents a clearly lower error, as shown in Table 9.

4.5. K-Fold Cross-Validation Approach

It is important to achieve the required accuracy of the models that are being used for the prediction of the models. This is a crucial component for ensuring the effectiveness of anticipated models, and validation is critical in this regard. The process of validation is used to evaluate the accuracy of data by mixing it randomly, and the K-fold validation test was used in this research to evaluate model accuracy. This strategy is used to eliminate the bias associated with randomly picking the training database from the training database. It separates the samples of data obtained from the experimental approach into ten equal subgroups using a random number generator. It absorbs nine out of ten subsets to shape the effective learner, with the remaining subset being used to test the model’s predictions. The process is repeated approximately ten times to achieve the average accuracy of the repetitions performed. Many researchers consider that the use of this ten-fold cross-validation approach is effective in demonstrating both the generality and reliability of a model’s performance [81].
Figure 13 depicts the error measurement-based validation test of the nonlinear model and ensemble model (RF and GBR) for error estimation. The model validation is checked by different error indicators, including mean absolute error (MAE), root mean square error (RMSE), and root mean square log error (RMSLE), mean absolute error (MAE), and root mean square log error (RMSE), as presented in Figure 13. Even though there were fluctuations reported in the results, the accuracy remained high, as demonstrated in Figure 13a–f. The result of K-fold cross-validation for the decision tree model via MAE, RMSE, and RMSLE produced an error below 10MP, as shown in Figure 13a–f and Table 10, respectively. Similarly, the MLPNN with bagging and Adaboost showed the same trend by showing minimal errors (see Table 11). This was due to their lower bias in predicting a model. Moreover, random forest regression and gradient boosting regression showed a strong correlation in terms of k-fold cross-validation. It can be seen in Figure 13e–f and Table 12 that the error indicators for model validation showed less error than 10 MPa.

4.6. Results Evaluation of the Employed Models

Model evaluation of the ensemble model with individual algorithms was also conducted using statistical analysis through various error indicators. It can be seen in Table 13 that individual models of DT and MLPNN exhibited higher errors, as compared to their ensemble approaches. This means that ensemble methods not only provide accurate predictions but also help narrow the error range between anticipated and actual findings.

5. Limitations and Future Work

This study considered the vast variety of ML approaches and their comparative performance in terms of predicting the strength property of HPC. However, it does have certain limitations. The data included is indispensable for the efficiency of the forecasted models. The range of the database employed in this study was limited to 471. In addition, this research was also limited to compressive strength prediction and did not consider the flexural and corrosive behavior of concrete at high temperatures. Indeed, proper database and testing must be applied as they are vital elements for engineering applications. This study was based on a wide range of data sets with eight input variables; however, the database and input parameters can be increased to generate a better response of the employed models.
It is recommended that additional variables, such as elevated temperature, humidity, and other environmental conditions, are added as input variables to evaluate the performance of the model for the prediction of required outcomes. As concrete properties are severely affected by environmental conditions, the impact under different conditions should be explored with other deep ML algorithms, such as the convolutional neural network (CNN) and the recurrent neural network, and using a restricted Boltzmann machine (RBM).

6. Conclusions

The objective of this paper was to use and analyze the accuracy of individual learners (MLPNN, DT) with ensemble learners, using techniques such as bagging, boosting and Python Anaconda. In this approach, the data was trained and tested and a high correlation was obtained between predictions and targets. The following conclusions are drawn from our analysis:
  • The result of individual learners, DT and ANN, showed a strong correlation between predictions and targets with R2 = 0.80 and R2 = 0.77, respectively. However, ensemble learner with bagging and boosting and mostly boosting with Adaboost for DT outburst from the individual learner produced a stronger correlation R2 = 0.899, and ANN with bagging also produced a stronger correlation R2 = 0.833.
  • Optimization of ensemble models was conducted with 20 models ranging from 10 to 200 estimators (sub-models). A decision tree with boosting (ensemble = 130) and random forest (ensemble = 130) provided a robust strong correlation with R2 = 0.89.
  • It is evident that using ensemble learner with a weak learner showed less average error compared to the individual learner. Furthermore, K-fold cross-validation was used to validate models with coefficients of correlation, mean square error, and root mean square error. All the models had low MAE and RMSE errors and a high correlation R2. Fluctuations in validation were noticed, using K-fold validation to acquire data in steps and then performing validation on unknown data.
  • Statistical analysis was also performed by means of MAE, MSE, RMSE, and MSLE. All ensemble learners produced less error compared to the individual learner, with random forest bagging giving a lesser error compared with MAE, MSE, RMSE, and MSLE. RF and AdaBoost are supervised learning algorithms that yielded strong relationships between prediction and targets.

Author Contributions

Y.S.: conceptualization, writing—original draft and validation; J.Z.: data curation, and review; K.A.O.: writing—revised draft and supervision; M.F.J.: software, writing and review; A.A.: resources and writing—revised draft; M.I.K.: formal analysis and investigation; F.A.: visualization and editing; R.K.: writing—revised draft. All authors have read and agreed to the published version of the manuscript.

Funding

This research did not receive external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data generated or analyzed in this research are included in this published article.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

HPCHigh performance concrete
SRMsSupplementary raw materials
GEP Genetic engineering programming
GGBSGround granulated blast slag
GPCGeopolymer concrete
GBAGround bottom ash
CFSTConcrete-filled steel tube
ANNArtificial neuron network
MLMachine learning
HPCHigh-performance concrete
DLDeep learning
DTDecision tree
MLPNNMultilayer perceptron neural network
DMDeep machine
RFRandom forest
GBGradient boosting

Appendix A

S.NoCement (kg/m3)Fly Ash (kg/m3)Water (kg/m3)Superplasticizer (kg/m3)Coarse Aggregate (kg/m3)Fine Aggregate (kg/m3)Age (Day)Compressive Strength (MPa)
154001622.510406762879.99
254001622.510556762861.89
3475022809325942839.29
4380022809326709052.91
54750228093259418042.62
63800228093267036552.52
73800228093267027053.3
847502280932594738.6
94750228093259427042.13
10475022809325949042.23
113800228093267018053.1
12349019201047806.9315.05
134750228093259436541.93
1431001920971850.639.87
154850146011208002871.99
16531.30141.828.2852.1893.7341.3
17531.30141.828.2852.1893.7746.9
18531.30141.828.2852.1893.72856.4
19531.30141.828.2852.1893.75658.8
20531.30141.828.2852.1893.79159.2
21222.496.7189.34.5967.1870.3311.58
22222.496.7189.34.5967.1870.31424.45
23222.496.7189.34.5967.1870.32824.89
24222.496.7189.34.5967.1870.35629.45
25222.496.7189.34.5967.1870.310040.71
26233.894.6197.94.6947852.2310.38
27233.894.6197.94.6947852.21422.14
28233.894.6197.94.6947852.22822.84
29233.894.6197.94.6947852.25627.66
30233.894.6197.94.6947852.210034.56
31194.7100.5165.67.51006.4905.9312.45
32194.7100.5165.67.51006.4905.91424.99
33194.7100.5165.67.51006.4905.92825.72
34194.7100.5165.67.51006.4905.95633.96
35194.7100.5165.67.51006.4905.910037.34
36190.7125.4162.17.81090804315.04
37190.7125.4162.17.810908041421.06
38190.7125.4162.17.810908042826.4
39190.7125.4162.17.810908045635.34
40190.7125.4162.17.8109080410040.57
41212.1121.6180.35.71057.6779.3312.47
42212.1121.6180.35.71057.6779.31420.92
43212.1121.6180.35.71057.6779.32824.9
44212.1121.6180.35.71057.6779.35634.2
45212.1121.6180.35.71057.6779.310039.61
46230118.3195.54.61029.4758.6310.03
47230118.3195.54.61029.4758.61420.08
48230118.3195.54.61029.4758.62824.48
49230118.3195.54.61029.4758.65631.54
50230118.3195.54.61029.4758.610035.34
51190.3125.2161.99.91088.1802.639.45
52190.3125.2161.99.91088.1802.61422.72
53190.3125.2161.99.91088.1802.62828.47
54190.3125.2161.99.91088.1802.65638.56
55190.3125.2161.99.91088.1802.610040.39
56166.1163.3176.54.51058.6780.1310.76
57166.1163.3176.54.51058.6780.11425.48
58166.1163.3176.54.51058.6780.12821.54
59166.1163.3176.54.51058.6780.15628.63
60166.1163.3176.54.51058.6780.110033.54
61229.7118.2195.26.11028.1757.6313.36
62229.7118.2195.26.11028.1757.61422.32
63229.7118.2195.26.11028.1757.62824.54
64229.7118.2195.26.11028.1757.65631.35
65229.7118.2195.26.11028.1757.610040.86
66238.194.1186.77949.9847319.93
67238.194.1186.77949.98471425.69
68238.194.1186.77949.98472830.23
69238.194.1186.77949.98475639.59
70238.194.1186.77949.984710044.3
7125095.7187.45.5956.9861.2313.82
7225095.7187.45.5956.9861.21424.92
7325095.7187.45.5956.9861.22829.22
7425095.7187.45.5956.9861.25638.33
7525095.7187.45.5956.9861.210042.35
76212.5100.4159.38.71007.8903.6313.54
77212.5100.4159.38.71007.8903.61426.31
78212.5100.4159.38.71007.8903.62831.64
79212.5100.4159.38.71007.8903.65642.55
80212.5100.4159.38.71007.8903.610042.92
81212.6100.4159.410.41003.8903.8313.33
82212.6100.4159.410.41003.8903.81425.37
83212.6100.4159.410.41003.8903.82837.4
84212.6100.4159.410.41003.8903.85644.4
85212.6100.4159.410.41003.8903.810047.74
86212124.81597.81085.4799.5319.52
87212124.81597.81085.4799.51431.35
88212124.81597.81085.4799.52838.5
89212124.81597.81085.4799.55645.08
90212124.81597.81085.4799.510047.82
91231.8121.61746.71056.4778.5315.44
92231.8121.61746.71056.4778.51426.77
93231.8121.61746.71056.4778.52833.73
94231.8121.61746.71056.4778.55642.7
95231.8121.61746.71056.4778.510045.84
96251.4118.3188.55.81028.4757.7317.22
97251.4118.3188.55.81028.4757.71429.93
98251.4118.3188.55.81028.4757.72829.65
99251.4118.3188.55.81028.4757.75636.97
100251.4118.3188.55.81028.4757.710043.58
101251.4118.3188.56.41028.4757.7313.12
102251.4118.3188.56.41028.4757.71424.43
103251.4118.3188.56.41028.4757.72832.66
104251.4118.3188.56.41028.4757.75636.64
105251.4118.3188.56.41028.4757.710044.21
106181.4167169.67.61055.6777.8313.62
107181.4167169.67.61055.6777.81421.6
108181.4167169.67.61055.6777.82827.77
109181.4167169.67.61055.6777.85635.57
110181.4167169.67.61055.6777.810045.37
111290.496.2168.19.4961.2865322.5
112290.496.2168.19.4961.28651434.67
113290.496.2168.19.4961.28652834.74
114290.496.2168.19.4961.28655645.08
115290.496.2168.19.4961.286510048.97
116277.197.4160.611.8973.9875.6323.14
117277.197.4160.611.8973.9875.61441.89
118277.197.4160.611.8973.9875.62848.28
119277.197.4160.611.8973.9875.65651.04
120277.197.4160.611.8973.9875.610055.64
121295.795.6171.58.9955.1859.2322.95
122295.795.6171.58.9955.1859.21435.23
123295.795.6171.58.9955.1859.22839.94
124295.795.6171.58.9955.1859.25648.72
125295.795.6171.58.9955.1859.210052.04
126251.899.9146.112.41006899.8321.02
127251.899.9146.112.41006899.81433.36
128251.899.9146.112.41006899.82833.94
129251.899.9146.112.41006899.85644.14
130251.899.9146.112.41006899.810045.37
131249.198.8158.112.8987.8889315.36
132249.198.8158.112.8987.88891428.68
133249.198.8158.112.8987.88892830.85
134249.198.8158.112.8987.88895642.03
135249.198.8158.112.8987.888910051.06
136252.398.8146.314.2987.8889321.78
137252.398.8146.314.2987.88891442.29
138252.398.8146.314.2987.88892850.6
139252.398.8146.314.2987.88895655.83
140252.398.8146.314.2987.888910060.95
141246.8125.1143.3121086.8800.9323.52
142246.8125.1143.3121086.8800.91442.22
143246.8125.1143.3121086.8800.92852.5
144246.8125.1143.3121086.8800.95660.32
145246.8125.1143.3121086.8800.910066.42
146275.1121.4159.59.91053.6777.5323.8
147275.1121.4159.59.91053.6777.51438.77
148275.1121.4159.59.91053.6777.52851.33
149275.1121.4159.59.91053.6777.55656.85
150275.1121.4159.59.91053.6777.510058.61
151297.2117.5174.89.51022.8753.5321.91
152297.2117.5174.89.51022.8753.51436.99
153297.2117.5174.89.51022.8753.52847.4
154297.2117.5174.89.51022.8753.55651.96
155297.2117.5174.89.51022.8753.510056.74
156213.7174.7154.810.21053.5776.4317.57
157213.7174.7154.810.21053.5776.41433.73
158213.7174.7154.810.21053.5776.42840.15
159213.7174.7154.810.21053.5776.45646.64
160213.7174.7154.810.21053.5776.410050.08
161213.5174.2154.611.71052.3775.5317.37
162213.5174.2154.611.71052.3775.51433.7
163213.5174.2154.611.71052.3775.52845.94
164213.5174.2154.611.71052.3775.55651.43
165213.5174.2154.611.71052.3775.510059.3
166218.9124.1158.511.31078.7794.9315.34
167218.9124.1158.511.31078.7794.91426.05
168218.9124.1158.511.31078.7794.92830.22
169218.9124.1158.511.31078.7794.95637.27
170218.9124.1158.511.31078.7794.910046.23
1713760214.601003.5762.4316.28
1723760214.601003.5762.41425.62
1733760214.601003.5762.42831.97
1743760214.601003.5762.45636.3
1753760214.601003.5762.410043.06
176500014049668532867.57
177475591421.910986412857.23
17850560195010306302864.02
179451016511.310307452878.8
18051601628.28018022841.37
18152001705.28558552860.28
18252801856.99207202856.83
18352001755.28708052851.02
184385136158209037682855.55
185500.1020031124.4613.22844.13
1864050175011206952852.3
18751601628.38018022841.37
18847501629.510446622858.52
1895000151910336552869.84
190165143.6163.801005.6900.9314.4
191190.3125.2166.69.91079798.9312.55
19225095.7191.85.3948.9857.238.49
193213.5174.2159.211.71043.6771.9315.61
194194.7100.5170.27.5998901.8312.18
195251.4118.3192.95.81043.6754.3311.98
196165143.6163.801005.6900.91416.88
197190.3125.2166.69.91079798.91419.42
19825095.7191.85.3948.9857.21424.66
199213.5174.2159.211.71043.6771.91429.59
200194.7100.5170.27.5998901.81424.28
201251.4118.3192.95.81043.6754.31420.73
202165143.6163.801005.6900.92826.2
203190.3125.2166.69.91079798.92824.85
20425095.7191.85.3948.9857.22827.22
205213.5174.2159.211.71043.6771.92844.64
206194.7100.5170.27.5998901.82837.27
207251.4118.3192.95.81043.6754.32833.27
208165143.6163.801005.6900.95636.56
209190.3125.2166.69.91079798.95631.72
21025095.7191.85.3948.9857.25639.64
211213.5174.2159.211.71043.6771.95651.26
212194.7100.5170.27.5998901.85643.39
213251.4118.3192.95.81043.6754.35639.27
214165143.6163.801005.6900.910037.96
215190.3125.2166.69.91079798.910033.56
21625095.7191.85.3948.9857.210041.16
217213.5174.2159.211.71043.6771.910052.96
218194.7100.5170.27.5998901.810044.28
219251.4118.3192.95.81043.6754.310040.15
22043602180838.4719.72823.85
22128901920913.2895.39032.07
22228901920913.2895.3311.65
22339301920940.6785.6319.2
22439301920940.6785.69048.85
22539301920940.6785.62839.6
22648001920936.2712.22843.94
22748001920936.2712.2734.57
22848001920936.2712.29054.32
22948001920936.2712.2324.4
23033301920931.2842.6315.62
23125501920889.89459021.86
23225501920889.8945710.22
23328901920913.2895.3714.6
23425501920889.89452818.75
23533301920931.2842.62831.97
23633301920931.2842.6723.4
23728901920913.2895.32825.57
23833301920931.2842.69041.68
23939301920940.6785.6727.74
24025501920889.894538.2
2413970185.701040.6734.32833.08
242382.50185.701047.8739.3724.07
243295.80185.701091.4769.3714.84
2443970185.701040.6734.3725.45
245381.40185.701104.6784.32822.49
246295.80185.701091.4769.32825.22
247238.10185.701118.8789.32817.58
248339.20185.701069.2754.3721.18
249381.40185.701104.6784.3714.54
250339.20185.701069.2754.32831.9
251238.10185.701118.8789.3710.34
252252.50185.701111.6784.32819.77
253382.50185.701047.8739.32837.44
254252.50185.701111.6784.3711.48
25533901970968781313.22
25633901970968781720.97
257339019709687811427.04
258339019709687812832.04
259339019709687819035.17
2603390197096878118036.45
2613390197096878136538.89
2622360194096888536.47
263236019409688851412.84
264236019409688852818.42
265236019409688859021.95
2662360193096888518024.1
2672360193096888536525.08
268277019109688561421.26
269277019109688562825.97
27027701910968856311.36
271277019109688569031.25
2722770191096885618032.33
2732770191096885636033.7
2742540198096886339.31
275254019809688639026.94
2762540198096886318027.63
2772540198096886336529.79
2783070193096881218034.49
2793070193096881236536.15
28030701930968812312.54
281307019309688122827.53
282307019309688129032.92
2832360193096888579.99
28420001800112584577.84
2852000180011258452812.25
286225018101113833711.17
2872250181011138332817.34
288325018401063783717.54
2893250184010637832830.57
290275018301088808714.2
2912750183010888082824.5
292300018401075795715.58
2933000184010757952826.85
294375018601038758726.06
2953750186010387582838.21
2964000187010257452843.7
297400018701025745730.14
298250018201100820712.73
2992500182011008202820.87
300350018601050770720.28
3013500186010507702834.29
302310019201012830311.85
303310019201012830717.24
3043100192010128302827.83
3053100192010128309035.76
30631001920101283012038.7
307331019201025821314.31
308331019201025821717.44
3093310192010258212831.74
3103310192010258219037.91
31133101920102582112039.38
312349019201056809315.87
31334901920105680979.01
3143490192010568092833.61
3153490192010568099040.66
31634901920105680912040.86
317238018601119789712.05
3182380186011197892817.54
319296018601090769718.91
3202960186010907692825.18
321297018601040734730.96
322480019209367212843.89
323480019209367219054.28
3243970186010407342836.94
325281018601104774714.5
3262810185011047742822.44
327500020001125613112.64
328500020001125613326.06
329500020001125613733.21
3305000200011256131436.94
3315000200011256132844.09
332540017301125613752.61
3335400173011256131459.76
3345400173011256132867.31
3355400173011256139069.66
33654001730112561318071.62
33754001730112561327074.17
33835002030974775718.13
339350020309747751422.53
340350020309747752827.34
341350020309747755629.98
342350020309747759031.35
3433500203097477518032.72
3443850186096676316.27
34538501860966763314.7
34638501860966763723.22
347385018609667631427.92
348385018609667632831.35
3493310192097882518039
3503310192097882536041.24
351349019201047806314.99
35233101920978825313.52
353382018601047739724
3543820186010477392837.42
355382018601111784711.47
3562810186011047742822.44
357339018501069754721.16
3583390185010697542831.84
359295018501069769714.8
3602950185010697692825.18
3612380185011187892817.54
362296019201085765714.2
3632960192010857652821.65
3642960192010857659029.39
36533101920879825313.52
36633101920978825716.26
367331019209788252831.45
368331019209788259037.23
369349019201047806718.13
3703490192010478062832.72
3713490192010478069039.49
37234901920104780618041.05
37334901920104780636042.13
374302020309748171418.13
3753020203097481718026.74
37652501890112561318061.92
3775000200011256139047.22
37850002000112561318051.04
37950002000112561327055.16
380540017301125613341.64
381252018501111784713.71
3822520185011117842819.69
3833390185010607542831.65
38439301920940758319.11
385393019209407582839.58
386393019209407589048.79
387382018501047739724
3883820185010477392837.42
389252018601111784711.47
3902520185011117842819.69
39131001920970850714.99
392310019209708502827.92
393310019209708509034.68
3943100192097085018037.33
3953100192097085036038.11
396525018901125613333.8
397525018901125613742.42
3985250189011256131448.4
3995250189011256132855.94
4005250189011256139058.78
40152501890112561327067.11
402322020309748001420.77
403322020309748002825.18
4043220203097480018029.59
405302020309748172821.75
4063970185010407342839.09
40748001920936721324.39
40852201460896896750.51
409522014608968962874.99
410144175158189438442815.42
4113740190710137302839.05
412305100196109597052830.12
413151184167129917722815.57
4141651501821210237292818.03
41529810718668798152842.64
41631812621068617372840.06
417356142193118017782840.87
418164200181138498462815.09
419314113170109257832838.46
420321128182118707802837.26
421298107210118807442831.87
422322116196108188132831.18
423313113178810026892836.8
424296107221118197782831.42
42515211218489928162812.18
426300120212108787282823.84
4271481371581610028302817.95
428326138199118017922840.68
42915819522011898713288.54
4301511851671610746782813.46
43127390199119317622832.24
432336018239868172844.86
433145134181119798122813.2
434155143193910476972812.46
435135166180109618052813.29
436148182181158398842815.52
437298107164139537842835.86
43814517920288248692810.54
4393130178810008222825.1
4401551431939877868289.74
441313.3113178.581001.9688.72836.8
442296106.7221.410.5819.2778.42831.42
443151.6111.9184.47.9992815.92812.18
444299.8119.8211.59.9878.2727.62823.84
445148.1136.6158.116.11001.8830.12817.96
446326.5137.919910.8801.1792.52838.63
447158.4194.9219.711897.7712.9288.54
448150.7185.3166.715.61074.56782813.46
449272.689.6198.710.6931.3762.22832.25
450336.50181.93.4985.8816.82844.87
451144.8133.6180.811.1979.5811.52813.2
452154.8142.8193.39.11047.4696.72812.46
453134.7165.7180.210961804.92813.29
454148.1182.1181.415838.9884.32815.53
455298.1107.5163.612.8953.27842835.87
456145.4178.9201.77.8824868.72810.54
457312.70178.18999.7822.22825.1
458154.8142.8193.39.1877.2867.7289.74
459143.6174.9158.417.9942.7844.52815.42
460374.30190.26.71013.2730.42839.06
461304.899.61969.8959.4705.22830.12
462150.9183.9166.611.6991.2772.22815.57
463164.6150.4181.611.71023.3728.92818.03
464298.1107186.46.1879815.22842.64
465317.9126.5209.75.7860.5736.62840.06
466355.9141.6193.311801.4778.42840.87
467164.2200.1181.212.6849.38462815.09
468313.8112.6169.910.1925.3782.92838.46
469321.4127.9182.511.5870.1779.72837.27
470298.2107209.711.1879.6744.22831.88
471322.2115.619610.4817.9813.42831.18

References

  1. Akbar, A.; Farooq, F.; Shafique, M.; Aslam, F.; Alyousef, R.; Alabduljabbar, H. Sugarcane bagasse ash-based engineered geopolymer mortar incorporating propylene fibers. J. Build. Eng. 2021, 33, 101492. [Google Scholar] [CrossRef]
  2. Visintin, P.; Xie, T.; Bennett, B. A large-scale life-cycle assessment of recycled aggregate concrete: The influence of functional unit, emissions allocation and carbon dioxide uptake. J. Clean. Prod. 2020, 248, 119243. [Google Scholar] [CrossRef]
  3. Cang, Y.; Yang, L.; Luo, Z.; Zhang, N. Prediction of embodied carbon emissions from residential buildings with different structural forms. Sustain. Cities Soc. 2020, 54, 101946. [Google Scholar] [CrossRef]
  4. Thilakarathna, P.S.M.; Seo, S.; Baduge, K.S.K.; Lee, H.; Mendis, P.; Foliente, G. Embodied carbon analysis and benchmarking emissions of high and ultra-high strength concrete using machine learning algorithms. J. Clean. Prod. 2020, 262, 121281. [Google Scholar] [CrossRef]
  5. Farooq, F.; Akbar, A.; Khushnood, R.A.; Muhammad, W.L.B.; Rehman, S.K.U.; Javed, M.F. Experimental investigation of hybrid carbon nanotubes and graphite nanoplatelets on rheology, shrinkage, mechanical, and microstructure of SCCM. Materials 2020, 13, 230. [Google Scholar] [CrossRef] [Green Version]
  6. Farooq, F.; Jin, X.; Javed, M.F.; Akbar, A.; Shah, M.I.; Aslam, F.; Alyousef, R. Geopolymer concrete as sustainable material: A state of the art review. Constr. Build. Mater. 2021, 306, 124762. [Google Scholar] [CrossRef]
  7. Ahmad, A.; Chaiyasarn, K.; Farooq, F.; Ahmad, W.; Suparp, S.; Aslam, F. Compressive Strength Prediction via Gene Expression Programming (GEP) and Artificial Neural Network (ANN) for Concrete Containing RCA. Buildings 2021, 11, 324. [Google Scholar] [CrossRef]
  8. Lu, Z.; Su, L.; Xian, G.; Lu, B.; Xie, J. Durability study of concrete-covered basalt fiber-reinforced polymer (BFRP) bars in marine environment. Compos. Struct. 2020, 234, 111650. [Google Scholar] [CrossRef]
  9. Liu, T.; Wei, H.; Zou, D.; Zhou, A.; Jian, H. Utilization of waste cathode ray tube funnel glass for ultra-high performance concrete. J. Clean. Prod. 2020, 249, 119333. [Google Scholar] [CrossRef]
  10. Shen, D.; Jiao, Y.; Gao, Y.; Zhu, S.; Jiang, G. Influence of ground granulated blast furnace slag on cracking potential of high performance concrete at early age. Constr. Build. Mater. 2020, 241, 117839. [Google Scholar] [CrossRef]
  11. Li, J.; Wu, Z.; Shi, C.; Yuan, Q.; Zhang, Z. Durability of ultra-high performance concrete—A review. Constr. Build. Mater. 2020, 255, 119296. [Google Scholar] [CrossRef]
  12. Biskri, Y.; Achoura, D.; Chelghoum, N.; Mouret, M. Mechanical and durability characteristics of High Performance Concrete containing steel slag and crystalized slag as aggregates. Constr. Build. Mater. 2017, 150, 167–178. [Google Scholar] [CrossRef]
  13. Zhang, J.; Zhao, Y.; Li, H. Experimental Investigation and Prediction of Compressive Strength of Ultra-High Performance Concrete Containing Supplementary Cementitious Materials. Adv. Mater. Sci. Eng. 2017, 2017, 4563164. [Google Scholar] [CrossRef] [Green Version]
  14. Ahmad, W.; Farooq, S.H.; Usman, M.; Khan, M.; Ahmad, A.; Aslam, F.; Yousef, R.A.; Abduljabbar, H.A.; Sufian, M. Effect of coconut fiber length and content on properties of high strength concrete. Materials 2020, 13, 1075. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Kaloop, M.R.; Kumar, D.; Samui, P.; Hu, J.W.; Kim, D. Compressive strength prediction of high-performance concrete using gradient tree boosting machine. Constr. Build. Mater. 2020, 264, 120198. [Google Scholar] [CrossRef]
  16. Naderpour, H.; Rafiean, A.H.; Fakharian, P. Compressive strength prediction of environmentally friendly concrete using artificial neural networks. J. Build. Eng. 2018, 16, 213–219. [Google Scholar] [CrossRef]
  17. Azim, I.; Yang, J.; Javed, M.F.; Iqbal, M.F.; Mahmood, Z.; Wang, F.; Liu, Q.F. Prediction model for compressive arch action capacity of RC frame structures under column removal scenario using gene expression programming. Structures 2020, 25, 212–228. [Google Scholar] [CrossRef]
  18. Javed, M.F.; Farooq, F.; Memon, S.A.; Akbar, A.; Khan, M.A.; Aslam, F.; Alyousef, R.; Alabduljabbar, H.; Rehman, S.K.U.; Rehman, S.K.U.; et al. New prediction model for the ultimate axial capacity of concrete-filled steel tubes: An evolutionary approach. Crystals 2020, 10, 741. [Google Scholar] [CrossRef]
  19. Javed, M.F.; Amin, M.N.; Shah, M.I.; Khan, K.; Iftikhar, B.; Farooq, F.; Aslam, F.; Alyousef, R.; Alabduljabbar, H. Applications of gene expression programming and regression techniques for estimating compressive strength of bagasse ash based concrete. Crystals 2020, 10, 737. [Google Scholar] [CrossRef]
  20. Shahmansouri, A.A.; Bengar, H.A.; Ghanbari, S. Compressive strength prediction of eco-efficient GGBS-based geopolymer concrete using GEP method. J. Build. Eng. 2020, 31, 101326. [Google Scholar] [CrossRef]
  21. Awoyera, P.O.; Kirgiz, M.S.; Viloria, A.; Ovallos-Gazabon, D. Estimating strength properties of geopolymer self-compacting concrete using machine learning techniques. J. Mater. Res. Technol. 2020, 9, 9016–9028. [Google Scholar] [CrossRef]
  22. Dantas, A.T.A.; Leite, M.B.; Nagahama, K.D. Prediction of compressive strength of concrete containing construction and demolition waste using artificial neural networks. Constr. Build. Mater. 2013, 38, 717–722. [Google Scholar] [CrossRef]
  23. Duan, Z.H.; Kou, S.C.; Poon, C.S. Prediction of compressive strength of recycled aggregate concrete using artificial neural networks. Constr. Build. Mater. 2013, 40, 1200–1206. [Google Scholar] [CrossRef]
  24. Nguyen, T.D.; Tran, T.H.; Hoang, N.D. Prediction of interface yield stress and plastic viscosity of fresh concrete using a hybrid machine learning approach. Adv. Eng. Inform. 2020, 44, 101057. [Google Scholar] [CrossRef]
  25. Zhang, J.; Li, D.; Wang, Y. Toward intelligent construction: Prediction of mechanical properties of manufactured-sand concrete using tree-based models. J. Clean. Prod. 2020, 258, 120665. [Google Scholar] [CrossRef]
  26. Siddique, R.; Aggarwal, P.; Aggarwal, Y. Prediction of compressive strength of self-compacting concrete containing bottom ash using artificial neural networks. Adv. Eng. Softw. 2011, 42, 780–786. [Google Scholar] [CrossRef]
  27. Tran, V.L.; Thai, D.K.; Nguyen, D.D. Practical artificial neural network tool for predicting the axial compression capacity of circular concrete-filled steel tube columns with ultra-high-strength concrete. Thin-Walled Struct. 2020, 151, 106720. [Google Scholar] [CrossRef]
  28. Iqbal, M.F.; Liu, Q.f.; Azim, I.; Zhu, X.; Yang, J.; Javed, M.F.; Rauf, M. Prediction of mechanical properties of green concrete incorporating waste foundry sand based on gene expression programming. J. Hazard. Mater. 2020, 384, 121322. [Google Scholar] [CrossRef]
  29. Congro, M.; Monteiro, V.M.D.; Brandão, A.L.T.; Santos, B.F.D.; Roehl, D.; Silva, F.D. Prediction of the residual flexural strength of fiber reinforced concrete using artificial neural networks. Constr. Build. Mater. 2021, 303, 124502. [Google Scholar] [CrossRef]
  30. Babanajad, S.K.; Gandomi, A.H.; Mohammadzadeh, D.; Alavi, A.H. Numerical modeling of concrete strength under multiaxial confinement pressures using linear genetic programming. Autom. Constr. 2013, 36, 136–144. [Google Scholar] [CrossRef]
  31. Nikoo, M.; Torabian Moghadam, F.; Sadowski, Ł. Prediction of concrete compressive strength by evolutionary artificial neural networks. Adv. Mater. Sci. Eng. 2015, 2015, 849126. [Google Scholar] [CrossRef]
  32. Han, Q.; Gui, C.; Xu, J.; Lacidogna, G. A generalized method to predict the compressive strength of high-performance concrete by improved random forest algorithm. Constr. Build. Mater. 2019, 226, 734–742. [Google Scholar] [CrossRef]
  33. Golafshani, E.M.; Behnood, A.; Arashpour, M. Predicting the compressive strength of normal and High-Performance Concretes using ANN and ANFIS hybridized with Grey Wolf Optimizer. Constr. Build. Mater. 2020, 232, 117266. [Google Scholar] [CrossRef]
  34. Bilim, C.; Atiş, C.D.; Tanyildizi, H.; Karahan, O. Predicting the compressive strength of ground granulated blast furnace slag concrete using artificial neural network. Adv. Eng. Softw. 2009, 40, 334–340. [Google Scholar] [CrossRef]
  35. Öztaş, A.; Pala, M.; Özbay, E.; Kanca, E.; Caglar, N.; Bhatti, M.A. Predicting the compressive strength and slump of high strength concrete using neural network. Constr. Build. Mater. 2006, 20, 769–775. [Google Scholar] [CrossRef]
  36. Behnood, A.; Behnood, V.; Gharehveran, M.M.; Alyamac, K.E. Prediction of the compressive strength of normal and high-performance concretes using M5P model tree algorithm. Constr. Build. Mater. 2017, 142, 199–207. [Google Scholar] [CrossRef]
  37. Mousavi, S.M.; Aminian, P.; Gandomi, A.H.; Alavi, A.H.; Bolandi, H. A new predictive model for compressive strength of HPC using gene expression programming. Adv. Eng. Softw. 2012, 45, 105–114. [Google Scholar] [CrossRef]
  38. Cheng, M.Y.; Chou, J.S.; Roy, A.F.; Wu, Y.W. High-performance concrete compressive strength prediction using time-weighted evolutionary fuzzy support vector machines inference model. Autom. Constr. 2012, 28, 106–115. [Google Scholar] [CrossRef]
  39. Ashrafian, A.; Shokri, F.; Amiri, M.J.T.; Yaseen, Z.M.; Rezaie-Balf, M. Compressive strength of Foamed Cellular Lightweight Concrete simulation: New development of hybrid artificial intelligence model. Constr. Build. Mater. 2020, 230, 117048. [Google Scholar] [CrossRef]
  40. Behnood, A.; Golafshani, E.M. Predicting the compressive strength of silica fume concrete using hybrid artificial neural network with multi-objective grey wolves. J. Clean. Prod. 2018, 202, 54–64. [Google Scholar] [CrossRef]
  41. Golafshani, E.M.; Behnood, A. Estimating the optimal mix design of silica fume concrete using biogeography-based programming. Cem. Concr. Compos. 2019, 96, 95–105. [Google Scholar] [CrossRef]
  42. Özcan, F.; Atiş, C.D.; Karahan, O.; Uncuoğlu, E.; Tanyildizi, H. Comparison of artificial neural network and fuzzy logic models for prediction of long-term compressive strength of silica fume concrete. Adv. Eng. Softw. 2009, 40, 856–863. [Google Scholar] [CrossRef]
  43. Golafshani, E.M.; Ashour, A. Prediction of self-compacting concrete elastic modulus using two symbolic regression techniques. Autom. Constr. 2016, 64, 7–19. [Google Scholar] [CrossRef] [Green Version]
  44. Behnood, A.; Olek, J.; Glinicki, M.A. Predicting modulus elasticity of recycled aggregate concrete using M5′ model tree algorithm. Constr. Build. Mater. 2015, 94, 137–147. [Google Scholar] [CrossRef]
  45. Sun, W.; Li, Z. Hourly PM2.5 concentration forecasting based on mode decomposition-recombination technique and ensemble learning approach in severe haze episodes of China. J. Clean. Prod. 2020, 263, 121442. [Google Scholar] [CrossRef]
  46. Ahani, I.K.; Salari, M.; Shadman, A. An ensemble multi-step-ahead forecasting system for fine particulate matter in urban areas. J. Clean. Prod. 2020, 263, 120983. [Google Scholar] [CrossRef]
  47. Ahmad, M.R.; Chen, B.; Yu, J. A comprehensive study of basalt fiber reinforced magnesium phosphate cement incorporating ultrafine fly ash. Compos. Part B Eng. 2019, 168, 204–217. [Google Scholar] [CrossRef]
  48. Erdal, H.I.; Karakurt, O.; Namli, E. High performance concrete compressive strength forecasting using ensemble models based on discrete wavelet transform. Eng. Appl. Artif. Intell. 2013, 26, 1246–1254. [Google Scholar] [CrossRef]
  49. Feng, D.C.; Liu, Z.T.; Wang, X.D.; Chen, Y.; Chang, J.Q.; Wei, D.F.; Jiang, Z.M. Machine learning-based compressive strength prediction for concrete: An adaptive boosting approach. Constr. Build. Mater. 2020, 230, 117000. [Google Scholar] [CrossRef]
  50. Han, T.; Siddique, A.; Khayat, K.; Huang, J.; Kumar, A. An ensemble machine learning approach for prediction and optimization of modulus of elasticity of recycled aggregate concrete. Constr. Build. Mater. 2020, 244, 118271. [Google Scholar] [CrossRef]
  51. Zounemat-Kermani, M.; Stephan, D.; Barjenbruch, M.; Hinkelmann, R. Ensemble data mining modeling in corrosion of concrete sewer: A comparative study of network-based (MLPNN & RBFNN) and tree-based (RF, CHAID, & CART) models. Adv. Eng. Inform. 2020, 43, 101030. [Google Scholar] [CrossRef]
  52. Thai, D.K.; Tu, T.M.; Bui, T.Q.; Bui, T.T. Gradient tree boosting machine learning on predicting the failure modes of the RC panels under impact loads. Eng. Comput. 2021, 37, 597–608. [Google Scholar] [CrossRef]
  53. Zhang, J.; Ma, G.; Huang, Y.; Sun, J.; Aslani, F.; Nener, B. Modelling uniaxial compressive strength of lightweight self-compacting concrete using random forest regression. Constr. Build. Mater. 2019, 210, 713–719. [Google Scholar] [CrossRef]
  54. Frank, A.; Asuncion, A. {UCI} Machine Learning Repository. 2010. Available online: https://archive.ics.uci.edu/ml/index.php (accessed on 2 January 2021).
  55. Shaqadan, A. Prediction of concrete mix strength using random forest model. Int. J. Appl. Eng. Res. 2016, 11, 11024–11029. [Google Scholar]
  56. Mining, W.I.D. Data mining: Concepts and techniques. Morgan Kaufinann. 2006, 10, 559–569. [Google Scholar]
  57. Auret, L.; Aldrich, C. Interpretation of nonlinear relationships between process variables by use of random forests. Miner. Eng. 2012, 35, 27–42. [Google Scholar] [CrossRef]
  58. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  59. Svetnik, V.; Liaw, A.; Tong, C.; Wang, T. Application of Breiman’s Random Forest to modeling structure-activity relationships of pharmaceutical molecules. Lect. Notes Comput. Sci. 2004, 3077, 334–343. [Google Scholar] [CrossRef]
  60. Dubeau, P.; King, D.J.; Unbushe, D.G.; Rebelo, L.M. Mapping the Dabus Wetlands, Ethiopia, using random forest classification of Landsat, PALSAR and topographic data. Remote Sens. 2017, 9, 1056. [Google Scholar] [CrossRef] [Green Version]
  61. Krkač, M.; Špoljarić, D.; Bernat, S.; Arbanas, S.M. Method for prediction of landslide movements based on random forests. Landslides 2017, 14, 947–960. [Google Scholar] [CrossRef]
  62. Fu, B.; Wang, Y.; Campbell, A.; Li, Y.; Zhang, B.; Yin, S.; Jin, X. Comparison of object-based and pixel-based Random Forest algorithm for wetland vegetation mapping using high spatial resolution GF-1 and SAR data. Ecol. Indic. 2017, 73, 105–117. [Google Scholar] [CrossRef]
  63. Hanselmann, M.; Köthe, U.; Kirchner, M.; Renard, B.Y.; Amstalden, E.R.; Glunde, K.; Heeren, R.M.A.; Hamprecht, F.A. Toward digital staining using imaging mass spectrometry and random forests. J. Proteome Res. 2009, 8, 3558–3567. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  64. Schwarz, D.F.; König, I.R.; Ziegler, A. On safari to Random Jungle: A fast implementation of Random Forests for high-dimensional data. Bioinformatics 2010, 26, 1752–1758. [Google Scholar] [CrossRef] [Green Version]
  65. Boulesteix, A.L.; Janitza, S.; Kruppa, J.; König, I.R. Overview of random forest methodology and practical guidance with emphasis on computational biology and bioinformatics. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2012, 2, 493–507. [Google Scholar] [CrossRef] [Green Version]
  66. Maghrebi, M.; Waller, T.; Sammut, C. Matching experts’ decisions in concrete delivery dispatching centers by ensemble learning algorithms: Tactical level. Autom. Constr. 2016, 68, 146–155. [Google Scholar] [CrossRef]
  67. Abidoye, L.K.; Mahdi, F.M.; Idris, M.O.; Alabi, O.O.; Wahab, A.A. ANN-derived equation and ITS application in the prediction of dielectric properties of pure and impure CO2. J. Clean. Prod. 2018, 175, 123–132. [Google Scholar] [CrossRef]
  68. Oliveira, V.; Sousa, V.; Dias-Ferreira, C. Artificial neural network modelling of the amount of separately-collected household packaging waste. J. Clean. Prod. 2019, 210, 401–409. [Google Scholar] [CrossRef]
  69. Hoque, M.M.; Rahman, M.T.U. Landfill area estimation based on solid waste collection prediction using ANN model and final waste disposal options. J. Clean. Prod. 2020, 256, 120387. [Google Scholar] [CrossRef]
  70. Xu, L.; Huang, C.; Li, C.; Wang, J.; Liu, H.; Wang, X. A novel intelligent reasoning system to estimate energy consumption and optimize cutting parameters toward sustainable machining. J. Clean. Prod. 2020, 261, 121160. [Google Scholar] [CrossRef]
  71. Ozoegwu, C.G. Artificial neural network forecast of monthly mean daily global solar radiation of selected locations based on time series and month number. J. Clean. Prod. 2019, 216, 1–13. [Google Scholar] [CrossRef]
  72. Hafeez, A.; Taqvi, S.A.A.; Fazal, T.; Javed, F.; Khan, Z.; Amjad, U.S.; Bokhari, A.; Shehzad, N.; Rashid, N.; Rehman, S.; et al. Optimization on cleaner intensification of ozone production using Artificial Neural Network and Response Surface Methodology: Parametric and comparative study. J. Clean. Prod. 2020, 252, 119833. [Google Scholar] [CrossRef]
  73. Ahmad, A.; Farooq, F.; Ostrowski, K.A.; Śliwa-Wieczorek, K.; Czarnecki, S. Application of novel machine learning techniques for predicting the surface chloride concentration in concrete containing waste material. Materials 2021, 14, 2297. [Google Scholar] [CrossRef]
  74. Asteris, P.G.; Mokos, V.G. Concrete compressive strength using artificial neural networks. Neural Comput. Appl. 2020, 32, 11807–11826. [Google Scholar] [CrossRef]
  75. Czarnecki, S.; Sadowski, Ł.; Hoła, J. Artificial neural networks for non-destructive identification of the interlayer bonding between repair overlay and concrete substrate. Adv. Eng. Softw. 2020, 141, 102769. [Google Scholar] [CrossRef]
  76. Zhou, G.; Moayedi, H.; Bahiraei, M.; Lyu, Z. Employing artificial bee colony and particle swarm techniques for optimizing a neural network in prediction of heating and cooling loads of residential buildings. J. Clean. Prod. 2020, 254, 120082. [Google Scholar] [CrossRef]
  77. Shahin, M.A.; Maier, H.R.; Jaksa, M.B. Data division for developing neural networks applied to geotechnical engineering. J. Comput. Civ. Eng. 2004, 18, 105–114. [Google Scholar] [CrossRef]
  78. Braga-Neto, U. Fundamentals of Pattern Recognition and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  79. Keprate, A.; Ratnayake, R.C. Using gradient boosting regressor to predict stress intensity factor of a crack propagating in small bore piping. In Proceedings of the 2017 IEEE International Conference on Industrial Engineering and Engineering Management (IEEM), Singapore, 10–13 December 2017. [Google Scholar]
  80. Li, C. A Gentle Introduction to Gradient Boosting. 2016. Available online: http://www.ccs.neu.edu/home/vip/teach/MLcourse/4_boosting/slides/gradient_boosting.pdf (accessed on 20 December 2021).
  81. Kohavi, R. A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection. Int. Jt. Conf. Artif. Intell. 1995, 14, 1137–1145. Available online: http://robotics.stanford.edu/~ronnyk (accessed on 2 January 2021).
  82. Song, H.; Ahmad, A.; Farooq, F.; Ostrowski, K.A.; Maślak, M.; Czarnecki, S.; Aslam, F. Predicting the compressive strength of concrete with fly ash admixture using machine learning algorithms. Constr. Build. Mater. 2021, 308, 125021. [Google Scholar] [CrossRef]
  83. Farooq, F.; Amin, M.N.; Khan, K.; Sadiq, M.R.; Javed, M.F.; Aslam, F.; Alyousef, R. A comparative study of random forest and genetic engineering programming for the prediction of compressive strength of high strength concrete (HSC). Appl. Sci. 2020, 10, 7330. [Google Scholar] [CrossRef]
Figure 1. Variable relationship with compressive strength.
Figure 1. Variable relationship with compressive strength.
Applsci 12 00361 g001
Figure 2. Relative frequency distribution of variables.
Figure 2. Relative frequency distribution of variables.
Applsci 12 00361 g002
Figure 3. Contour plots indicating the concentrations used in analyzing compressive strength.
Figure 3. Contour plots indicating the concentrations used in analyzing compressive strength.
Applsci 12 00361 g003
Figure 4. Flowchart of the methodology of machine learning.
Figure 4. Flowchart of the methodology of machine learning.
Applsci 12 00361 g004
Figure 5. Execution process of ANN model [73].
Figure 5. Execution process of ANN model [73].
Applsci 12 00361 g005
Figure 6. Indication of the neural model.
Figure 6. Indication of the neural model.
Applsci 12 00361 g006
Figure 7. Results of ensemble ML models with their sub-models. (a) DT-Bagging; (b) DT-Adaboost; (c) MLPNN-Bagging; (d) MLPNN-Boosting.
Figure 7. Results of ensemble ML models with their sub-models. (a) DT-Bagging; (b) DT-Adaboost; (c) MLPNN-Bagging; (d) MLPNN-Boosting.
Applsci 12 00361 g007
Figure 8. Indication of K-fold cross-validation approach.
Figure 8. Indication of K-fold cross-validation approach.
Applsci 12 00361 g008
Figure 9. (a) DT prediction vs target; (b) DT error distribution; (c) actual vs prediction of DT-bagging; (d) DT-bagging error distribution; (e) actual vs forecasted results for DT-boosting; (f) error distribution of DT-boosting.
Figure 9. (a) DT prediction vs target; (b) DT error distribution; (c) actual vs prediction of DT-bagging; (d) DT-bagging error distribution; (e) actual vs forecasted results for DT-boosting; (f) error distribution of DT-boosting.
Applsci 12 00361 g009
Figure 10. (a) Forecasted vs target for MLPNN; (b) MLPNN error distribution; (c) forecasted vs actual for MLPNN-bagging; (d) MLPNN-bagging error distribution; (e) forecasted vs target MLPNN- boosting; (f) error distribution of MLPNN-boosting.
Figure 10. (a) Forecasted vs target for MLPNN; (b) MLPNN error distribution; (c) forecasted vs actual for MLPNN-bagging; (d) MLPNN-bagging error distribution; (e) forecasted vs target MLPNN- boosting; (f) error distribution of MLPNN-boosting.
Applsci 12 00361 g010
Figure 11. (a) Predicted data through RF; (b) Error distribution of actual and predictions.
Figure 11. (a) Predicted data through RF; (b) Error distribution of actual and predictions.
Applsci 12 00361 g011
Figure 12. (a) Prediction vs targets of GB; (b) Distribution of errors of GB model.
Figure 12. (a) Prediction vs targets of GB; (b) Distribution of errors of GB model.
Applsci 12 00361 g012
Figure 13. Result of validation; (a,b) DT-bagging and DT-boosting; (c,d) MLPNN-bagging and MLPNN-boosting; (e) random forest; (f) gradient boosting.
Figure 13. Result of validation; (a,b) DT-bagging and DT-boosting; (c,d) MLPNN-bagging and MLPNN-boosting; (e) random forest; (f) gradient boosting.
Applsci 12 00361 g013
Table 1. Literature-based application of ML approaches.
Table 1. Literature-based application of ML approaches.
Concrete TypePropertiesTechniquesReferences
Normal concreteCompressive strengthGenetic programming[30]
ANN[31]
High-performance concreteCompressive strengthRandom forest[32]
ANN[33,34,35]
M5P[36]
GEP[37]
Genetic weighted pyramid operation tree[38]
Foamed cellular lightweight concreteCompressive strengthANN[39]
Silica fume concreteCompressive strengthHybrid ANN[40]
Biogeography-based programming (BBP)[41]
ANN and ANFIS[42]
Self-compacting concreteModulus of ElasticityBiogeography-based programming (BBP)[43]
Recycled aggregate concreteModulus of ElasticityM5P[44]
Concrete-filled steel tubeCompressive strengthGEP[18]
Table 2. Ensemble ML algorithms used in the literature.
Table 2. Ensemble ML algorithms used in the literature.
Concrete TypePropertiesTechniquesReferences
High-performance concreteCompressive strengthBANN[48]
GBANN
Adaptive boosting[49]
RF[32]
Gradient tree boosting[15]
Recycled aggregate concreteModulus of ElasticityRF + SVM[50]
Corrosion of concrete sewerMicrobially induced concrete corrosionBagging/BoostingMLPNN/RBFNN/CHAID/CART[51]
Corrosion of concrete sewerMicrobially induced concrete corrosionEnsemble RF[51]
High-performance concreteCompressive strengthAdaptive boosting
RC panelsFailure modesGBML[52]
Lightweight self-compacting concreteCompressive strengthRF[53]
Table 3. Description of input variable ranges.
Table 3. Description of input variable ranges.
Variables UsedAbbreviation UsedMinimum
Value
Maximum
Value
Input variables
BinderCEM134.7540
Fine aggregate
Coarse aggregate
FA
CA
594
801
945
1125
Fly ashFA0200.1
WaterW140228
Superplasticizer
Age
SP
AG
0
1
28.2
365
Output variable
Compressive strengthFc6.2779.99
Table 4. Statistical indication of parameters.
Table 4. Statistical indication of parameters.
StatisticsCementFly AshWaterSuperplasticizerCoarse AggregateFine AggregateDaysStrength
Count471471471471471471471471
Mean298.0862.59181.885.021004.04793.1747.8431.60
Std100.6964.8818.015.4974.1773.8665.5314.74
Min134.70140080159416.27
25%229.70167.550961.2758.31419.73
50%281901864.610067922831.35
75%349118.31929.510568505640.87
Max540200.122828.2112594536579.99
Table 5. Detailed information of the sub-models of ensemble approaches.
Table 5. Detailed information of the sub-models of ensemble approaches.
Approaches
Used
Ensemble
Techniques
Machine Learning MethodsEnsemble ModelsOptimum EstimatorR-Value
Individual-DT--0.80
-MLPNN--0.77
EnsembleBaggingDecision tree- bagging(10, 20, 30…200)200.87
Multilayer perceptron neuron network- bagging(10, 20, 30…200)200.83
EnsembleBoostingDecision tree- Adaboost(10, 20, 30…200)1300.89
Multilayer perceptron neuron network- Adaboost(10, 20, 30…200)400.80
Modified learner Random forest(10, 20, 30…200)1300.89
Boosting regressor Gradient boosting regressor(10, 20,…30.200)500.84
Table 6. Statistical evaluation of an individual with ensemble DT model.
Table 6. Statistical evaluation of an individual with ensemble DT model.
AnalysisDTDT-BaggingDT-AdaBoost
Average EDT4.413.542.96
Minimum EDT0.0360.0060.057
Maximum EDT22.8419.8220.53
Results lies below 10 MPa849090
Results between 10 MPa and 15 MPa831
Results between 15 MPa and 20 MPa111
Results between 20 MPa and 25 MPa100
Total data in testing data949494
Average below 10 MPa89.3695.7495.74
Results between 10 MPa and 15 MPa8.513.193.19
Results between 15 MPa and 20 MPa1.061.061.06
Results between 20 MPa and 25 MPa1.0600
Table 7. Statistical results of the individual with ensemble MLPNN model.
Table 7. Statistical results of the individual with ensemble MLPNN model.
AnalysisMLPNNMLPNN-BaggingMLPNN-Adaboost
Average of EMLPNN5.243.944.42
Minimum of EMLPNN0.090.050.02
Maximum of EMLPNN19.5722.3022.77
Result lies below 10 MPa768789
Result lies between 10 MPa and 15 MPa1452
Result lies between 15 MPa and 20 MPa411
Result lies between 20 MPa and 25 MPa012
Total data in testing data949494
Average result below 10 MPa80.8592.5594.68
Average result between 10 MPa and 15 MPa14.895.312.11
Average result between 15 MPa and 20 MPa4.251.0647.34
Average result between 20 MPa and 25 MPa01.064.22
Table 8. Statistical analysis of RF.
Table 8. Statistical analysis of RF.
Statistical AnalysisRF
Average EMLPNN2.89
Minimum EMLPNN0.06
Maximum EMLPNN20.39
Entries lies below 10 MPa91
Entries lies between 10 MPa and 15 MPa2
Entries lies between 15 MPa and 20 MPa1
Entries lies between 20 MPa and 25 MPa0
Total data in testing data94
Average below 10 MPa96.80
Average between 10 MPa and 15 MPa2.12
Average between 15 MPa and 20 MPa1.06
Average between 20 MPa and 25 MPa0
Table 9. Statistical analysis of GB.
Table 9. Statistical analysis of GB.
Statistical AnalysisGB
Average EMLPNN3.59
Minimum EMLPNN0.00
Maximum EMLPNN23.97
Entries lies below 10 MPa88
Entries lies between 10 MPa and 15 MPa4
Entries lies between 15 MPa and 20 MPa1
Entries lies between 20 MPa and 25 MPa1
Total data in testing data94
Average below 10 MPa93.61
Average between 10 MPa and 15 MPa4.25
Average between 15 MPa and 20 MPa1.06
Average between 20 MPa and 25 MPa1.06
Table 10. Result of K-Fold cross-validation for ensemble models.
Table 10. Result of K-Fold cross-validation for ensemble models.
FoldsDT-BaggingDT-Boosting
MAERMSLERMSEMAERMSLERMSE
17.8120.1738.8227.2590.1598.665
27.1560.1347.9508.5020.1179.150
33.9700.0335.7464.8540.04297.342
45.3160.0758.3063.6200.04224.943
54.8500.0587.4414.8650.05195.970
65.3080.0797.2894.7610.04136.305
72.9240.0344.9062.6670.03283.946
89.3070.10311.0956.8000.05268.433
97.6840.12411.7595.7610.08788.926
106.6220.0859.4586.4950.06598.018
Table 11. K-Fold cross-validation result of ensemble models.
Table 11. K-Fold cross-validation result of ensemble models.
FoldsMLPNN-BaggingMLPNN-Boosting
MAERMSLERMSEMAERMSLERMSE
15.2490.1258.2005.2550.1096.294
27.3780.1178.2006.3190.0918.234
35.7210.0706.5986.5370.0598.834
43.8880.0775.8323.8720.0435.124
57.7940.07411.8177.0750.0778.093
64.4120.0425.8404.450.0316.428
72.6930.0192.9402.2420.0463.358
Table 12. K-Fold cross-validation result of ensemble models.
Table 12. K-Fold cross-validation result of ensemble models.
FoldsRF-BaggingGBR-Boosting
MAERMSLERMSEMAERMSLERMSE
16.1360.1518.1977.6280.1269.229
27.2740.1178.1588.6140.13710.723
31.5110.0102.0326.0080.0729.055
44.4920.0385.1967.2170.14214.321
54.4830.0565.8075.5000.0676.347
64.7690.0545.5555.4970.0566.783
72.6880.0314.1755.7990.0817.901
86.4200.0518.6285.7740.05410.006
96.0590.0949.9907.4320.1319.404
106.0160.0537.7195.2150.0446.912
Table 13. Statistical evaluation.
Table 13. Statistical evaluation.
Approaches UseML MethodsMAEMSERMSEMSLE
Individual learnerDecision tree5.4055.707.460.052
Multilayer perceptron neuron network4.5737.346.110.049
Decision tree-bagging4.1934.515.870.034
Ensemble learning baggingMultilayer perceptron neuron network- bagging4.4133.495.780.043
Decision tree-Adaboost3.5324.284.920.029
Ensemble learning boostingMultilayer perceptron neuron network- Adaboost4.3939.296.260.045
Modified EnsembleRandom forest3.2622.264.710.026
Boosting ensembleGradient boosting 4.1133.605.790.042
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Song, Y.; Zhao, J.; Ostrowski, K.A.; Javed, M.F.; Ahmad, A.; Khan, M.I.; Aslam, F.; Kinasz, R. Prediction of Compressive Strength of Fly-Ash-Based Concrete Using Ensemble and Non-Ensemble Supervised Machine-Learning Approaches. Appl. Sci. 2022, 12, 361. https://doi.org/10.3390/app12010361

AMA Style

Song Y, Zhao J, Ostrowski KA, Javed MF, Ahmad A, Khan MI, Aslam F, Kinasz R. Prediction of Compressive Strength of Fly-Ash-Based Concrete Using Ensemble and Non-Ensemble Supervised Machine-Learning Approaches. Applied Sciences. 2022; 12(1):361. https://doi.org/10.3390/app12010361

Chicago/Turabian Style

Song, Yang, Jun Zhao, Krzysztof Adam Ostrowski, Muhammad Faisal Javed, Ayaz Ahmad, Muhammad Ijaz Khan, Fahid Aslam, and Roman Kinasz. 2022. "Prediction of Compressive Strength of Fly-Ash-Based Concrete Using Ensemble and Non-Ensemble Supervised Machine-Learning Approaches" Applied Sciences 12, no. 1: 361. https://doi.org/10.3390/app12010361

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop