Literature DB >> 35335405

Application of Soft Computing Techniques to Predict the Strength of Geopolymer Composites.

Qichen Wang1, Waqas Ahmad2, Ayaz Ahmad2,3, Fahid Aslam4, Abdullah Mohamed5, Nikolai Ivanovich Vatin6.   

Abstract

Geopolymers may be the best alternative to ordinary Portland cement because they are manufactured using waste materials enriched in aluminosilicate. Research on geopolymer composites is accelerating. However, considerable work, expense, and time are needed to cast, cure, and test specimens. The application of computational methods to the stated objective is critical for speedy and cost-effective research. In this study, supervised machine learning approaches were employed to predict the compressive strength of geopolymer composites. One individual machine learning approach, decision tree, and two ensembled machine learning approaches, AdaBoost and random forest, were used. The coefficient correlation (R2), statistical tests, and k-fold analysis were used to determine the validity and comparison of all models. It was discovered that ensembled machine learning techniques outperformed individual machine learning techniques in forecasting the compressive strength of geopolymer composites. However, the outcomes of the individual machine learning model were also within the acceptable limit. R2 values of 0.90, 0.90, and 0.83 were obtained for AdaBoost, random forest, and decision models, respectively. The models' decreased error values, such as mean absolute error, mean absolute percentage error, and root-mean-square errors, further confirmed the ensembled machine learning techniques' increased precision. Machine learning approaches will aid the building industry by providing quick and cost-effective methods for evaluating material properties.

Entities:  

Keywords:  artificial intelligence; compressive strength; geopolymer composites; machine learning; prediction models; sustainable materials

Year:  2022        PMID: 35335405      PMCID: PMC8956037          DOI: 10.3390/polym14061074

Source DB:  PubMed          Journal:  Polymers (Basel)        ISSN: 2073-4360            Impact factor:   4.329


1. Introduction

Cement-based conventional concrete (CBCC) is the most broadly utilized type of construction material on a global scale [1,2,3]. The primary constituents of CBCC are aggregates, water, and ordinary Portland cement (OPC) [4,5]. Following aluminum and steel, OPC is the third most energy-demanding substance on the earth, consuming 7% of the total energy of global industry [6,7]. Regrettably, the manufacture of OPC produces large quantities of greenhouse gases, i.e., CO2, which substantially add to climate change [8,9,10]. The production of OPC is anticipated to release 1.35 billion tons of greenhouse emissions annually [11,12,13]. Thus, scholars have focused their attempts on minimizing OPC usage through the use of alternate binder types. Alternatives to CBCC may include alkali-activated compounds such as geopolymers [14,15,16]. When precursors and activators react, alkali-activated compounds are formed. They have been categorized into two kinds based on the calcium proportion of the products formed during the reaction: those that are calcium-rich, with a Ca/(Si+Al) fraction above 1, and those that are calcium-deficient, i.e., geopolymers [17,18,19]. Geopolymer is a novel binder that was established to substitute OPC in concrete production [20,21,22]. The purpose is to acquire a building material that is free of OPC, environmentally caring, and sustainable. As industry and people grow, a significant amount of waste material (waste glass powder, fly ash, sugarcane bagasse ash, ground granulated blast furnace slag (GGBS), silica fume, and rice husk ash, for example) is generated and disposed of in landfills. Due to the fact that these waste products contribute to pollution, their disposal in landfills is dangerous to the ecosystem [23,24,25,26]. Since geopolymer composites (GPCs) require raw ingredients with a high aluminosilicate content, which are found in these waste materials, recycling these waste materials can help to reduce environmental pollution [27,28,29,30]. The method of producing GPC is represented in Figure 1, along with the various components and curing regimes employed. Consumption of these waste materials benefits both the environment and the economy, as the demand for inexpensive housing will increase as the population expands [31,32,33]. GPC has been the topic of broad research and development on a global scale, and it may one day become the best green construction material [34,35,36,37]. GPC, on the other hand, has the potential to make a significant contribution to the prolonged sustainability of both CBCC technology and the building sector.
Figure 1

Production process of geopolymer concrete.

Artificial intelligence (AI) advancements have resulted in the widespread usage of machine learning (ML) techniques for anticipating the properties of a variety of materials [38,39,40,41]. Ahmad et al. [14] conducted a comparative investigation of three ML approaches for assessing the compressive strength (CS) of fly ash-based GPC, including decision tree (DT), AdaBoost, and bagging regressor (BR). It was noted that the BR model was the most precise of the models examined. Ahmad et al. [42] forecasted the CS of concrete, including recycled coarse aggregate using an artificial neural network (ANN) and gene expression programming (GEP). The GEP technique was found to be more predictively exact than the ANN technique. Song et al. [43] used an ANN method to estimate the CS of concrete, and they observed a satisfactory forecasting ability. Nguyen et al. [44] predicted the CS and tensile strength of high-performance concrete using a range of ML approaches. They concluded that ensembled ML techniques outperformed individual ML techniques in terms of precision. Thus, numerous scientists have reported on diverse ML strategies that improve the accuracy of material property estimation. As a result, it is critical to conduct additional in-depth investigations to elucidate this topic. This research concentrates on the use of ML approaches to estimate the CS of GPCs. Three different types of ML approaches were used, including decision tree (DT), AdaBoost, and random forest (RF), and their performance was assessed using statistical tests and correlation coefficient (R2). Additionally, the validity of each strategy was determined using k-fold analysis and error distributions. DT is an individual ML technique, whereas AdaBoost and RF are ensemble ML algorithms. This research is novel in that it estimates the CS of GPC using both individual and ensembled ML techniques, whereas experimental investigations need significant human work, experimental costs, and time for material acquisition, casting, curing, and testing. The use of modern techniques, like ML, in the field of civil engineering to predict material properties will reduce human effort and save time since experimental work for said purpose can be eliminated. ML approaches require a data set that might be retrieved from the literature as considerable research has been conducted to experiment with the material properties, and the data set can be used to train the ML models and estimate the various characteristics of a material. This study aims to identify the most suitable ML technique for the CS of GPCs in terms of results prediction and the influence of input parameters on the model’s performance.

2. Methods

2.1. Description of Data

To obtain an appropriate result, SML algorithms require a varied range of input variables [45,46,47]. The CS of GPC was estimated using data retrieved from the literature (attached as a Supplementary File). To prevent bias representation, experimental data were randomly chosen from the literature. The literature published on the use of comparable materials for the CS of GPC was assessed. While most papers examined additional features of GPC, this study acquired CS-based data points to run the algorithms. Nine variables were included as inputs in the algorithms, containing water/solids ratio, NaOH molarity, gravel 4/10 mm, gravel 10/20 mm, NaOH, Na2SiO3, fly ash, GGBS, and fine aggregate, with CS as the output variable. The number of inputs and datasets have a considerable effect on the model’s outcome [48,49,50]. In the current investigation, 363 data points were used to run ML algorithms. Table 1 summarizes the descriptive statistic evaluation of each input variable. The term “descriptive statistics” refers to a group of concise, factual measurements that generate an outcome, which may be the whole population or a subset of the population. The mean, median, and mode variables represent fundamental tendency, whereas the maximum, minimum, and standard deviation represent variability. The table provides all the mathematical terms for the model’s input variables. The relative frequency distribution of all variables used in the analysis is depicted in Figure 2. It depicts the total number of interpretations related to each value or combination of values. It is intrinsically related to probability dispersal, a widely used statistical term.
Table 1

Descriptive measurements of input variables employed.

ParameterWater/Solids RatioNaOH MolarityGravel 4/10 mm (kg/m3)Gravel 10/20 mm (kg/m3)NaOH (kg/m3)Na2SiO3 (kg/m3)Fly Ash (kg/m3)GGBS (kg/m3)Fine Aggregate (kg/m3)
Minimum01003.51800459
Maximum0.63201293.412981473425234501360
Range0.63191293.41298143.5324523450901
Median0.349.220878956108120300728
Mode0.5310006410800651
Mean0.348.14288.39737.3753.74111.66174.34225.15729.88
Standard Error0.010.2419.5418.821.672.538.828.526.87
Standard Deviation0.114.56372.31358.5531.9148.16167.95162.27130.97
Sum124.82955.1104,684.3267,664.919,508.840,532.763,286.081,728.1264,947.8
Figure 2

Relative frequency dispersal of inputs: (a) Water/solids ratio; (b) NaOH Molarity; (c) Gravel 4/10 mm; (d) Gravel 10/20 mm; (e) NaOH; (f) Na2SiO3; (g) Fly ash; (h) GGBS; (i) Fine aggregate.

2.2. Machine Learning Algorithms Employed

Individual ML approaches (DT) and ensemble ML techniques (AdaBoost and RF) were used to accomplish the study’s objectives, with Python scripting via the Anaconda Navigator package. To run DT, AdaBoost, and RF models, the tool Spyder (version 4.3.5) was chosen. These algorithms are often used to anticipate desired outcomes based on input variables. These algorithms, among other things, are capable of forecasting the temperature effect, strength properties, and durability of materials [51,52]. Throughout the modeling phase, nine input variables and one output variable (i.e., CS) were employed. The R2 value for the projected result reflected the validity/precision of all models. The R2 represents the extent of divergence; a value close to zero suggests higher divergence, but a value close to one shows that the model and data are almost completely suited [14]. The sub-sections below detail the ML techniques applied in this research. Additionally, to validate models, statistical and k-fold analysis and error assessments are carried out on all techniques, involving root-mean-square error (RMSE), mean absolute percentage error (MAPE) and mean absolute error (MAE). Moreover, sensitivity analysis is employed to discover the effect of each input parameter on the outcome estimation. The flowchart in Figure 3 depicts the research strategy.
Figure 3

Flowchart of research methodology.

2.2.1. Decision Tree

DTs are created by the development of algorithms that divide a dataset into branch-like portions. These portions combine to create an upturned tree, which begins with a root node at the top [53]. Figure 4 demonstrates an illustration of such a tree with five nodes and six leaves. As seen in the figure, a DT tree can contain both uninterrupted and isolated features. Correlations between the object of analysis and the input fields are used to produce the decision rule for branching or segmenting underneath the root node. After establishing the link, one or more decision rules specifying the relationships between the inputs and targets can be produced. Decision rules reliably estimate the values of new or unknown observations that include input values but not targets. The errors are computed at each division point, and the variable with the lowest fitness function value is chosen as a split point, followed by the procedure for the other variables.
Figure 4

Decision tree schematic representation [14]. Reprinted with permission from ref. [14]. Copyright 2022 Elsevier B.V.

2.2.2. AdaBoost

The AdaBoost approach is the most often used ensembled ML algorithm from the boosting group of ensembled ML techniques. AdaBoost’s distinguishing characteristic is that it utilizes the initial training data to construct a weak learner, after which it modifies its distribution of training data depending on its projection performance in the following turn of weak learner training. Remember that the training samples with less forecast precision from the previous stage will be given more attention in the following phase. After that, the weak learners are then coupled with a strong learner by applying a variety of weights to form a final combination [39]. AdaBoost is simple to implement. In general, it consists of four stages: (i) data collection; (ii) development of a strong learner; (iii) testing or confirmation of the learner; and (iv) use of the learner for engineering challenges. Clearly, the second step is important to the AdaBoost algorithm. As mentioned previously, it consists of two components: a framework for integrating weak learners into a strong one and a regression learning algorithm for producing the weak learner from the training data. The weak learner is generated using the decision tree (DT) algorithm [39], and the weak learners are combined using the median of the weighted weak learners. Figure 5 illustrates the flow diagram for this technique.
Figure 5

Schematic representation of AdaBoost technique.

2.2.3. Random Forest

RF is implemented via the random split selection on bagging DTs [54]. Figure 6 illustrates the production and process of the RF model schematically. Every tree in the forest is constructed using a randomly chosen training set, and each split within each tree is built using a randomly selected subset of input variables, resulting in a forest of trees [55]. The addition of this unpredictability boosts the tree’s diversity. The forest is entirely composed of fully-grown binary trees. The RF method has been exceedingly effective as a common-purpose classification and regression tool. The technique, which combines the predictions of several randomized DTs, has shown higher precision in circumstances where the quantity of variables exceeds the quantity of observations. Moreover, it is adaptive to both large-scale and ad hoc learning tasks, returning metrics of different importance [54].
Figure 6

Schematic representation of random forest algorithm [54]. Reprinted with permission from ref. [54]. Copyright 2019 Elsevier Ltd.

3. Models Results

3.1. Decision Tree Model

Figure 7 illustrates the results of the DT model for the CS of GPC. Figure 7a shows the link between experimental and projected outcomes. The DT technique generated results with a satisfactory level of accuracy and a small disparity between experimental and predicted outcomes. The R2 of 0.83 confirms the satisfactory performance of the DT model in forecasting the CS of GPC. The dispersion of predicted and error values for the DT model is represented in Figure 7b. The error values were analyzed, and it was determined that the minimum, average, and highest values were 0.00, 7.02, and 36.59 MPa, respectively. Additionally, the percentage distribution of error values was determined, and it was discovered that 37.4% of values were less than 3 MPa, 38.5% were between 3 and 10 MPa, and only 24.2% were above 10 MPa. Additionally, the dispersion of error values implies that the DT model performs satisfactorily.
Figure 7

Decision tree model: (a) correlation between experimental and estimated results; (b) dispersal of predicted and error values.

3.2. AdaBoost Model

Figure 8 depicts the AdaBoost model’s findings for estimating the CS of GPC. The relationship between experimental and anticipated outcomes is depicted in Figure 8a. The AdaBoost approach produced output with higher precision and the least amount of divergence between actual and anticipated outcomes. With an R2 of 0.90, the AdaBoost model is quite accurate at forecasting the CS of GPC. Figure 8b illustrates the dispersion of predicted and error values for the AdaBoost model. The training set’s lowest, average, and maximum error values were determined to be 0.00, 5.20, and 20.40 MPa, respectively. The error distribution was 46.2%less than 3 MPa, 35.2% between 3 and 10 MPa, and only 18.7% greater than 10 MPa. The distribution of error values indicates the AdaBoost model’s higher precision in forecasting outcomes.
Figure 8

AdaBoost model: (a) correlation between experimental and estimated results; (b) dispersal of predicted and error values.

3.3. Random Forest Model

Figure 9a,b demonstrate an assessment of the RF model’s experimental and estimated results. Figure 9a depicts the link between experimental and estimated findings, with an R2 of 0.90 signifying that the RF model has a comparable precision to the AdaBoost model in estimating the GPCs CS. Figure 9b shows the dispersal of experimental, expected, and error values for the RF model. The lowest, average, and highest error values were determined to be 0.06, 5.33, and 23.45 MPa, respectively. The error distribution was 47.3% less than 3 MPa, 34.1% between 3 and 10 MPa, and only 18.7% larger than 10 MPa. These reduced error values demonstrate the RF model’s higher exactness than the DT model and similar accuracy to the AdaBoost model.
Figure 9

Random forest model: (a) correlation between experimental and estimated results; (b) dispersal of predicted and error values.

4. Validation of Models

Statistical and k-fold analysis approaches were used to validate the models. The k-fold technique is frequently used to ascertain a technique’s validity [42], during which relevant data are arbitrarily scattered and divided into 10 classes. As seen in Figure 10, nine groups will be used to train the model, while one group will be utilized to validate it. Approximately 75% of the data was utilized for training the models, whereas 25% was utilized to assess the models that were employed. When the errors (MAE and RMSE) are low and the R2 value is high, the model is more accurate. In addition, the operation ought to be reiterated ten times to achieve a reasonable conclusion. This extensive effort contributes significantly to the model’s remarkable accuracy. Furthermore, as shown in Table 2, all models were statistically evaluated in terms of errors (MAE, MAPE, and RMSE). These assessments also confirmed the AdaBoost and RF model’s higher accuracy as a result of their lower error readings when compared to the DT model. The predictive performance of the techniques was determined statistically using Equations (1)–(3), which were acquired from previous studies [38,56,57]. where = sum of data samples, = predicted values, and = experimental values from the data set.
Figure 10

K-fold cross-validation procedure.

Table 2

Statistical assessments of the models employed in this study.

ModelMAE (MPa)MAPE (%)RMSE (MPa)
Decision tree7.01616.02010.432
AdaBoost5.19912.3027.467
Random forest5.32512.4207.602
In order to figure out how well the k-fold cross-validation worked, MAE, MAPE, RMSE, and R2 were calculated, and their values are provided in Table 3. The DT model’s MAE values ranged from 7.02 to 19.78 MPa, with an average of 11.08 MPa. When comparing, the MAE values for the AdaBoost model ranged between 5.20 and 14.68 MPa, with an average of 8.68 MPa. As for the AdaBoost model, the MAE values were between 5.33 and 18.47 MPa, with an average of 8.97 MPa. Similarly, the average MAPE for DT, AdaBoost, and RF models was noted to be 17.04%, 13.16%, and 13.47%, respectively. The average RMSE values for the DT, AdaBoost, and RF models were 15.89, 11.18, and 11.94 MPa. On the other hand, the average R2 values for DT, AdaBoost, and RF models were 0.59, 0.67, and 0.65, respectively. The AdaBoost and RF models with the lower error values and the higher R2 values are more accurate in forecasting the CS of GPC when compared to the DT model.
Table 3

Outcomes of k-fold analysis.

K-FoldDecision TreeAdaBoostRandom Forest
MAEMAPERMSER2MAEMAPERMSER2MAEMAPERMSER2
116.0318.7021.920.608.7013.2513.040.4310.7012.9713.430.54
27.0216.9311.570.765.6512.308.010.495.3313.768.090.72
39.1516.0310.940.206.5614.038.160.795.5414.888.370.64
411.7617.2110.430.708.1812.558.430.678.0613.6611.400.52
57.3116.0212.410.596.1112.987.470.905.3412.907.850.77
612.9616.5517.070.3712.9414.4514.340.579.8513.7713.820.53
77.7218.6719.580.729.5013.6612.060.609.4312.4215.560.74
810.9216.0315.260.419.3313.0814.330.8611.1214.0213.930.34
98.1517.2216.500.725.2012.957.680.745.8013.797.600.79
1019.7817.0223.230.8314.6812.3518.280.6118.4712.5019.310.90

5. Sensitivity Analysis

The intention of this assessment is to ascertain the effect of input parameters on GPC’s CS predicting. The expected outcome is greatly affected by the input variables [14]. Figure 11 demonstrates the impact of the inputs on the CS estimate of GPC. The investigation determined that fly ash was the most important constituent, accounting for 26.37% of the total, followed by GGBS at 14.74% and NaOH molarity at 13.12%. The remaining input variables, on the other hand, contributed less to the forecast of GPC’s CS, with NaOH accounting for 11.60%, the water/solids ratio accounting for 9.52%, fine aggregate accounting for 7.53%, gravel 4/10 mm accounting for 6.48%, gravel 10/20 mm accounting for 5.84%, and Na2SiO3 accounting for 4.80%. Sensitivity analysis generated outcomes related to the number of input parameters and data points employed to construct the models. The influence of an input parameter on the technique’s output was determined using Equations (4) and (5). where and are the peak and bottom of the expected result on the output, respectively, whereas other input variables are maintained constant at their mean values. is the achieved contribution proportion for a particular variable.
Figure 11

Input variables’ contributions to predicting outcomes.

6. Discussions

6.1. Comparison of Machine Learning Models

The objective of this study was to contribute to the existing study area regarding the implementation of contemporary approaches for estimating the CS of GPC. This type of research will aid the building industry by developing rapid and cost-effective solutions for material property prediction. Additionally, by utilizing these strategies to promote eco-friendly construction, the adoption and use of GPC in construction will be accelerated. Since GPC may be made from waste materials containing aluminosilicates, its use in construction will have a number of advantages, as seen in Figure 12. This study demonstrates how ML methods can be employed to anticipate the CS of GPC. Three ML techniques were used in the study: one individual (DT) and two ensembled (AdaBoost and RF). Each technique was examined for accuracy in order to discover which is the most efficient predictor. The AdaBoost and RF models produced more exact results with an R2 of 0.90, compared to the DT model, which yielded an R2 of 0.83.
Figure 12

Advantages of geopolymer composites produced with waste materials.

Furthermore, all models’ accuracy was validated using the statistical k-fold analysis approach. The fewer error values in the model, the more precise it is. The higher accuracy of AdaBoost and RF models towards the prediction of outcomes is also reported by other researchers [39,58,59]. Feng et al. [39] noticed the superior performance of the AdaBoost model compared to individual models, including ANN and support vector machine (SVM), based on higher R2 and lower error values. Similarly, Farooq et al. [59] compared the performance of RF with ANN, GEP, and DT techniques and reported the higher precision of the RF model than the others with an R2 of 0.96. However, determining and recommending the optimal ML model for forecasting outcomes through a variety of areas is complicated, as the performance of a model is greatly reliant on the input parameters and quantity of data points utilized to execute the algorithm. The previous studies concluded that up to 300 data points and a minimum of 8 input variables could result in the higher precision of the ML models [56,60]. Hence, the data set retrieved for the current investigation is suitable for the ML model’s best performance. The ensembled ML algorithms commonly exploit the weak learner by generating sub-models that may be trained on data and adjusted to optimize the R2 value. The dispersion of R2 values for the AdaBoost and RF sub-models is shown in Figure 13. The minimum, average, and highest R2 values for AdaBoost sub-models were 0.854, 0.876, and 0.900, respectively. Similarly, the minimum, average, and highest R2 values for RF sub-models were 0.872, 0.892, and 0.900, respectively. These results demonstrate that both the AdaBoost and RF sub-models have comparable values and a high degree of precision in forecasting GPC’s CS. Additionally, a sensitivity analysis was done to ascertain the effect of all inputs on the expected CS of GPC. The model’s performance might be affected by the input parameters and the dataset’s size. The sensitivity analysis determined how each of the nine input characteristics contributed to the projected output. Fly ash, GGBS, and NaOH molarity were determined to be the three most significant input variables.
Figure 13

Coefficient of determination of sub-models.

6.2. Comparison of Experimental and Predicted Results

To compare the experimental and predicted results for all the models employed in this study, Figure 14, Figure 15 and Figure 16 are generated for 91 mixes. The intention of this comparison was to determine the deviation of the predicted results from the experimental results for the validation of the employed models in estimating the CS of GPCs. This analysis revealed that for the DT model, the deviation from the experimental results was between 0.00 and 36.59 MPa, with an average of 7.02 MPa. Furthermore, for 34 mixes, the deviation from the experimental results was less than 3 MPa, from 3 to 10 MPa deviation was noted in 35 mixes, and above 10 MPa deviation was noted in 22 mixes (Figure 14). This showed a moderate deviation of the predicted results compared to the experimental results for the DT model. A similar comparison for the AdaBoost model revealed that the deviation of the results was in the range of 0.00 to 20.40 MPa with an average of 5.20 MPa. Additionally, for 42 mixes, the deviation was less than 3 MPa. Deviation from 3 to 10 MPa was observed in 32 mixes, and deviation greater than 10 MPa was observed in only 17 mixes (Figure 15). This showed the higher precision of the AdaBoost model compared to the DT. Similarly, the RF model results were like the AdaBoost in estimating the CS of GPCs. The deviation among the experimental and predicted results was in the range of 0.06 to 23.45 MPa, with an average of 5.33 MPa. For 43 mixes, the deviation of the results was less than 3 MPa; deviation from 3 to 10 MPa was observed in 31 mixes, and deviation higher than 10 MPa was observed in only 17 mixes (Figure 16). This analysis further validated the comparable accuracy of the AdaBoost and RF model with higher accuracy than the DT model. Additionally, the higher accuracies of the AdaBoost and RF models was confirmed since, for around 81.3% of mixes, the deviation of the predicted results from the experimental was less than 10 MPa. However, the DT model performed less accurately in estimating the CS of GPCs than the AdaBoost and RF models, as more deviation of results was noted among the experimental and predicted results. Hence, this study recommends the application of AdaBoost and RF models for the prediction of the CS of GPCs.
Figure 14

Comparison of experimental and predicted results for the decision tree model.

Figure 15

Comparison of experimental and predicted results for the AdaBoost model.

Figure 16

Comparison of experimental and predicted results for the random forest model.

7. Conclusions

The intention of this research was to employ both individual and ensemble machine learning (ML) algorithms to anticipate the compressive strength (CS) of geopolymer composites (GPCs). To forecast outcomes, one individual technique, decision tree (DT), was used, as well as two ensemble techniques, AdaBoost and random forest (RF). The following conclusions have been drawn as a result of this research: Ensemble ML approaches (AdaBoost and RF) performed better than the individual ML technique (DT) at predicting the CS of GPCs, with the AdaBoost and RF models performing with a similar degree of precision. The correlation coefficients (R2) for the AdaBoost, RF and DT models were 0.90, 0.90, and 0.83, respectively. Statistical checks and k-fold analysis verified the model’s performance. Furthermore, these checks also confirmed the comparable accuracy of the AdaBoost and RF models. The lower deviation (MAE, MAPE, and RMSE) of the predicted results and higher R2 values of the ensembled models validated their higher precision. The comparison of the experimental and predicted results further validated the higher accuracy of AdaBoost and RF models due to less deviation of the predicted results than the experimental results. On the other hand, the deviation of the DT model’s results was higher than the AdaBoost and RF models and is less recommended for estimating the CS of GPCs. Sensitivity analysis revealed that fly ash, ground granulated blast furnace slag, and NaOH molarity have a greater influence on the model’s outcome and account for 26.37%, 14.74%, and 13.12% of the contribution, respectively. However, NaOH, water/solids ratio, fine aggregate, gravel 4/10 mm, gravel 10/20 mm, and Na2SiO3 contributed 11.60%, 9.52%, 7.53%, 6.48%, 5.84%, and 4.80%, respectively, to the prediction of the outcome. This type of research will aid the construction sector by enabling the development of quick and cost-effective methods for predicting material strength. Additionally, by promoting eco-friendly construction using these strategies, the acceptance and use of GPC in construction will be expedited. This study proposes that in upcoming studies, the number of data points and results should be enhanced by experimental research, field trials, and other numerical evaluation techniques (e.g., Monte Carlo simulation). Additionally, to improve the models’ responsiveness, environmental parameters (e.g., elevated/low temperature and humidity) and a detailed description of the raw materials could be incorporated as input factors. Additionally, data from the literature should be retrieved and arranged in such a manner that the influence of different kinds of activators and precursors on the strength of GPCs can be determined using ML techniques.
  10 in total

Review 1.  Eco-friendly geopolymer prepared from solid wastes: A critical review.

Authors:  Bo Ren; Yunliang Zhao; Haoyu Bai; Shichang Kang; Tingting Zhang; Shaoxian Song
Journal:  Chemosphere       Date:  2020-11-13       Impact factor: 7.086

2.  A novel acidic phosphoric-based geopolymer binder for lead solidification/stabilization.

Authors:  Shaoyun Pu; Zhiduo Zhu; Weilong Song; Hairong Wang; Wangwen Huo; Jie Zhang
Journal:  J Hazard Mater       Date:  2021-03-20       Impact factor: 10.588

3.  Decision tree methods: applications for classification and prediction.

Authors:  Yan-Yan Song; Ying Lu
Journal:  Shanghai Arch Psychiatry       Date:  2015-04-25

4.  Machine Learning Techniques in Concrete Mix Design.

Authors:  Patryk Ziolkowski; Maciej Niedostatkiewicz
Journal:  Materials (Basel)       Date:  2019-04-17       Impact factor: 3.623

5.  Strength Development and Elemental Distribution of Dolomite/Fly Ash Geopolymer Composite under Elevated Temperature.

Authors:  Emy Aizat Azimi; Mohd Mustafa Al Bakri Abdullah; Petrica Vizureanu; Mohd Arif Anuar Mohd Salleh; Andrei Victor Sandu; Jitrin Chaiprapa; Sorachon Yoriya; Kamarudin Hussin; Ikmal Hakem Aziz
Journal:  Materials (Basel)       Date:  2020-02-24       Impact factor: 3.623

Review 6.  Palm Oil Fuel Ash-Based Eco-Efficient Concrete: A Critical Review of the Short-Term Properties.

Authors:  Mugahed Amran; Gunasekaran Murali; Roman Fediuk; Nikolai Vatin; Yuriy Vasilev; Hakim Abdelgader
Journal:  Materials (Basel)       Date:  2021-01-11       Impact factor: 3.623

7.  An Experimental and Empirical Study on the Use of Waste Marble Powder in Construction Material.

Authors:  Muhammad Sufian; Safi Ullah; Krzysztof Adam Ostrowski; Ayaz Ahmad; Asad Zia; Klaudia Śliwa-Wieczorek; Muhammad Siddiq; Arsam Ahmad Awan
Journal:  Materials (Basel)       Date:  2021-07-08       Impact factor: 3.623

Review 8.  Fly Ash-Based Eco-Efficient Concretes: A Comprehensive Review of the Short-Term Properties.

Authors:  Mugahed Amran; Roman Fediuk; Gunasekaran Murali; Siva Avudaiappan; Togay Ozbakkaloglu; Nikolai Vatin; Maria Karelina; Sergey Klyuev; Aliakbar Gholampour
Journal:  Materials (Basel)       Date:  2021-07-30       Impact factor: 3.623

  10 in total
  15 in total

1.  Assessment of Artificial Intelligence Strategies to Estimate the Strength of Geopolymer Composites and Influence of Input Parameters.

Authors:  Kaffayatullah Khan; Waqas Ahmad; Muhammad Nasir Amin; Ayaz Ahmad; Sohaib Nazar; Majdi Adel Al-Faiad
Journal:  Polymers (Basel)       Date:  2022-06-20       Impact factor: 4.967

2.  Compressive Strength of Steel Fiber-Reinforced Concrete Employing Supervised Machine Learning Techniques.

Authors:  Yongjian Li; Qizhi Zhang; Paweł Kamiński; Ahmed Farouk Deifalla; Muhammad Sufian; Artur Dyczko; Nabil Ben Kahla; Miniar Atig
Journal:  Materials (Basel)       Date:  2022-06-14       Impact factor: 3.748

3.  Using Machine Learning Algorithms to Estimate the Compressive Property of High Strength Fiber Reinforced Concrete.

Authors:  Li Dai; Xu Wu; Meirong Zhou; Waqas Ahmad; Mujahid Ali; Mohanad Muayad Sabri Sabri; Abdelatif Salmi; Dina Yehia Zakaria Ewais
Journal:  Materials (Basel)       Date:  2022-06-24       Impact factor: 3.748

Review 4.  A Systematic Review of the Research Development on the Application of Machine Learning for Concrete.

Authors:  Kaffayatullah Khan; Waqas Ahmad; Muhammad Nasir Amin; Ayaz Ahmad
Journal:  Materials (Basel)       Date:  2022-06-27       Impact factor: 3.748

5.  Comparison of Prediction Models Based on Machine Learning for the Compressive Strength Estimation of Recycled Aggregate Concrete.

Authors:  Kaffayatullah Khan; Waqas Ahmad; Muhammad Nasir Amin; Fahid Aslam; Ayaz Ahmad; Majdi Adel Al-Faiad
Journal:  Materials (Basel)       Date:  2022-05-10       Impact factor: 3.748

6.  Comparative Study of Experimental and Modeling of Fly Ash-Based Concrete.

Authors:  Kaffayatullah Khan; Ayaz Ahmad; Muhammad Nasir Amin; Waqas Ahmad; Sohaib Nazar; Abdullah Mohammad Abu Arab
Journal:  Materials (Basel)       Date:  2022-05-24       Impact factor: 3.748

7.  Compressive Strength Evaluation of Ultra-High-Strength Concrete by Machine Learning.

Authors:  Zhongjie Shen; Ahmed Farouk Deifalla; Paweł Kamiński; Artur Dyczko
Journal:  Materials (Basel)       Date:  2022-05-13       Impact factor: 3.748

8.  Compressive Strength Estimation of Geopolymer Composites through Novel Computational Approaches.

Authors:  Muhammad Nasir Amin; Kaffayatullah Khan; Waqas Ahmad; Muhammad Faisal Javed; Hisham Jahangir Qureshi; Muhammad Umair Saleem; Muhammad Ghulam Qadir; Muhammad Iftikhar Faraz
Journal:  Polymers (Basel)       Date:  2022-05-23       Impact factor: 4.967

9.  Machine Learning Prediction Models to Evaluate the Strength of Recycled Aggregate Concrete.

Authors:  Xiongzhou Yuan; Yuze Tian; Waqas Ahmad; Ayaz Ahmad; Kseniia Iurevna Usanova; Abdeliazim Mustafa Mohamed; Rana Khallaf
Journal:  Materials (Basel)       Date:  2022-04-12       Impact factor: 3.748

10.  Use of Artificial Intelligence Methods for Predicting the Strength of Recycled Aggregate Concrete and the Influence of Raw Ingredients.

Authors:  Xinchen Pan; Yixuan Xiao; Salman Ali Suhail; Waqas Ahmad; Gunasekaran Murali; Abdelatif Salmi; Abdullah Mohamed
Journal:  Materials (Basel)       Date:  2022-06-13       Impact factor: 3.748

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.