Abstract

The prediction of concrete strength is an interesting point of investigation and could be realized well, especially for the concrete with the complex system, with the development of machine learning and artificial intelligence. Therefore, an excellent algorithm should put emphasis to receiving increased attention from researchers. This study presents a novel predictive system as follows: extreme gradient boosting (XGBoost) based on grey relation analysis (GRA) for predicting the compressive strength of concrete containing slag and metakaolin. One of its highlights is a feature selection methodology, i.e., GRA, which was used to determine the main input variables. Another highlight is that its performance was compared with the frequently used artificial neural network (ANN) and genetic algorithm-artificial neural network (GA-ANN) by using random dataset and the same testing datasets. For three same testing datasets, the average R2 values of ANN, GA-ANN, and XGBoost are 0.674, 0.829, and 0.880, respectively, indicating that XGBoost has the highest absolute fraction of variance (R2). XGBoost can provide best result by testing the root mean squared error (RMSE) and mean absolute percentage error (MAPE). The average RMSE values of ANN, GA-ANN, and XGBoost are 15.569 MPa, 10.530 MPa, and 9.532 MPa, respectively, and those of MAPE of ANN, GA-ANN, and XGBoost are 11.224%, 9.140%, and 8.718%, respectively. Thus, the XGBoost definitely performed better than the ANN and GA-ANN. Finally, a type of application software based on XGBoost was developed for practical applications. This vivid software interfaces could help users in prediction and easy and efficient analysis.

1. Introduction

A series of dilemmas including waste emission and overconsumption of energy and natural resources have been currently pressing worldwide concerns because of global population explosion and rapid urbanization. According to the International Energy Agency report, 4.8 billion tons of cement will be produced worldwide by 2050, resulting in the emission of 3.8 billion tons of CO2 [1]. Therefore, it is beneficial to find and utilize active admixtures with high-quality and low-energy consumption as alternatives of cement, partly or totally [1, 2]. In addition, those active admixtures can even enhance the properties of concrete such as compressive strength, antipermeability, and corrosion resistance [3]. Therefore, mix proportion of concrete, especially the high-performance concrete, has been in increasing demand. Mix design of concrete is difficult to achieve for researchers, as it must reach a specific strength level, which is the most important property of concrete.

Currently, combining with machine learning, some researchers utilize various basic prediction models, such as logistic regression (LR), random forest model (BRF), support vector machine (SVM), and artificial neural network (ANN) [46]. These models are commonly used for predicting compressive strength of concrete, irrespective of costly and time-consuming nature. The artificial neural network, inspired by biological systems of the human brain, can learn and generalize from experience without prior knowledge. It has powerful pattern classification and pattern recognition capabilities [7]. Bilim et al. [8] conducted various ANNs and indicated their strong potential for predicting the compressive strength of granulated blast furnace slag concrete. Saridemir et al. [9] investigated ANNs for predicting the compressive strength of concrete including metakaolin and silica fume, exhibiting good performance. However, the determination of weights and biases of ANN is a complicated process, and the standard ANN algorithm has a slow convergence rate and is easy to fall into the local minimum. As a matter of course, there are many kinds of intelligent algorithms used to optimize the superparameter parameters of the basic model such as conjugate simulated annealing [10], gradient algorithm [11], and genetic algorithm (GA) [12]. Among those algorithms, the performance of GA is excellent, attributing to its advantages such as scalability, adaptation, speed, fault tolerance, modularity, autonomy, and parallelism. Firouzi et al. [13] proposed an integrated GA-ANN approach, appropriate for finding optimum reliability-based inspection plans based on minimum life cycle costs.

The ANN model needs a lot of parameters, and the learning process is like a black box. Hence, it is difficult to carry out the actual derivation step by step, affecting the credibility. In addition, in the learning process, it is easy to appear overfitting state; therefore, some researchers used GA to optimize the internal structure parameters of the neural network. Although GA is a solution for optimizing the neural network, its programming process is complex and the search speed is slow. Therefore, another new model, XGBoost, will be introduced later.

The extreme gradient boosting (XGBoost) is a novel machine learning algorithm mainly used for supporting classification, regression, and ranking. In recent years, it has gradually attracted attention owing to its excellent performance of strong learning ability and fast convergence during parallel learning computations shown in many AI competitions. In KDDCup competition including commercial sales forecast, the team of Top 10 used the XGBoost algorithm for web page text classification, customer behavior prediction, ad click rate prediction, and hazard risk prediction and other fields [14]. In consideration of those outstanding achievements, in this study, XGBoost was applied for predicting the compressive strength of concrete containing slag and metakaolin. In addition, the performance of the model was compared with the commonly used ANN and superior GA-ANN in order to reflect the advantages or disadvantages of XGBoost.

Before predicting the compressive strength of concrete based on ANN, GA-ANN, and XGBoost, the determination of main influencing factors of concrete strength (i.e., the input variables of prediction models) is a primary thing because it is beneficial to reduce the dimensions of the model and improve prediction accuracy [15]. In this study, there is a highlight that a feature selection methodology, grey relation analysis (GRA), was used to determine the main input variables. Figure 1 shows the overall process of predicting compressive strength of concrete containing slag and metakaolin. To train and test those prediction models, 600 groups of data selected from 18 research papers (Table 1) were utilized, after a necessary pretreatment process, i.e., unifying the compressive strength of concrete under different dimensions to avoid the influence of dimension effect. Those data were divided into training dataset and testing dataset. The quantitative analysis of prediction performance of ANN, GA-ANN, and XGBoost was obtained by testing the value of absolute fraction of variance (R2), mean absolute percentage error (MAPE), and root mean squared error (RMSE).

2. Preprocessing and Preanalysis of Selected Data

2.1. Preprocessing of Data
2.1.1. Unifying the Compressive Strength of Concrete

Because of the dimension effect, the values of compressive strength will be valid just when they are compared under a uniform dimension. A cube with a length of 100 mm was frequently used in 600 groups of collected data. Therefore, it acts as the uniform dimension in this study. For other values of compressive strength of concrete under different dimensions, they would be first transformed by the formula of Neville [32], as shown in the following equation:where P is the compressive strength of concrete under other dimensions; P0 is the compressive strength of cube (length = 150 mm, approximately 5.9 inches); V is the specimen volume; h is the height; and d is the maximum transverse dimension of specimen. Notably, all the numerical values must be calculated with a unit of inch. By the formula of Neville and another simple conversion (cube with 150 mm length to cube with 100 mm length), the scaling factor (the value of P/P0) based on the uniform dimension of cube with a length of 100 mm was calculated, as listed in Table 2.

2.1.2. Normalization of Data

In the system of data-driven modelling, in order to eliminate the influences of outliers, missed values, and bad data, all raw data should be normalized to fit the range of [0, 1]. It ensures that the raw data collected by major journals are more suitable for modelling, and it also helps to achieve better results and significantly speeds up the calculation [32]. The linear mapping function is as follows:where X represents the original value; Xmin and Xmax are the minimum value and maximum value of dataset, respectively; and Xm is the value after mapping, noting that the outputs will be remapped to their corresponding real values by the inverse mapping function before calculating any performance criterion [31].

3. Preanalysis of Data: Feature Selection Methodology

Generally speaking, the compressive strength of traditional concrete was determined by various factors, including cement type, water-binder ratio (W/B), sand ratio, dosage of water-reducing agent, and curing age. For concrete containing slag and metakaolin, the content of slag, MK, and the Si/Al of metakaolin (Si/Al of MK) should also be included in consideration of some samples; therefore, the Fly ash content should be considered as an important factor. The selection of main influencing factors among various factors is significant in order to improve the efficiency and accuracy of the prediction models. In this study, GRA, a feature selection methodology, was used to address the abovementioned problem for obtaining the most influential and significant factors.

3.1. Determination of the Analysis Sequence

First, we construct the reference matrix (compressive strength of concrete) X0 (k) (where k = 1, 2, …, 600, i.e., the number of sample datasets) and comparative matrix (cement type, water-binder ratio (W/B), sand ratio, dosage of water-reducing agent, curing age, slag content, fly ash content, MK content, and Si/Al of MK) Xi (k) (where i = 1, 2, …, 9, i.e., the number of factors and k = 1, 2, …, 600). The equations in Table 3 are the mathematical forms for the construction of the reference matrix and the comparative matrix [33].

3.2. Dimensionless Treatment

For the purpose of reducing the numerical fluctuation, the normalization process can be determined as follows:

3.3. Calculation of Grey Relational Coefficient

Subsequently, the grey relational coefficient can be calculated, according to the following equation [33]:where represents the absolute difference between the two sequences; , the distinguishing coefficient, is usually 0.5, and it represents the significance of . The smaller the is, the higher its distinguishability is. Therefore,  = 0.5 was considered, as it can offer more moderate distinguishing effects and better stability [34].

3.4. Calculation of Grey Relational Grade () and Rank of Results

As shown in equation (5), it is worth noting that the grey relational grade () can be used for measuring the degree of correlation between the comparative matrix and the reference matrix. The being close to 1 indicates the strong correlation, whereas being close to 0 depicts the weak correlation.

The size of the grey correlation grade objectively reflects the influencing degree of each factor on the compressive strength of concrete; therefore, choosing the influential and significant factors is important. In this study, MATLAB 7.0 was used to calculate of nine factors. According to the results of grey relational grade, the rank (cement type > sand ratio > W/B > MK content > slag content > curing age > dosage of water-reducing agent > Si/Al of MK > fly ash content) was obtained, as shown in Figure 2. However, too many input factors will decrease the calculation efficiency of the models and complicate the model architecture of the input layer and output layer pattern and thus will be unfavorable for the building models. Thus, the factors such as Si/Al of MK and fly ash content, considered as the weak factors, were not taken into account to optimize the model structure because of its low grey correlation grade ( < 0.9).

After the abovementioned three steps of preprocessing and preanalysis of data source, i.e., unifying the compressive strength of concrete, normalization of data, and selection of main influencing factors, the datasets are ready for training and testing the prediction models. The characteristics of the prepared dataset are listed in Table 4.

4. Main Theories and Parameters of ANN, GA-ANN, and XGBoost

4.1. Artificial Neural Network (ANN)
4.1.1. Artificial Neuron

ANN is inspired by understanding of the biological nervous system [35] and is a mathematical model that imitates the behavioral feature of the human neural network for distributed and parallel processing of information. These are massively parallel complex systems made up of many processing neurons connected by connection weights (wi), as shown in Figure 3.

A single neuron’s structure is simple. However, a complete network system containing a mass of basic neurons can be skilled in processing nonlinear problems. Figure 3 shows the notational convention where the orange graphics represents a computational unit and Xi is an input variable and multiplied by the respective weight Wi. After that, the results combined with bias term b will be transferred to every neuron in the next layer of the network. The summed values (one per neuron in the layer) are then forwarded to the next layer of the network usually via an activation function such as sigmoid function, as shown in equation (6) [36]. The output from the activation function Y can be determined by equation (7), and it will act as the final result or the input variable of the next layer.

4.1.2. Feed-Forward Neural Network and Backpropagation Algorithm

The feed-forward neural network is one of the simplest neural networks. Each neuron is only connected to the one in the previous layer. Take the output of the previous layer and output it to the next layer. There is no feedback between the layers. It is one of the most widely used and rapidly developed ANNs [37]. Commonly, the feed-forward neural network consists of one input layer, several hidden layers, and one output layer of neurons [38]. The original data are considered as the input information accepted by the first layer (input layer) and then is transmitted to the corresponding neurons of the second layer (hidden layer); at last, it will pass forward to the output layer. Besides, the number of input nodes can be regarded as the input parameters (main influencing factors mentioned previously), and the output node is the target prediction result. Figure 4 shows the framework of the ANN with 7 input nodes and 1 output node used in this study. Notably, in terms of empirical and experimental data types, the number of neurons of the hidden layer needs to be debugged multiple times for creating an optimal predictive neural network model.

There is no feedback process between the layers in the abovementioned ANN model. In order to get a better prediction model, some algorithms such as the frequently used backpropagation (BP) algorithm were added to the ANN for adjusting parameters. The essence of the BP algorithm is decreasing the difference between the actual result and the desirable result in each output layer by the fine adjustment of the weights in the back-propagated process. When the BP algorithm is injected into the forward-feeding neural network, the feed-forward neural network shows excellent performance such as strong generalization ability, fast convergence, and simple structure [39].

During the process of back-propagation, all of the weights should be updated by the BP algorithm in the neural network according to the following formula [40]:

In equation (8), n is the number of times, η is the learning rate, means the error signal, and is the expression of momentum factor. The difference between the target value and the true value can be evaluated by the error signal for an output unit j [41]:

The error signal is a function of the error signals of those units in the next higher layer, connected to unit j and the weights of those connections [9]:

As shown in equation (10), it is worth noting that the kth layer means the layer in front of the jth layer. The training procedure should be working until the iterative process converges. The updated connection weights are captured from the final trained network [42].

Although the BP algorithm has been widely used, it still has some shortcomings, such as long training time and easy to fall into local minimum. Therefore, some optimized algorithms were developed by researchers for combining with the ANN model to get a better prediction performance.

5. Parameters Optimized by Genetic Algorithm

5.1. Genetic Algorithm

The training process of ANN is a procedure of optimization of model parameters, and it starts from a random initial solution and iteratively finds out the optimal parameter values. During each iteration, the error function is first calculated at the current gradient, and then the search direction is determined based on the gradient. However, on this occasion, the optimization of parameters is likely to sink into local minimum (e.g., using the BP algorithm). At this point, the genetic algorithm (GA) was chosen for training the networks to better approximate the global minimum [43].

As one of the most commonly used intelligent optimization algorithms, GA is a mathematical computation model simulating the natural selection of Darwin’s biological evolution theory and the process of biological evolution of genetic variation. Moreover, it is a characteristic methodology inspired by the physical phenomenon of biological evolution (selection, crossover, and mutation) to seek the global optimal solutions in space. In the field of artificial intelligence, it produces significant influence on optimization main parameters of the artificial neural network (ANN) [44]. Thus, for creating a powerful model with a higher prediction accuracy, it is necessary to utilize the genetic algorithm to optimize initial weights and thresholds of the ANN model.

5.2. The Operating Procedure of GA

Figure 5 shows the main operation procedure of GA used for the optimization of ANN. GA’s key elements such as population size, numbers of generation, and fitness function and key operation procedure including selection, evolution, and mutation directly affect the final optimization result, making them definitely important in the operating procedure.

In this study, GA was combined with ANN to get its desired weights and thresholds. During the optimization process, the initial weights and thresholds of the network could be construed as individuals with prediction errors of the initial ANN as its genetic information. Then, the genetic operation of selection, crossover, and mutation is repeated, to search the optimal individual, namely, the optimal weights and thresholds of ANN (GA-ANN).

6. The Extreme Gradient Boosting Methodology (XGBoost)

XGBoost is a machine learning algorithm based on decision tree with a process of gradient promotion [45]. It is one of the most successful machine learning algorithms at present and has been extensively used by researchers in some competitions of machine learning attributing to its excellent learning performance and fast calculation speed, even more used than the popular deep neural network method. In addition, the XGBoost algorithm is generally considered to be superior to ANNs in processing small- and medium-sized structured data although ANNs show excellent performance in analyzing unstructured model data such as pictures, text, audio, and video.

In a word, the main calculation process of XGBoost is an accumulation of iterative results after T times, as shown in equation (11) [46], where i is the number of samples; T is the number of decision trees, and is the final predicted value of the ith sample in the Decision tree with number T. Function represents the calculation formula of the ith sample in the T decision tree equation (12)), is the weight vector corresponding to the leaf node, and is a function of the feature vector mapped to the leaf node of the decision tree.

To better understand the running process of the XGBoost model, the training process of a traditional decision tree was first introduced, as shown in Figure 6, indicating W/B, curing age, and sand ratio (not all the main factors for simple example), as the decision items for processing the data of four samples. The training process steps are as follows: (i) sample 1 is screened out when the limit is W/B > 0.4; (ii) sample 3 is separated from other samples for curing age ≤3; (iii) samples 2 and 4 are divided by the sand ratio of 0.3; and (iv) a traditional decision tree completes an accurate prediction model with some potential problems.

Figures 79 show the training process of the XGBoost model. Figure 7 represents the Tree 1 with W/B as the discriminant condition, similar to the first branch of Figure 6. The difference is that the output results of the leaf nodes are averaged to represent the training output values of the four samples (i.e., , where i = 1, 2, 3, 4), and then the training error of each sample is obtained by calculating the deviation of the predicted value from the true value. The obtained errors in Tree 1 will be used as the input values of Tree 2 (Figure 8) and the curing age (≤3 or >3) will act as the discrimination standard. Then, the value of will be obtained based on the average value of decided errors. The predictive values of samples were obtained by the summation of and . The calculated training errors in Tree 2 will be used as input into Tree 3. Similarly, a decision tree based on sand ratio was established, as shown in Figure 9, and the average values are all 0, i.e., , where i = 1, 2, 3, 4. Then, the final predicted values of samples 1, 2, 3, and 4 were calculated by equation and found as 69.5, 80.5, 79.5, and 80.5 MPa, respectively. At this point, the model training is finished. You can find that the output values in XGBoost are different than those in the traditional decision tree. To be more specific, the predictive values of four samples are not equal to their true values in XGBoost and thus seem to be worse than the traditional decision tree; however, it can solve a big problem of traditional decision tree, i.e., overfitting.

For instance, the traditional decision tree and the XGBoost were used, respectively, to predict the strength value of sample 5, indicating that the traditional decision tree will output the strength value of 81 MPa and that by the XGBoost model is 80.5 MPa. The following analysis can prove the competitive prediction ability of XGBoost. Compared with sample 2, except that the slag content of sample 5 is less than sample 2, other variables have exactly the same values. In combination with the general conclusion of the influence of slag content on the concrete strength, the compressive strength of sample 5 was obtained as predicted to be less than 81 MPa. In the same way, it was easy to know that the strength value of sample 5 will be more than that of sample 3 (i.e., 79 MPa) on account of the increasing curing age of samples under the same mix proportion. Hence, the approximate range of the strength of sample 5 will be in the range 79–81 MPa, indicating that the predictive result of XGBoost (80.5 MPa) is more reasonable and the predictive performance of XGBoost is superior than that of the traditional decision tree. This phenomenon is caused by the overfitting training process of the traditional decision tree. It is sometimes inclined to focus on the characteristics of the training data and summarizes the internal rules of the training set (four samples). However, for the new test dataset (sample 5), it cannot reveal its internal change rule. Therefore, a single decision tree model is not generalizable.

7. Setting Structural Parameters and Learning Rate of ANN, GA-ANN, and XGBoost

In this study, the prediction performances of ANN, GA-ANN, and XGBoost were tested and compared. In order to avoid the influence of network structures on the ability of prediction models, the structural parameters of ANN and GA-ANNs were fixed, namely, with the same hidden layers and nodes, as listed in Table 5. Because XGBoost is a definitely different model with a tree structure compared with ANN and GA-ANN, it has different parameters, but some parameters similar to those in ANN and GA-ANN were set.

The programming of ANN, GA-ANN, and XGBoost was performed on the MATLAB 7.0 platform, combined with the dataset (600 groups of data selected from 18 research papers) prepared for the training and testing prediction models.

The prediction ability of those models was tested by comparing the value of absolute fraction of variance (R2), root mean squared error (RMSE), and mean absolute percentage error (MAPE). The detailed results are as follows.

8. Results and Discussion

8.1. Comparison of R2, RMSE, and MAPE of ANN, GA-ANN, and XGBoost with Random Testing Dataset

To train and test those prediction models, 600 groups of data after preprocessing were randomly divided into a training dataset (containing 480 groups of data, i.e., 80% of the total dataset) and a testing dataset (containing 120 groups of data, i.e., 20% of total dataset). The values of the absolute fraction of variance (R2), root mean squared error (RMSE), and mean absolute percentage error (MAPE) were adopted to comprehensively evaluate the prediction performances of these models. The detailed equations are as follows:where ti is the true value, Oi is the predictive value, and n is the total number [41, 47].

Figure 10 shows the values of R2, RMSE, and MAPE of the three prediction models by using the test dataset. It is easy to find that the values of R2, RMSE, and MAPE of ANN are 0.708, 13.38 MPa, and 9.72%, respectively. With the parameters optimized by the genetic algorithm (GA), the prediction performance is better attributing to the better values of R2 (0.837), RMSE (11.74 MPa), and MAPE (8.26%) of GA-ANN. The third image shows the prediction results of XGBoost. As we know, XGBoost is often known to deal with the classifying problems owing to its excellent performance, but surprisingly, it has also achieved good performance for dealing with regression prediction problems. As shown in Figure 10 (the third), the data points are more concentrated on the line y = x and nearly without the deviation beyond ± 25 MPa (the red dotted lines) of predictive values compared with target values. The results of R2 (0.872), RMSE (8.62 MPa), and MAPE (8.25%) are also good, better than those obtained by GA-ANN.

The above analysis and discussion indicate superior prediction performance of XGBoost than those by ANN and GA-ANN; however, the results were obtained based on the random testing dataset, indicating that the dataset applied to XGBoost is different with those of ANN and GA-ANN. A special case where the dataset applied to XGBoost is easy to predict was tested, exhibiting good result. In order to eliminate the possibility of that special case and get a more persuasive conclusion, the R2, RMSE, and MAPE were tested by using the same testing dataset as described in the following section.

8.2. Comparison of R2, RMSE, and MAPE of ANN, GA-ANN, and XGBoost with the Same Testing Dataset

The R2, RMSE, and MAPE values were calculated thrice by using the three datasets, i.e., dataset 1, 2, and 3 with 120 data in each dataset generated by random selection. In order to easily compare the deviation degree of predictive values (red hollow points) and true values (black solid points) of ANN, GA-ANN, and XGBoost, the diagrams of the results are presented side-by-side (as shown in Figure 11). It is obvious that the predictive values of XGBoost have the maximal degree of coincidence with true values shown in Figure 11, and the degree of coincidence of those points in the ANN is the lowest. It is a subjective conclusion; therefore, the objective results of ANN, GA-ANN, and XGBoost were compared. The specific R2 values obtained by the ANN, GA-ANN, and XGBoost are 0.708, 0.833, and 0.874, respectively. Second, for the same 120 groups of testing data, the RMSE values obtained by the ANN, GA-ANN, and XGBoost are 13.377, 9.413, and 10.982 MPa, respectively. Lastly, the MAPE values obtained by the ANN, GA-ANN, and XGBoost are 9.733%, 9.289%, and 9.544%, respectively.

The values in Figure 11 indicate that only the R2 value obtained by XGBoost is the highest, whereas the RMSE and MAPE obtained by the GA-ANN are the lowest than those obtained by the ANN and XGBoost, indicating better prediction performance of GA-ANN than that of XGBoost. However, the results obtained from dataset 2 and dataset 3, as shown in Figures 12 and 13, respectively, show different conclusions. For getting the comprehensive results of comparison, the average R2, RMSE, and MAPE values were calculated by the ANN, GA-ANN, and XGBoost. Table 6 shows the average R2 values (i.e., ) of the ANN, GA-ANN, and XGBoost as 0.674, 0.829, and 0.880, respectively, indicating that the highest absolute fraction of variance (R2) by the XGBoost. The RMSE and MAPE values obtained by XGBoost were the best. The average RMSE value of the ANN, GA-ANN, and XGBoost was 15.569, 10.530, and 9.532 MPa, respectively, and the average MAPE value of the ANN, GA-ANN, and XGBoost was 11.224%, 9.140%, and 8.718%, respectively, certainly verifying the superior prediction performance of XGBoost.

The above analyses and discussions of the results obtained by using the random testing dataset and three same testing datasets clearly indicate more stable prediction performance of XGBoost with the highest prediction accuracy. Therefore, XGBoost is a recommended prediction model and could be used for predicting the compressive strength of concrete containing slag and metakaolin.

9. Application Software for Easy Prediction Based on XGBoost and GRA

At present, the design and development of application software based on various prediction models is receiving increasing attention. This behavior can encourage to develop software from theory to practical application. Therefore, in this study, C# programming was utilized for integrating the involved research results and methods into an application software (named PSCS V1.0) with straightforward interactive interfaces to help users without the knowledge of MATLAB to predict and analyze easily and efficiently. With the help of PSCS, users can just input the dimension of sample (unifying the compressive strength of concrete under different dimensions automatically) and main parameters (the parameters obtained by GRA, i.e., type of cement, W/B ratio, sand ratio, slag content, MK content, dosage of water-reducing agent, and curing age), as shown in Figure 14. After a short wait time for computing of XGBoost (Figure 15), the final prediction results obtained are shown in Figure 16.

PSCS is a short but strong system and could be used for predicting the compressive strength of concrete containing slag and metakaolin. As the dataset used for the training XGBoost model was collected from various journal articles, it can ensure the applicability for different cases. In the following study, relevant literature data were collected continually for building up a complete and more comprehensive database.

10. Conclusions

In conclusion, the proposed novel predictive system XGBoost was successfully applied to predict the compressive strength of concrete containing slag and metakaolin, and its performance was compared with the commonly used ANN and optimized GA-ANN models. The datasets used for the training and testing three models were selected from 18 research articles (containing 600 groups of data), with the treatments of unifying the compressive strength of concrete, normalization of data, and selection of main influencing factors. With the help of GRA, the main input parameters such as the type of cement, W/B ratio, sand ratio, slag content, MK content, dosage of water-reducing agent, and curing age were determined scientifically and reasonably. After that, the prediction models were trained and tested by calculating the values of R2, RMSE, and MAPE. The main results and conclusions can be drawn as follows:(1)For the random testing dataset, the R2, RMSE, and MAPE obtained by the ANN are 0.708, 13.38 MPa, and 9.72%, respectively, and those obtained by the GA-ANN are 0.837, 11.74 MPa, and 8.26%, respectively, and the results obtained by XGBoost are 0.872, 8.62 MPa, and 8.25%, respectively, proving a superior prediction performance of XGBoost than those of ANN and GA-ANN.(2)XGBoost also exhibited superior prediction performance for prediction of the models based on the same testing dataset. The average R2 values of ANN, GA-ANN, and XGBoost are 0.674, 0.829, and 0.880, respectively. For the RMSE and MAPE, the average RMSE values of the ANN, GA-ANN, and XGBoost are 15.569, 10.530, and 9.532 MPa, respectively, and the average MAPE values of the ANN, GA-ANN, and XGBoost are 11.224%, 9.140%, and 8.718%, respectively. Therefore, XGBoost was the best prediction model for testing the same datasets.(3)Based on the analyses and discussions obtained by using the random testing dataset and the same testing datasets, XGBoost exhibited better performance than the ANN and GA-ANN for prediction capability. Therefore, XGBoost is the best method for predicting the compressive strength of concrete containing slag and metakaolin in this study. Lastly, in order to make XGBoost user friendly and easy to operate, the application software (PSCS V1.0) was encoded by C# in this study, making it a short but strong system that could predict the compressive strength of concrete containing slag and metakaolin efficiently.

Data Availability

The data used in this paper are from references, and the corresponding literature sources have been marked, so the data have reliability and authenticity.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

The authors would like to express gratitude to all those who helped them during the data analysis and writing of this thesis. This work was supported by the National Natural Science Foundation of China (no. 51828201), China Postdoctoral Science Foundation (no. 2018M632518), Anhui Postdoctoral Science Foundation (nos. 2017B150 and 2018B248), Natural Science Foundation of Anhui University (no. KJ2018A0074), and Key Research and Development Program Project of Anhui Province (no. 201904a07020081).