Next Article in Journal
Research on Microstructure, Mechanical Properties, and High-Temperature Stability of Hot-Rolled Tungsten Hafnium Alloy
Previous Article in Journal
The Diatomite Grinding Technology Concept for the Protection of Diatomite Shells and the Control of Product Grading
Previous Article in Special Issue
Investigation of Using Calcined Coal Gangue as the Co-Blended Precursor in the Alkali-Activated Metakaolin
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Compressive Strength of Concrete Specimens Based on Interpretable Machine Learning

Power China Chengdu Engineering Corporation Ltd., Chengdu 610031, China
*
Authors to whom correspondence should be addressed.
Materials 2024, 17(15), 3661; https://doi.org/10.3390/ma17153661
Submission received: 14 June 2024 / Revised: 10 July 2024 / Accepted: 17 July 2024 / Published: 24 July 2024

Abstract

:
The aim of this paper is to explore an effective model for predicting the compressive strength of concrete using machine learning technology, as well as to interpret the model using an interpretable method, which overcomes the limitation of the unknowable prediction processes of previous machine learning models. An experimental database containing 228 samples of the compressive strength of standard cubic specimens was built in this study, and six algorithms were applied to build the predictive model. The results show that the XGBoost model has the highest prediction accuracy among all models, as the R2 of the training set and testing set are 0.982 and 0.966, respectively. Further analysis was conducted on the XGBoost model to discuss its applicability. The main steps include the following: (i) obtaining key features, (ii) obtaining trends in the evolution of features, (iii) single-sample analysis, and (iv) conducting a correlation analysis to explore methods of visualizing the variations in the factors that exert influence. The interpretability analyses on the XGBoost model show that the contribution to the compressive strength by each factor is highly in line with the conventional theory. In summary, the XGBoost model proved to be effective in predicting concrete’s compressive strength.

1. Introduction

Concrete is a ubiquitous material in civil engineering, and determining its compressive strength accurately is crucial to engineering construction and operation. As a composite material, it is challenging to accurately predict its compressive strength due to its complex and variable composition, differences in curing times, and curing environments. Currently, the main methods for obtaining the compressive strength of concrete are the rebound method, core drilling method, and model prediction method. The rebound method is easy to carry out; however, using different rebound meters for the same type of concrete can yield large discrepancies and fail to accurately estimate the internal strength. The core drilling method is time-consuming, and the results are not stable. Predictive modeling involves creating mathematical models that correlate the material mix, age, and measured compressive strength, but due to the nonlinear relationships between these factors, the establishment of an explicit mathematical model is more difficult. Therefore, there is an urgent need for a fast, accurate, and reliable method for predicting the compressive strength of concrete.
With advances in artificial intelligence and big data analytics, some of these challenges are being mitigated. Machine learning algorithms, including decision trees (DTs), Naive Bayes, and ensemble methods such as Random Forest, AdaBoost, Gradient Boosting Decision Trees (GBDTs), and XGBoost, have been increasingly employed. For instance, Rankbar et al. [1] utilized a deep learning algorithm and evaluated six models with varying hyperparameters. Adhikary et al. [2] developed an artificial neural network model using five input features that provided a more accurate prediction of the ultimate shear strength in reinforced concrete than in traditional formulas. Mansouri et al. [3,4] applied two algorithms, ANN and ANFIS, successfully predicting the debonding strength of masonry. Extensive experimentation by previous researchers on concrete’s compressive strength illustrates the potential of machine learning algorithms to significantly enhance data utility. Nevertheless, current research has not adequately addressed how variations in input parameters influence prediction outcomes. This gap highlights the growing need for methods that can elucidate the predictive modeling process, balancing accuracy with interpretability. Therefore, constructing a precise mathematical model is challenging.
In order to improve the credibility of machine learning models, a variety of interpretive methods have been proposed. Among these methods, the SHAP method is a more comprehensive interpretable method. It includes global and local interpretations. Local interpretation, or single-sample analysis, quantifies the contribution of each input variable to the predicted values, whereas global interpretation assesses feature importance and dependency. For example, Feng et al. [5] aligned the feature importance rankings in a deep-beam prediction model with established theories. Mangalathu S et al. [6] performed an interpretive analysis for seismic performance assessment, and further work by Mangalathu S et al. [7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23] clarified the importance of specific characteristics and their impact on the seismic performance of infrastructures using SHAP. Bentz predicted changes in the heat release, chemical shrinkage, and compressive strength over time using kinetic calibration [24,25,26,27,28]. The method suggested by Mechling et al. greatly improves the accuracy of slurry strength calculations. These results can be used to calculate concrete’s strength [29,30,31,32]. Abuodeh et al. used ANN with Sequence Feature Selection (SFS) and a Neural Interpretation Diagram (NID) to greatly improve the accuracy of the model and provide valuable insights into the ANN’s compressive strength predictions of different UHPC mixes [33,34,35,36]. However, the ambiguous influence patterns of input features on concrete’s compressive strength have so far limited the full validation of interpretive analyses in machine learning models.
Accordingly, this study employed various machine learning algorithms to predict concrete’s compressive strength, verifying the soundness of the prediction model through the analysis of key features, trends in feature evolution, single-sample interpretation, and a correlation analysis.

2. Research Methods and Modeling Process

2.1. Algorithmic Principle

In this study, six algorithms were selected to predict the concrete’s compressive strength. By comparing the six algorithms, the best performing prediction model was obtained. The principles of the selected algorithms are briefly described below.
K Nearest Neighbors (KNN) is one of the simplest methods in data mining techniques; to predict the target value based on the distance between data points, it selects K nearest neighbor data points and then uses their labels or averages to classify or regress them [37,38]. A Decision Tree (DT) is a supervised learning method that can summarize the decision rules from a series of featured and labeled data and present these rules in a tree-like structure to solve classification and regression problems. The main parameters of the model are the maximum depth of the tree, maximum number of leaf nodes, minimum number of partitioned samples, and so on [39].
Random Forest (RF) [40,41] is an integrated algorithm based on bagging, which combines multiple decision trees, where the dataset is selected randomly and put back each time, and some of the features are also randomly selected as inputs. RF reduces the risk of overfitting a single tree by integrating multiple decision trees. Each tree sees only part of the data and some of the features, thus improving the generalization of the overall model.
A boosting algorithm boosts a weak learner to a strong learner algorithm by means of concatenation [42]. Depending on the strategy, there are three common integrated algorithms, namely GBDT [43], Adaboost [44], and XGBoost [45,46]. GBDT is designed to continuously reduce the residuals (regression) by continuously adding new trees designed to build a new model in the direction of residual reduction (negative gradient). The loss function is designed to reduce the residuals as fast as possible. In order to obtain the residuals, all decision trees use CART regression trees. Adaboost emphasizes adaptability by constantly modifying the sample weights (increasing the weights of the wrong samples and decreasing the weights of the right samples) and constantly adding weak classifiers for boosting. The principle of XGBoost is similar to that of GBDT, which is optimized as a distributed gradient. XGBoost is an optimized distributed gradient boosting library and is also a tool for massively parallel boosting trees. Compared with GBDT [47], XGBoost has three main improvements: (1) XGBoost takes into account the complexity of the tree, while GBDT does not; (2) XGBoost fits the second-order derivative expansion of the loss function of the previous round, while GBDT fits the first-order derivative of the loss function of the previous round, and thus, the former has fewer iterations and is more accurate; and (3) XGBoost has a faster running speed when selecting the optimal cut-points as it uses multi-threading to carry out the process.
In summary, the first two are single algorithms, and the last four are integrated algorithms. While RF is a bagging-class integrated algorithm, GBDT, Adaboost, and XGBoost are boosting-class integrated algorithms, and the best performing model is selected by comparing the six algorithms.

2.2. Evaluation Metrics

For regression tasks, the common evaluation metrics are the MSE (Mean Square Error), MAE (Mean Absolute Error), RMSE (Root Mean Square Error), MAPE (Mean Absolute Percentage Error), and the Coefficient of determination R-Squared (R2). And both the MSE and MAE are measures of the gap between the predicted value and the true value. Therefore, the latter four were chosen in this paper as the indicators for evaluating the model. Here, four commonly used evaluation metrics were selected and defined as follows:
--R2:
R 2 = 1 i = 1 m ( P i T i ) 2 i = 1 m ( P i T ¯ ) 2
--RMSE:
RMSE = i = 1 m ( P i T i ) 2 m
--MAE:
MAE = i = 1 m P i T i m
--MAPE:
MAPE = 100 % m i = 1 m P i T i

2.3. The Interpretable Method

Lundberg and Lee [48,49,50] proposed the interpretable method known as the Shapley additive explanation, which is an interpretive framework based on the concept of Shapley values in cooperative game theory. Shapley values are a method of assigning a degree of contribution to each parameter in a cooperative game, and SHAP builds on this foundation by applying the analysis of the contribution of features to the model output. SHAP considers the effect of each feature on different combinations of features and calculates the contribution of each feature to a certain predicted value. By weighing all possible combinations of features, the SHAP value provides a weight for each feature that explains the relative contribution of that feature to the model output.
To interpret the predictions for individual samples, Python’s SHAP package creates a visualization: The influence of each eigenvalue on the prediction result is expressed as a numerical value and summed to obtain the prediction value. As shown in Figure 1, the baseline value is the average of the predictions for the entire sample, with arrows to the left indicating a negative impact value and arrows to the right indicating a positive impact value. SHAP is the conditional expectation function of the model. A0, A1, A2, A3, and A4 denote the weights of individual factors influencing the outcome, and E[f(x)], etc. denote the mathematical expectation of individual factors on the predictive model, the sum of which is f(x), i.e., the model prediction.

2.4. Implementation of ML Methods

As shown in Figure 2, the modeling process of this study was as follows: Firstly, the database of the samples of concrete’s compressive strength was established. Among them, the training set accounted for 80%, and the testing set accounted for 20%. Then, the parameters were adjusted to avoid model overfitting or underfitting, and the prediction model was established. Then, several prediction models were evaluated for performance and the best performing prediction model (XGBoost) was selected. Finally, the credibility of the XGBoost model was verified using (1) the key feature; (2) trends in the evolution of features; (3) a single-sample analysis; and (4) a correlation analysis.

3. Experimental Database

3.1. Selection of Input and Output Features

In this study, several parameters were selected as input features for the model; these parameters included the water–binder ratio, water, sand ratio, superplasticizer, air-entraining agent, slump, air content, and age. The model’s output was the concrete specimen’s compressive strength (vu).

3.2. Details of the Database

The database contains 228 samples. The database characteristics are as follows: (1) the specimens are 150 × 150 × 150 standard cubic blocks; (2) all of the specimens are standard cured; (3) all of the specimens are of secondary concrete, which is consistent except for the control variables, such as the selection of aggregates, and the water content of the sand, etc.
Table 1 describes the scope of important features of the database. The water–binder ratio ranges from 0.35 to 0.55, the water varies greatly from 124 kg/L to 154 kg/L, and the sand ratio ranges from 37% to 43%. The superplasticizer and air-entraining agent vary from 0.50% to 0.75% and from 0.20% to 1.5%, respectively. The slump and air content vary from 163 mm to 202 mm and from 1.2% to 6.4%, respectively. The ages were 7 d, 28 d, and 90 d. The experimental values of the compressive strength of concrete (vtest = Vtest/area) range from 17.1 MPa to 61.4 MPa. In addition, the distribution of these features can be visualized in Figure 3.

4. Model Results and Discussion

This section compares the performance of the models based on six algorithms, namely, the single algorithms KNN and DT, and the integrated algorithms RF, GBDT, Adaboost, and XGBoost, in predicting the compressive strength of concrete. According to Section 2.4, the important hyperparameters in the six ML models were determined, as shown in Table 2. Four evaluation metrics were selected to measure the performance of the models.
Table 3 presents the performance advantages and disadvantages of several ML models on the training and testing set. In the training set, the R2 of several algorithms (except KNN) is around 0.980, but XGBoost has the smallest RMSE, MAE, and MAPE values. Meanwhile, in the testing set, XGBoost has the closest R2 to 1 and the smallest RMSE, MAE, and MAPE values, indicating that the XGBoost model has the optimal generalization ability.
In Figure 4, the horizontal axis shows the experimental values, and the vertical axis shows the predicted values. The points distributed on the diagonal line indicate that the predicted values are the same as the test values. The points to the upper left of the diagonal indicate that the predicted value is greater than the experimental value. It can be seen that the scatter points of the XGBoost model are more concentrated in and around the diagonal. The KNN model has the most dispersed scatter distribution area, and several other algorithmic models have roughly the same concentration of distribution. It can be concluded that the integrated model outperforms the single ML model. The horizontal axis in Figure 5 shows the ratio of predicted to experimental values, denoted by χ; the vertical axis indicates the number of the interval. In Figure 5f, all models range from 0.95 to 1.05.
In summary, the XGBoost model has high accuracy and good generalization ability. However, the model’s prediction process is invisible, and it is unknown how the input features affect the prediction. In the next chapter, the prediction process will be uncovered through methods such as a SHAP value analysis.

5. Model’s Interpretable Analysis

5.1. Key Features

The absolute value of the SHAP value of each feature for all individuals is summed to give the overall feature importance. Then, the global importance of each feature is divided by the sum to obtain the relative importance value of each individual feature in the database. By ranking the importance of the features, we can know which input features are very important and which input features can be ignored. Figure 6 shows the importance of the input features.
As shown in Figure 6, the relative importance value of age (d) is 20.92%, and its effect on the concrete’s compressive strength is the largest. Then, the next most important influence on compressive strength is the air content (%), accounting for about 19.34%. The slump (mm) is the third most important input feature, and it has a value of 18.57%. The water-–binder ratio and water (kg) have the same effect on compressive strength, and their SHAP values are 15.9% and 15.19%, respectively. The sand ratio (%) also has some effect on the compressive strength, while the superplasticizer and air-entraining agent dosage have negligible impacts on the compressive strength.

5.2. Trends in the Evolution of Features

Trends in the evolution of features plots can represent the response of features to model predictions, reflecting whether the relationship between the input and output features is linear or nonlinear and monotonic or complex. Figure 7 shows the trends in SHAP values for age (d), air content (%), slump (mm), water-binder ratio, water (kg), and sand ratio (%).
It is clear that the age (d) and slump (mm) have a strong positive impact on compressive strength, while the air content (%) and water–binder ratio have a strong passive impact on compressive strength. The approximately linear relationship between the age, air content, water–binder ratio, and SHAP values shows that the higher the age, the lower the air content, and the lower the water–binder ratio, the higher the compressive strength of the concrete. The compressive strength increases faster when the slump increases from 163 mm to 170 mm, while the change in compressive strength is less pronounced when the slump is greater than 170 mm. When the water is more than 135 kg, a change in water has a negligible effect on the compressive strength. When the sand ratio is between 37% and 40%, the compressive strength increases as the sand ratio increases, while when the sand ratio is greater than 41%, the compressive strength decreases instead as the sand ratio increases; therefore, it can be concluded that a sand ratio of 40–41% is the optimum sand ratio. In conclusion, the trends in the evolution of features for important variables plot clearly reveals the impact of features on the prediction, helping us to make informed decisions using the ML model. In conclusion, the feature dependence diagram gives us an understanding of the law of the influence of input features on compressive strength, which is of great significance in guiding the actual production work.

5.3. Single-Sample Analysis

In this study, the test results of one of the specimens were interpreted by a single-sample model, as shown in Figure 8, which demonstrates that the SHAP value decomposes the prediction of compressive strength into the sum of the contributions of each input variable. The “base value” in Figure 8 represents the baseline value, i.e., the average of the XGBoost model predictions for all variables in the database (35.8 kN). The red arrows indicate the variables that contribute positively to the base compressive strength, while the blue arrows indicate the variables that lead to a decrease in the base compressive strength, and the lengths of the arrows indicate the incremental and decremental amounts. For this specimen, the water–binder ratio, air content, slump, and air-entraining agent content are the key variables that contribute positively to the baseline compressive strength, and their increments decrease in that order. The remaining variables, such as age, are the key variables that contribute to the reduction in the benchmark compressive strength. The algebraic summation of the increments or decrements of all the variables with the benchmark compressive strength is the model-predicted value of 31.00 kN in the figure, which is close to the test value of 31.1 kN.

5.4. Correlation Analysis

A correlation heat map represents the correlation between the input features and the output features. The conclusions drawn from the XGBoost model were validated by correlation analysis. In Figure 9, the correlation heat map uses different value sizes and color depths to describe the correlation between different variables. A larger value indicates a higher correlation and vice versa, a lower correlation. Colors are used to distinguish between positive and negative correlations.
As shown in Figure 9, the value at the intersection of age and compressive strength is 0.71 with a light green color, indicating that the longer the age, the higher the compressive strength of the concrete. The value at the intersection of water-binder ratio and compressive strength is negative 0.68 with a darker pink color. This shows that the compressive strength decreases with the increase in the water-binder ratio. In addition, the age, water-binder ratio, air content, and slump have a significantly higher effect on the compressive strength of concrete than other variables, which is in line with Figure 6, proving that these four features are the key factors affecting the compressive strength.

6. Conclusions

This paper presents the application of machine learning algorithms in the prediction of concrete’s compressive strength. The theoretical background and implementation process of the algorithms are presented. A database containing 228 specimens was collected and used to train and test the ML model. The results show that the integrated algorithm outperforms the single algorithm, with XGBoost having the best performance. An interpretable analysis of the XGBoost model was also carried out to overcome the “black box” problem of previous machine learning models and increase the credibility of the model. In addition, it is possible to understand how the input parameters affect the target value, which is meaningful to guiding the actual production.
However, there were some limitations in this study: (1) the effects of aggregate and admixture types on compressive strength were not investigated; (2) this paper only analyzed the compressive strength, so important parameters such as seepage resistance, frost resistance, and other properties in the actual project were not involved; and (3) the calculation method for compressive strength was not accurate. These aspects need to be further studied.

Author Contributions

Methodology, Conceptualization, Q.D.; Software, Investigation, X.L.; Resources, Supervision, Validation, T.Z.; Writing—original draft, Investigation, W.W.; Writing—review & editing, Methodology, G.L.; Project administration, Resources, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Nyingchi Science and Technology Plan Project (No. 2023-SYQ-010).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

All authors were employed by the company Power China Chengdu Engineering Corporation Ltd. The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Ranjbar, I.; Toufigh, V.; Boroushaki, M. A combination of deep learning and genetic algorithm for predicting the compressive strength of high-performance concrete. Struct. Concr. 2022, 23, 2405–2418. [Google Scholar] [CrossRef]
  2. Adhikary, B.B.; Mutsuyoshi, H. Prediction of shear strength of steel fiber RC beams using neural networks. Constr. Build. Mater. 2006, 20, 801–811. [Google Scholar] [CrossRef]
  3. Mansouri, I.; Kisi, O. Prediction of debonding strength for masonry elements retrofitted with FRP composites using neuro-fuzzy and neural network approaches. Compos. B Eng. 2015, 70, 247–255. [Google Scholar] [CrossRef]
  4. Luo, H.; Paal, S.G. Machine learning-based backbone curve model of reinforced concrete columns subjected to cyclic loading reversals. J. Comput. Civ. Eng. 2018, 32, 04018042. [Google Scholar] [CrossRef]
  5. Feng, D.C.; Wang, W.J.; Mangalathu, S.; Hu, G.; Wu, T. Implementing ensemble learning methods to predict the shear strength of RC deep beams with/without web reinforcements. Eng. Struct. 2021, 235, 111979. [Google Scholar] [CrossRef]
  6. Mangalathua, S.; Karthikeyanb, K.; Feng, D.C.; Jeon, J.S. Machine-learning interpretability techniques for seismic performance assessment of infrastructure systems. Eng. Struct. 2022, 250, 112883. [Google Scholar] [CrossRef]
  7. Mangalathua, S.; Shinb, H.; Choic, E.; Jeon, J.S. Explainable machine learning models for punching shear strength estimation of flat slabs without transverse reinforcement. J. Build. Eng. 2021, 39, 102300. [Google Scholar] [CrossRef]
  8. Vakhshouri, B.; Nejadi, S. Prediction of compressive strength of self-compacting concrete by ANFIS models. Neurocomputing 2018, 280, 13–22. [Google Scholar] [CrossRef]
  9. Namyong, J.; Sangchun, Y.; Hongbum, C. Prediction of compressive strength of in-situ concrete based on mixture proportions. J. Asian Archit. Build. Eng. 2004, 3, 9–16. [Google Scholar] [CrossRef]
  10. Hlobil, M.; Smilauer, V.; Chanvillard, G. Micromechanical multiscale fracture model for compressive strength of blended cement pastes. Cem. Concr. Res. 2016, 83, 188–202. [Google Scholar] [CrossRef]
  11. Kabir, A.; Hasan, M.; Miah, K. Predicting 28 days compressive strength of concrete from 7 days test result. In Proceedings of the International Conference on Advances in Design and Construction of Structures, Bangalore, India, 19–20 October 2012; pp. 18–22. [Google Scholar]
  12. Zain MF, M.; Abd, S.M. Multiple regression model for compressive strength prediction of high performance concrete. J. Appl. Sci. 2009, 9, 155–160. [Google Scholar] [CrossRef]
  13. Hamid, Z.N.; Jamali, A.; Narimanzadeh, N.; Akbarzadeh, H. A polynomial model for concrete compressive strength prediction using GMDH-type neural networks and genetic algorithm. In Proceedings of the 5th WSEAS International Conference on System Science and Simulation in Engineering, Tenerife, Canary Islands, Spain, 16–18 December 2006; pp. 13–18. [Google Scholar]
  14. Li, Q.F.; Song, Z.M. High-performance concrete strength prediction based on ensemble learning. Constr. Build. Mater. 2022, 324, 126694. [Google Scholar] [CrossRef]
  15. Moon, S.; Munira, C.A. Utilization of prior information in neural network training for improving 28-day concrete strength prediction. J. Constr. Eng. Manag. 2021, 147, 04021028. [Google Scholar] [CrossRef]
  16. Feng, D.C.; Liu, Z.T.; Wang, X.D.; Chen, Y.; Chang, J.Q.; Wei, D.F.; Jiang, Z.M. Machine learning-based compressive strength prediction for concrete: An adaptive boosting approach. Constr. Build. Mater. 2020, 230, 117000. [Google Scholar] [CrossRef]
  17. Wang, M.; Zhao, G.; Liang, W.; Wang, N. A comparative study on the development of hybrid SSA-RF and PSO-RF models for predicting the uniaxial compressive strengthof rocks. Case Stud. Constr. Mater. 2023, 18, e02191. [Google Scholar]
  18. Ni, H.-G.; Wang, J.-Z. Prediction of compressive strength of concrete by neural networks. Cem. Concr. Res. 2000, 30, 1245–1250. [Google Scholar] [CrossRef]
  19. Tran, V.Q.; Dang, V.Q.; Ho, L. Evaluating compressive strength of concrete made with recycled concreteaggregates using machine learning approach. Constr. Build. Mater. 2022, 323, 126578. [Google Scholar] [CrossRef]
  20. Dutta, S.; Samui, P.; Kim, D. Comparison of machine learning techniques to predict compressive strength of concrete. Comput. Concr. 2018, 21, 463–470. [Google Scholar]
  21. Zhang, B.W.; Geng, X.L. Prediction of concrete compressive strength based on tuna swarm algorithm optimization extreme learning machine. Appl. Res. Comput. 2024, 41, 444–449. [Google Scholar] [CrossRef]
  22. Xue, G.B.; Hu, A.L.; Wei, Y.; Fneg, Y.J.; Liang, K.; Li, L.H. Compressive Strength Predict. Concr. Based Cost-Sensitive Coefficients. J. Xi’an Univ. Technol. 2022, 38, 588–593. [Google Scholar] [CrossRef]
  23. Al-Shamiri, A.K.; Kim, J.H.; Yuan, T.-F.; Yoon, Y.S. Modeling the compressive strength of high-strength concrete: An extreme learning approach. Constr. Build. Mater. 2019, 208, 204–219. [Google Scholar] [CrossRef]
  24. Bentz, D.P. Modelling cement microstructure: Pixels, particles, and property prediction. Mater. Struct. Constr. 1999, 32, 187–195. [Google Scholar] [CrossRef]
  25. Karni, J. Prediction of compressive strength of concrete. Mater. Constr. 1974, 73, 197–200. [Google Scholar] [CrossRef]
  26. Silva, R. Use of Recycled Aggregates from Construction and Demolition Waste in the Production of Structural Concrete. Res. Net. 2015. [Google Scholar] [CrossRef]
  27. De Brito, J.; Kurda, R.; da Silva, P.R. Can we truly predict the compressive strength of concrete without knowing the properties of aggregates? Appl. Sci. 2018, 8, 1095. [Google Scholar] [CrossRef]
  28. Bentz, D.P.; Garboczi, E.J.; Bullard, J.W.; Ferraris, C.F.; Martys, N.S.; Stutzman, P.E. Virtual testing of cement and concrete. In Significance of Tests and Properties of Concrete and Concrete-Making Materials; ASTM: West Conshohocken, PA, USA, 2006. [Google Scholar]
  29. Mechling, J.M.; Lecomte, A.; Diliberto, C. Relation between cement composition and compressive strength of pure pastes. Cem. Concr. Compos. 2009, 31, 255–262. [Google Scholar] [CrossRef]
  30. Cheng, M.Y.; Prayogo, D.; Wu, Y.W. Novel Genetic Algorithm-based Evolutionary Support Vector Machine for Optimizing High-performance Concrete Mixture. J. Comput. Civ. Eng. 2014, 28, 06014003. [Google Scholar] [CrossRef]
  31. Taffese, W.Z.; Sistonen, E. Machine learning for durability and service-life assessment of reinforced concrete structures: Recent advances and future directions. Autom. Constr. 2017, 77, 1–14. [Google Scholar] [CrossRef]
  32. Deng, F.; He, Y.; Zhou, S.; Yu, Y.; Cheng, H.; Wu, X. Compressive strength prediction of recycled concrete based on deep learning. Constr. Build. Mater. 2018, 175, 562–569. [Google Scholar] [CrossRef]
  33. Abuodeh, O.R.; Abdalla, J.A.; Hawileh, R.A. Assessment of compressive strength of ultra-high performance concrete using deep machine learning techniques. Appl. Soft Comput. 2020, 95, 106552. [Google Scholar] [CrossRef]
  34. Rubene, S.; Vilnitis, M.; Noviks, J. Frequency analysis and measurements of moisture content of AAC masonry constructions by EIS. Procedia Eng. 2015, 123, 471–478. [Google Scholar] [CrossRef]
  35. Usmen, M.A.; Vilnitis, M. Evaluation of safety, quality and productivity in construction. IOP Conf. Ser. Mater. Sci. Eng. 2015, 96, 012061. [Google Scholar] [CrossRef]
  36. Barbosa, A.A.R.; Vilnītis, M. Innovation and construction management in Brazil: Challenges of companies in times of quality and productivity. IOP Conf. Ser. Mater. Sci. Eng. 2017, 251, 012040. [Google Scholar] [CrossRef]
  37. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  38. Hang, L. Statistical Learning Method; Tsinghua University Press: Beijing, China, 2012. [Google Scholar]
  39. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  40. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  41. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
  42. Freund, Y.; Schapire, R.E. A short introduction to boosting. J.-Jpn. Soc. Artif. Intell. 1999, 14, 771–780. [Google Scholar]
  43. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  44. Schapire, R.E.; Singer, Y. Improved boosting algorithms using confidence-rated predictions. Mach. Learn. 1999, 37, 297–336. [Google Scholar] [CrossRef]
  45. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar]
  46. Schapire, R.E. The strength of weak learnability. Mach. Learn. 1990, 5, 197–227. [Google Scholar] [CrossRef]
  47. Bentéjac, C.; Csörgő, A.; Martínez-Muñoz, G. A comparative analysis of gradient boosting algorithms. Artif. Intell. Rev. 2021, 54, 1937–1967. [Google Scholar] [CrossRef]
  48. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  49. Lundberg, S.M.; Erion, G.G.; Lee, S.I. Consistent individualized feature attribution for tree ensembles. arXiv 2018, arXiv:1802.03888. [Google Scholar]
  50. Molnar, C. Interpretable Machine Learning; Lulu Press: Morrisville, NC, USA, 2020. [Google Scholar]
Figure 1. SHAP attributes.
Figure 1. SHAP attributes.
Materials 17 03661 g001
Figure 2. Flow chart of prediction model of concrete’s compressive strength.
Figure 2. Flow chart of prediction model of concrete’s compressive strength.
Materials 17 03661 g002
Figure 3. Statistical distributions of the input variables.
Figure 3. Statistical distributions of the input variables.
Materials 17 03661 g003
Figure 4. Scatter plot of compressive strength distribution.
Figure 4. Scatter plot of compressive strength distribution.
Materials 17 03661 g004
Figure 5. Bar graph distribution of the χ.
Figure 5. Bar graph distribution of the χ.
Materials 17 03661 g005
Figure 6. Importance of each input feature.
Figure 6. Importance of each input feature.
Materials 17 03661 g006
Figure 7. Trends in the evolution of features for important variables.
Figure 7. Trends in the evolution of features for important variables.
Materials 17 03661 g007
Figure 8. Interpretation of single-sample feature.
Figure 8. Interpretation of single-sample feature.
Materials 17 03661 g008
Figure 9. Heatmap of variable correlation.
Figure 9. Heatmap of variable correlation.
Materials 17 03661 g009
Table 1. Scope of important features of the database.
Table 1. Scope of important features of the database.
VariablesUnitMinMaxMeanMedianSD
Water–Binder Ratio-0.350.550.440.450.06
Waterkg/L124.00154.00140.30140.005.78
Sand ratio%37.0043.0039.2539.001.59
Superplasticizer%0.500.750.520.500.06
Air-entraining agent0.201.500.450.500.28
Slumpmm163202177.87180.008.10
Air content%1.206.404.084.001.00
Aged7.0090.0044.2128.0035.95
Compressive strengthMPa17.1061.4035.6935.859.58
Notes: SD = Standard deviation.
Table 2. Algorithm hyperparameter optimization.
Table 2. Algorithm hyperparameter optimization.
AlgorithmHyperparameter Optimization
KNNn_neighbors = 5
DTcriterion = ‘mse’, splitter = ‘best’, min_samples_split = 2
RFn_estimators = 68, random_state = 90, max_depth = 12
GBDTlearning_rate = 0.2, n_estimators = 5, max_depth = 3
Adaboostmax_depth = 19, learning_rate = 0.9,n_estimators = 40
XGBoostn_estimators = 42, max_depth = 5, gamma = 0.2, learning_rate = 0.2
Table 3. Model performance on training and testing sets.
Table 3. Model performance on training and testing sets.
ModelsSetsMeasures
R2RMSEMAEMAPE(%)
KNNTraining0.8483.6992.9448.291
Testing0.7255.0794.33412.133
DTTraining0.9821.2660.9292.621
Testing0.9432.3081.7354.802
RFTraining0.9801.3371.0072.858
Testing0.9502.1671.6284.409
GBDTTraining0.9791.3711.0252.906
Testing0.9562.0341.5344.138
AdaBoostTraining0.9801.3290.9702.741
Testing0.9402.3681.8115.038
XGBoostTraining0.9821.2660.9292.622
Testing0.9662.3071.7344.801
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, W.; Zhong, Y.; Liao, G.; Ding, Q.; Zhang, T.; Li, X. Prediction of Compressive Strength of Concrete Specimens Based on Interpretable Machine Learning. Materials 2024, 17, 3661. https://doi.org/10.3390/ma17153661

AMA Style

Wang W, Zhong Y, Liao G, Ding Q, Zhang T, Li X. Prediction of Compressive Strength of Concrete Specimens Based on Interpretable Machine Learning. Materials. 2024; 17(15):3661. https://doi.org/10.3390/ma17153661

Chicago/Turabian Style

Wang, Wenhu, Yihui Zhong, Gang Liao, Qing Ding, Tuan Zhang, and Xiangyang Li. 2024. "Prediction of Compressive Strength of Concrete Specimens Based on Interpretable Machine Learning" Materials 17, no. 15: 3661. https://doi.org/10.3390/ma17153661

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop