Cargando…

Research on prediction of compressive strength of fly ash and slag mixed concrete based on machine learning

Every year, a large amount of solid waste such as fly ash and slag is generated worldwide. If these solid wastes are used in concrete mixes to make concrete, it can effectively save resources and protect the environment. The compressive strength of concrete is an essential indicator for testing its...

Descripción completa

Detalles Bibliográficos
Autores principales: Wang, Meng, Kang, Jiaxu, Liu, Weiwei, Su, Jinshuai, Li, Meng
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9794082/
https://www.ncbi.nlm.nih.gov/pubmed/36574382
http://dx.doi.org/10.1371/journal.pone.0279293
Descripción
Sumario:Every year, a large amount of solid waste such as fly ash and slag is generated worldwide. If these solid wastes are used in concrete mixes to make concrete, it can effectively save resources and protect the environment. The compressive strength of concrete is an essential indicator for testing its quality, and its prediction is affected by many factors. It is difficult to predict its strength accurately. Therefore, based on the current popular machine learning supervised learning algorithms: Random Forest (RF), Extreme Gradient Boosting (XGBoost), and Support Vector Machine (SVR), three models established a nonlinear mapping between multi-factor features and target feature concrete compressive strength. Using the three completed training models, we validated the test set with 206 example sets, and the Root Mean Square Error (RMSE), fitting coefficient (R(2)), and Mean Absolute Error (MAE) were used as evaluation metrics. The validation results showed that the values of RMSE, R(2), and MAE for the RF model were 0.1, 0.9, and 0.21, respectively; the values of XGBoost model were 0.05, 0.95, and 0.15, respectively. The values of SVR were 0.15, 0.86, and 0.3, respectively. As a result, Extreme Gradient Boosting (XGBoost) has better generalization ability and prediction accuracy than the other two algorithms.