Cargando…
Exploring machine learning algorithms for accurate water level forecasting in Muda river, Malaysia
Accurate water level prediction for both lake and river is essential for flood warning and freshwater resource management. In this study, three machine learning algorithms: multi-layer perceptron neural network (MLP-NN), long short-term memory neural network (LSTM) and extreme gradient boosting XGBo...
Autores principales: | , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Elsevier
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10344711/ https://www.ncbi.nlm.nih.gov/pubmed/37456046 http://dx.doi.org/10.1016/j.heliyon.2023.e17689 |
Sumario: | Accurate water level prediction for both lake and river is essential for flood warning and freshwater resource management. In this study, three machine learning algorithms: multi-layer perceptron neural network (MLP-NN), long short-term memory neural network (LSTM) and extreme gradient boosting XGBoost were applied to develop water level forecasting models in Muda River, Malaysia. The models were developed using limited amount of daily water level and meteorological data from 2016 to 2018. Different input scenarios were tested to investigate the performance of the models. The results of the evaluation showed that the MLP model outperformed both the LSTM and the XGBoost models in predicting water levels, with an overall accuracy score of 0.871 compared to 0.865 for LSTM and 0.831 for XGBoost. No noticeable improvement has been achieved after incorporating meteorological data into the models. Even though the lowest reported performance was reported by the XGBoost, it is the faster of the three algorithms due to its advanced parallel processing capabilities and distributed computing architecture. In terms of different time horizons, the LSTM model was found to be more accurate than the MLP and XGBoost model when predicting 7 days ahead, demonstrating its superiority in capturing long-term dependencies. Therefore, it can be concluded that each ML model has its own merits and weaknesses, and the performance of different ML models differs on each case because these models depends largely on the quantity and quality of data available for the model training. |
---|