Machine Learning Models for Early Warning of Coastal Flooding and Storm Surges

  • Abstract
  • Literature Map
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

Floods and storm surges pose significant threats to coastal regions worldwide, demanding timely and accurate early warning systems (EWS) for disaster preparedness. Traditional numerical and statistical methods often fall short in capturing complex, nonlinear, and real-time environmental dynamics. In recent years, machine learning (ML) and deep learning (DL) techniques have emerged as promising alternatives for enhancing the accuracy, speed, and scalability of EWS. This review critically evaluates the evolution of ML models—such as Artificial Neural Networks (ANN), Convolutional Neural Networks (CNN), and Long Short-Term Memory (LSTM)—in coastal flood prediction, highlighting their architectures, data requirements, performance metrics, and implementation challenges. A unique contribution of this work is the synthesis of real-time deployment challenges including latency, edge-cloud tradeoffs, and policy-level integration, areas often overlooked in prior literature. Furthermore, the review presents a comparative framework of model performance across different geographic and hydrologic settings, offering actionable insights for researchers and practitioners. Limitations of current AI-driven models, such as interpretability, data scarcity, and generalization across regions, are discussed in detail. Finally, the paper outlines future research directions including hybrid modelling, transfer learning, explainable AI, and policy-aware alert systems. By bridging technical performance and operational feasibility, this review aims to guide the development of next-generation intelligent EWS for resilient and adaptive coastal management.

Similar Papers
  • Research Article
  • Cite Count Icon 14
  • 10.1371/journal.pone.0317619
Evaluating Machine Learning and Deep Learning models for predicting Wind Turbine power output from environmental factors.
  • Jan 23, 2025
  • PloS one
  • Montaser Abdelsattar + 4 more

This study presents a comprehensive comparative analysis of Machine Learning (ML) and Deep Learning (DL) models for predicting Wind Turbine (WT) power output based on environmental variables such as temperature, humidity, wind speed, and wind direction. Along with Artificial Neural Network (ANN), Long Short-Term Memory (LSTM), Recurrent Neural Network (RNN), and Convolutional Neural Network (CNN), the following ML models were looked at: Linear Regression (LR), Support Vector Regressor (SVR), Random Forest (RF), Extra Trees (ET), Adaptive Boosting (AdaBoost), Categorical Boosting (CatBoost), Extreme Gradient Boosting (XGBoost), and Light Gradient Boosting Machine (LightGBM). Using a dataset of 40,000 observations, the models were assessed based on R-squared, Mean Absolute Error (MAE), and Root Mean Square Error (RMSE). ET achieved the highest performance among ML models, with an R-squared value of 0.7231 and a RMSE of 0.1512. Among DL models, ANN demonstrated the best performance, achieving an R-squared value of 0.7248 and a RMSE of 0.1516. The results show that DL models, especially ANN, did slightly better than the best ML models. This means that they are better at modeling non-linear dependencies in multivariate data. Preprocessing techniques, including feature scaling and parameter tuning, improved model performance by enhancing data consistency and optimizing hyperparameters. When compared to previous benchmarks, the performance of both ANN and ET demonstrates significant predictive accuracy gains in WT power output forecasting. This study's novelty lies in directly comparing a diverse range of ML and DL algorithms while highlighting the potential of advanced computational approaches for renewable energy optimization.

  • Research Article
  • Cite Count Icon 2
  • 10.1007/s43832-025-00207-z
A review on the applications of machine learning and deep learning to groundwater salinity modeling: present status, challenges, and future directions
  • Feb 27, 2025
  • Discover Water
  • Dilip Kumar Roy + 4 more

Coastal aquifers are vital for sustaining ecosystems and providing freshwater for agriculture and domestic use. However, rising groundwater salinity in these regions demands innovative approaches for accurate modeling and prediction. In recent decades, machine learning (ML) and deep learning (DL) techniques have been extensively used across various water resource management fields, often yielding promising outcomes. This review examines the growing usage of ML and DL techniques in modeling groundwater salinity in coastal aquifers. The analysis is based on 104 peer-reviewed journal articles including the review articles indexed in PubMed, ScienceDirect, Google Scholar, Scopus, and SpringerLink using keywords like "saltwater intrusion", "machine learning", “deep learning”, and "groundwater salinity". The review discusses recent advancements, challenges, and future directions for improving ML model accuracy and applicability. Studies reveal an increasing reliance on ML techniques to predict groundwater salinity and manage seawater intrusion across various coastal geological settings. Researchers have employed a range of ML models, from traditional methods like Artificial Neural Networks and Adaptive Neuro Fuzzy Inference Systems to advanced DL-based models like Convolutional Neural Networks and Long Short-Term Memory Networks, often using ensemble methods to enhance accuracy. Key findings include the top-tier output of ensemble and optimized models, including Random Forest and Grasshopper Optimization Algorithm-XGBoost. These studies highlight the importance of model selection, input variable ranking, and computational efficiency, demonstrating ML's global relevance in environmental management. The review highlights the prospects of ML, DL, and ensemble models to redefine groundwater management in coastal aquifers, offering more effective and sustainable solutions to the challenges posed by groundwater salinity and saltwater intrusion. Future advancements will focus on further integrating DL, explainable artificial intelligence, and ensemble approaches, along with other innovative techniques, to explore sparsely studied variables, model new or unique study areas, and apply ML methods for managing groundwater salinity.

  • Research Article
  • 10.1029/2024jh000507
Machine Learning Techniques for Predicting Typhoon‐Induced Storm Surge Using a Hybrid Wind Field
  • Apr 10, 2025
  • Journal of Geophysical Research: Machine Learning and Computation
  • Changyu Su + 3 more

Accurate and timely storm surge prediction is critical information in coastal zone management and risk reduction strategies. The Bohai Sea, a semi‐enclosed bay in the Northwest Pacific that used to be less prone to typhoon disasters, has been witnessing a paradigm shift in typhoon activities in the recent past. Since there have been limited typhoon‐induced storm surges in the Bohai Sea, an innovative prediction system is warranted to address frequent and intense typhoon‐induced impacts. Four Machine Learning (ML) models (Long Short‐Term Memory (LSTM), Convolutional Neural Networks (CNN), CNN‐LSTM, and ConvLSTM) were built to predict storm surges and significantly improve prediction when combined with a three‐dimensional Finite Volume Community Ocean Model (FVCOM), that is, FVCOM‐ML. In this study, the FVCOM‐ML model was driven by a hybrid wind field that superimposed the Holland wind and the reanalysis wind field. The ML models were trained via Advanced Circulation Model simulations to compensate for the limited in‐situ observations. The prediction performances were analyzed for both spatial (e.g., single and multiple sites) and temporal (e.g., single and multiple steps) scale variability. ML is trained to overcome the residual error of the FVCOM, effectively reducing the inherent uncertainty of traditional methods. FVCOM‐ML offers a significant advantage over standalone FVCOM or ML while better incorporating realistic physical constraints and improving the accuracy of storm surge forecasts.

  • Research Article
  • Cite Count Icon 3
  • 10.13031/jnrae.15647
A Review of Machine Learning Models for Harmful Algal Bloom Monitoring in Freshwater Systems
  • Jan 1, 2023
  • Journal of Natural Resources and Agricultural Ecosystems
  • Ibrahim Busari + 3 more

Highlights Machine Learning (ML) models are identified, reviewed, and analyzed for HAB predictions. Data preprocessing is vital for efficient ML model development. ML models for toxin production and monitoring are limited. Abstract. Harmful algal blooms (HABs) are detrimental to livestock, humans, pets, the environment, and the global economy, which calls for a robust approach to their management. While process-based models can inform practitioners about HAB enabling conditions, they have inherent limitations in accurately predicting harmful algal blooms. To address these limitations, Machine Learning (ML) models can potentially leverage large volumes of IoT data to aid in near real-time predictions. ML models have evolved as efficient tools for understanding patterns and relationships between water quality parameters and HAB expansion. This review describes ML models currently used for predicting and forecasting HABs in freshwater ecosystems and presents model structures and their application for predicting algal parameters and related toxins. The review revealed that regression trees, random forest, Artificial Neural Network (ANN), Support Vector Regression (SVR), Long Short-Term Memory (LSTM), and Gated Recurrent Unit (GRU) are the most frequently used models for HABs monitoring. This review shows ML models' prowess in identifying significant variables influencing algal growth, HAB drivers, and multistep HAB prediction. Hybrid models also improve the prediction of algal-related parameters through improved optimization techniques and variable selection algorithms. While ML models often focus on algal biomass prediction, few studies apply ML models for toxin monitoring and prediction. This limitation can be associated with a lack of high-frequency toxin datasets for model development, and exploring this domain is encouraged. This review serves as a guide for policymakers and researchers to implement ML models for HAB prediction and reveals the potential of ML models for decision support and early prediction for HAB management. Keywords: Cyanobacteria, Freshwater, Harmful algal blooms, Machine learning, Water quality.

  • Research Article
  • 10.21015/vtse.v12i4.1928
Leveraging Machine Learning And Deep Learning Models for Proactive Churn Customer Retention
  • Nov 23, 2024
  • VFAST Transactions on Software Engineering
  • Hira Farman + 5 more

Customer attrition is especially an issue in industries such as retail, banking, and telecommunications where customer acquisition costs are significantly higher than the costs of retaining repeat customers. The customer lack of interest is now predictable through machine learning models, and deep learning has become instrumental in early intervention for retention. In order to assess the quality of churn prediction, the study tests six basic machine learning techniques: random forest, logistic regression, and the k-nearest neighbors method, as well as four deep learning techniques: long short term memory (LSTM), bidirectional LSTM, convolutional neural networks (CNN), and artificial neural networks (ANN). The performance of the model is then assessed via the evaluation matrices, including the accuracy, precision, recall, and F1-score from the customer's behavioral data after feature extraction from large datasets. The study reveals that DL models offer improved handling of the churn and non-churn customer classification and Random Forest as well as other ML models comparable accuracy. This research can conclude that LSTM and ANN models outshine in actual-world churn prediction circumstances, especially when long-term consumer behavior evaluation is required. To enhance the current outcomes of a given prediction model, this research focuses on data preprocessing and the utilization of bootstrapping, feature extraction, and the combination of multiple models. The implications of the study provide specific practical recommendations for firms to effectively manage customer churn and increase customer retention by employing data-dealing techniques.

  • Research Article
  • Cite Count Icon 11
  • 10.1186/s44147-023-00307-4
Machine learning algorithm and neural network architecture for optimization of pharmaceutical and drug manufacturing industrial effluent treatment using activated carbon derived from breadfruit (Treculia africana)
  • Nov 14, 2023
  • Journal of Engineering and Applied Science
  • Prosper Eguono Ovuoraye + 6 more

In a recent development, attention has shifted to the application of artificial intelligence for the optimization of wastewater treatment processes. This research compared the performances of the machine learning (ML) model: random forest, decision tree, support vector machine, artificial neural network, convolutional neural network, long-short term memory, and multiple linear regressors for optimization in effluent treatment. The training, testing, and validation datasets were obtained via the design of an experiment conducted on the removal of total dissolved solids (TDS) from pharmaceutical effluent. The breadfruit-activated carbon (BFAC) adsorbent was characterized using scanning electron microscopy and X-ray diffraction techniques. The predictive capacity of an ML algorithm, and neural network architecture implemented to optimize the treatment process using statistical metrics. The results showed that MSE ≤ 1.68, MAE ≤ 0.95, and predicted-R2 ≥ 0.9035 were recorded across all ML. The ML output with minimum error functions that satisfied the criterion for clean discharge was adopted. The predicted optimum conditions correspond to BFAC dosage, contact time, particle size, and pH of 2.5 mg/L, 10 min, 0.60 mm, and 6, respectively. The optimum transcends to a reduction in TDS concentration from 450 mg/L to a residual ≤ 40 mg/L and corresponds to 90% removal efficiency, indicating ± 1.01 standard deviation from the actual observation practicable. The findings established the ML model outperformed the neural network architecture and affirmed validation for the optimization of the adsorption treatment in the pharmaceutical effluent domain. Results demonstrated the reliability of the selected ML algorithm and the feasibility of BFAC for use in broad-scale effluent treatment.

  • Research Article
  • Cite Count Icon 7
  • 10.1016/j.jhydrol.2023.130350
Value of process understanding in the era of machine learning: A case for recession flow prediction
  • Oct 21, 2023
  • Journal of Hydrology
  • Prashant Istalkar + 2 more

Value of process understanding in the era of machine learning: A case for recession flow prediction

  • Preprint Article
  • Cite Count Icon 1
  • 10.5194/egusphere-egu23-13083
Machine learning modelling of compound flood events
  • May 15, 2023
  • Agnieszka Indiana Olbert + 2 more

Flood early warning systems are vital for preventing flood damages and for reducing disaster risks. Such systems are particularly important for forecasting compound events where multiple, often dependent flood drivers co-occur and interact. In this research an early warning system for prediction of coastal-fluvial floods is developed to provide a robust, cost-effective and time-efficient framework for management of flood risks and impacts. This three-step method combines a cascade of three linked models: (1) statistical model that determines probabilities of multiple-driver flood events, (2) hydrodynamic model forced by outputs from the statistical model, and finally (3) machine learning (ML) model that uses hydrodynamic outputs from various probability flood events to train the ML algorithm in order to predict the spatially and temporarily variable inundation patterns resulting from a combination of coastal and fluvial flood drivers occurring simultaneously.The method has been utilized for the case of Cork City, located in the south-west of Ireland, which has a long history of fluvial-coastal flooding. The Lee  River channelling through the city centre may generate a substantial flood when the downstream river flow draining to the estuary coincides with the sea water propagating upstream on a flood tide. For this hydrological domain the statistical model employs the univariate extreme values analysis and copula functions to calculate joint probabilities of river discharges and sea water levels (astronomical tides and surge residuals) occurring simultaneously. The return levels for these two components along a return level curve produced by the copula function are used to generate synthetic timeseries, which serve as water level boundary conditions for a hydrodynamic flood model. The multi-scale nested flood model (MSN_Flood) was configured for Cork City at 2m resolution to simulate an unsteady, non-uniform flow in the Lee  River and a flood wave propagation over urban floodplains. The ensemble hydrodynamic model outputs are ultimately used to train and test a range machine learning models for prediction of flood extents and water depths. In total, 23 machine learning algorithms including: Artificial Neural Network, Decision Tree, Gaussian Process Regression, Linear Regression, Radial Basis Function, Support Vector Machine, and Support Vector Regression were employed to confirm that the ML algorithm can be used successfully to predict the flood inundation depths over urban floodplains for a given set of compound flood drivers. Here, the limited flood conditioning factors taken into account to analyse floods are the upstream flood hydrographs and downstream sea water level timeseries. To evaluate model performance, different statistical skill scores were computed. Results indicated that in most pixels, the Gaussian Process Regression model performs better than the other models.The main contribution of this research is to demonstrate the ML models can be used in early warning systems for flood prediction and to give insight into the most suitable models in terms of robustness, accuracy, effectiveness, and speed. The findings demonstrate that ML models do help in flood water propagation mapping and assessment of flood risk under various compound flood scenarios.

  • Abstract
  • 10.1182/blood-2024-212217
Systematic Review of Machine Learning in Diagnosis of Myeloproliferative Neoplasia
  • Nov 5, 2024
  • Blood
  • Karna Desai + 5 more

Systematic Review of Machine Learning in Diagnosis of Myeloproliferative Neoplasia

  • Research Article
  • Cite Count Icon 119
  • 10.1007/s11356-021-17084-3
Prediction of irrigation groundwater quality parameters using ANN, LSTM, and MLR models.
  • Nov 8, 2021
  • Environmental Science and Pollution Research
  • Saber Kouadri + 4 more

Forecasting the irrigation groundwater parameters helps plan irrigation water and crop, and it is commonly expensive because it needs various parameters, mainly in developing nations. Therefore, the present research's core objective is to create accurate and reliable machine learning models for irrigation parameters. To accomplish this determination, three machine learning (ML) models, viz. long short-term memory (LSTM), multi-linear regression (MLR), and artificial neural network (ANN), have been trained. It is validated with mean squared error (MSE) and correlation coefficients (r), root mean square error (RMSE), and mean absolute error (MAE). These machine learning models have been used and applied for predicating the six irrigation water quality parameters such as sodium absorption ratio (SAR), percentage of sodium (%Na), residual sodium carbonate (RSC), magnesium hazard (MH), Permeability Index (PI), and Kelly ratio (KR). Therefore, the two scenario performances of ANN, LSTM, and MLR have been developed for each model to predict irrigation water quality parameters. The first and second scenario performance was created based on all and second reduction input variables. The ANN, LSTM, and MLR models have discovered that excluding for ANN and MLR models shows high accuracy in first and second scenario models, respectively. These model's accuracy was checked based on the mean squared error (MSE), correlation coefficients (r), and root mean square error (RMSE) for training and testing processes serially. The RSC values are highly accurate predicated values using ANN and MLR models. As a result, machine learning models may improve irrigation water quality parameters, and such types of results are essential to farmers and crop planning in various irrigation processes.

  • Conference Article
  • Cite Count Icon 2
  • 10.1109/iscon57294.2023.10112002
Heart Disease Evaluation with Deep Learning and Machine Learning
  • Mar 3, 2023
  • Sachin Upadhyay + 3 more

All over the world is affected by the disease which is named as heart disease. The main reason behind the heart disease is our busy life, as the person is affected not only by office work but also by personal problems. This mortality rate is too high. We can predict this disease with the help of Machine Learning (ML) and Deep Learning (DL) prediction models. In this paper, to reach the accuracy we worked on three ML & DL models. For this paper we use ML models named as: SVM (Support Vector Machine), LR (Logistic Regression) & Naïve Bayes. DL models named as: CNN (Convolutional Neural Network), RNN (Recurrent Neural Network) & LSTM (Long Short Term Memory). The accuracy obtained in this study is made up of the 85% accuracy of Logistic Regression, the 89% accuracy of SVM, and the 85% accuracy of Naive Bayes. LSTM has an accuracy of 83%, RNN has an accuracy of 91%, and CNN has an accuracy of 83%. The study’s findings indicate that the RNN model is the most accurate, coming in at 90%, and from this we can say that it is the best at predicting heart disease.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 6
  • 10.3390/en16114271
A Prediction Model for Spot LNG Prices Based on Machine Learning Algorithms to Reduce Fluctuation Risks in Purchasing Prices
  • May 23, 2023
  • Energies
  • Sun-Feel Yang + 2 more

The ongoing Russia–Ukraine conflict has exacerbated the global crisis of natural gas supply, particularly in Europe. During the winter season, major importers of liquefied natural gas (LNG), such as South Korea and Japan, were directly affected by fluctuating spot LNG prices. This study aimed to use machine learning (ML) to predict the Japan Korea Marker (JKM), a spot LNG price index, to reduce price fluctuation risks for LNG importers such as the Korean Gas Corporation (KOGAS). Hence, price prediction models were developed based on long short-term memory (LSTM), artificial neural network (ANN), and support vector machine (SVM) algorithms, which were used for time series data prediction. Eighty-seven variables were collected for JKM prediction, of which eight were selected for modeling. Four scenarios (scenarios A, B, C, and D) were devised and tested to analyze the effect of each variable on the performance of the models. Among the eight variables, JKM, national balancing point (NBP), and Brent price indexes demonstrated the largest effects on the performance of the ML models. In contrast, the variable of LNG import volume in China had the least effect. The LSTM model showed a mean absolute error (MAE) of 0.195, making it the best-performing algorithm. However, the LSTM model demonstrated a decreased in performance of at least 57% during the COVID-19 period, which raises concerns regarding the reliability of the test results obtained during that time. The study compared the ML models’ prediction performances with those of the traditional statistical model, autoregressive integrated moving averages (ARIMA), to verify their effectiveness. The comparison results showed that the LSTM model’s performance deviated by an MAE of 15–22%, which can be attributed to the constraints of the small dataset size and conceptual structural differences between the ML and ARIMA models. However, if a sufficiently large dataset can be secured for training, the ML model is expected to perform better than the ARIMA. Additionally, separate tests were conducted to predict the trends of JKM fluctuations and comprehensively validate the practicality of the ML models. Based on the test results, LSTM model, identified as the optimal ML algorithm, achieved a performance of 53% during the regular period and 57% d during the abnormal period (i.e., COVID-19). Subject matter experts agreed that the performance of the ML models could be improved through additional studies, ultimately reducing the risk of price fluctuations when purchasing spot LNG.

  • Research Article
  • Cite Count Icon 4
  • 10.1016/j.egyai.2024.100398
Deep learning from three-dimensional Lithium-ion battery multiphysics model Part II: Convolutional neural network and long short-term memory integration
  • Jul 18, 2024
  • Energy and AI
  • Yiheng Pang + 3 more

Deep learning from three-dimensional Lithium-ion battery multiphysics model Part II: Convolutional neural network and long short-term memory integration

  • Research Article
  • Cite Count Icon 1
  • 10.21926/aeer.2404020
Comparative Analysis of Machine Learning Models and Explainable Artificial Intelligence for Predicting Wastewater Treatment Plant Variables
  • Oct 17, 2024
  • Advances in Environmental and Engineering Research
  • Fuad Bin Nasir + 1 more

Increasing urban wastewater and rigorous discharge regulations pose significant challenges for wastewater treatment plants (WWTP) to meet regulatory compliance while minimizing operational costs. This study explores the application of several machine learning (ML) models specifically, Artificial Neural Networks (ANN), Gradient Boosting Machines (GBM), Random Forests (RF), eXtreme Gradient Boosting (XGBoost), and hybrid RF-GBM models in predicting important WWTP variables such as Biochemical Oxygen Demand (BOD), Total Suspended Solids (TSS), Ammonia (NH₃), and Phosphorus (P). Several feature selection (FS) methods were employed to identify the most influential WWTP variables. To enhance ML models’ interpretability and to understand the impact of variables on prediction, two widely used explainable artificial intelligence (XAI) methods-Local Interpretable Model-Agnostic Explanations (LIME) and SHapley Additive exPlanations (SHAP) were investigated in the study. Results derived from FS and XAI methods were compared to explore their reliability. The ML model performance results revealed that ANN, GBM, XGBoost, and RF-GBM have great potential for variable prediction with low error rates and strong correlation coefficients such as R<sup>2</sup> value of 1 on the training set and 0.98 on the test set. The study also revealed that XAI methods identify common influential variables in each model’s prediction. This is a novel attempt to get an overview of both LIME and SHAP explanations on ML models for a WWTP variable prediction.

  • Research Article
  • 10.3389/fcvm.2025.1568907
Predicting mortality in critically ill patients with hypertension using machine learning and deep learning models
  • Aug 8, 2025
  • Frontiers in Cardiovascular Medicine
  • Ziyang Zhang + 1 more

BackgroundAccurate prediction of mortality in critically ill patients with hypertension admitted to the Intensive Care Unit (ICU) is essential for guiding clinical decision-making and improving patient outcomes. Traditional prognostic tools often fall short in capturing the complex interactions between clinical variables in this high-risk population. Recent advances in machine learning (ML) and deep learning (DL) offer the potential for developing more sophisticated and accurate predictive models.ObjectiveThis study aims to evaluate the performance of various ML and DL models in predicting mortality among critically ill patients with hypertension, with a particular focus on identifying key clinical predictors and assessing the comparative effectiveness of these models.MethodsWe conducted a retrospective analysis of 30,096 critically ill patients with hypertension admitted to the ICU. Various ML models, including logistic regression, decision trees, and support vector machines, were compared with advanced DL models, including 1D convolutional neural networks (CNNs) and long short-term memory (LSTM) networks. Model performance was evaluated using area under the receiver operating characteristic curve (AUC) and other performance metrics. SHapley Additive exPlanations (SHAP) values were used to interpret model outputs and identify key predictors of mortality.ResultsThe 1D CNN model with an initial selection of predictors achieved the highest AUC (0.7744), outperforming both traditional ML models and other DL models. Key clinical predictors of mortality identified across models included the APS-III score, age, and length of ICU stay. The SHAP analysis revealed that these predictors had a substantial influence on model predictions, underscoring their importance in assessing mortality risk in this patient population.ConclusionDeep learning models, particularly the 1D CNN, demonstrated superior predictive accuracy compared to traditional ML models in predicting mortality among critically ill patients with hypertension. The integration of these models into clinical workflows could enhance the early identification of high-risk patients, enabling more targeted interventions and improving patient outcomes. Future research should focus on the prospective validation of these models and the ethical considerations associated with their implementation in clinical practice.

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.