Distribution of concentration in hybrid computation and finite element method in assessment of the flow properties in the two-phase contactor
Distribution of concentration in hybrid computation and finite element method in assessment of the flow properties in the two-phase contactor
- Research Article
2
- 10.13031/jnrae.15812
- Jan 1, 2024
- Journal of Natural Resources and Agricultural Ecosystems
Highlights The monitoring of HABs can be improved using ML models for chlorophyll-a prediction. ML model selection for HABs monitoring depends on target objectives. Random forest model predicts chlorophyll-a better when the temporal dimension is not considered. The LSTM model is essential for making time-dependent chlorophyll-a predictions for HABs monitoring. Abstract. The complex dynamics of freshwater harmful algal blooms (HABs) necessitate proactive monitoring approaches to mitigate their impacts. The rapid breakthrough in computing prowess and statistical advances is triggering the development of data-driven techniques such as machine learning (ML) models, which have been shown in different fields to be instrumental in finding patterns for explaining relationships in observed data. This study assesses the ability of ML models for HABs monitoring in a lake using chlorophyll-a concentration as the index. The selected models for this study were regression tree, random forest (RF), multilayer perceptron (MLP), support vector regression (SVR), long short-term memory (LSTM), and gated recurrent unit (GRU) models, with the last two models able to consider the temporal sequence of obtained water quality datasets. The results showed that the RF model with R2, mean absolute error (MAE), and root mean square error (RMSE) of 0.87 µgL-1, 0.97 µgL-1, and 3.53 µgL-1, respectively, outperformed the SVR, MLP, and regression tree models. LSTM model with MAE and RMSE of 2.39 µgL-1 and 3.29 µgL-1, respectively, predicted temporal dynamics of chlorophyll-a better than GRU, although with more runtime, and showed the potential for developing real-time HAB monitoring and early warning systems. The findings reveal the robustness of the chosen ML models, thereby shedding light on crucial factors that necessitate careful deliberation by researchers and policymakers in determining the most suitable approaches for monitoring HABs. Keywords: Cyanobacteria, Early warning systems, Freshwater, HABs, Machine learning models.
- Research Article
24
- 10.3390/ma15134436
- Jun 23, 2022
- Materials
Bacterial-based self-healing concrete (BSHC) is a well-known healing technology which has been investigated for a few decades for its excellent crack healing capacity. Nevertheless, considered as costly and time-consuming, the healing performance (HP) of concrete with various types of bacteria can be designed and evaluated only in laboratory environments. Employing machine learning (ML) models for predicting the HP of BSHC is inspired by practical applications using concrete mechanical properties. The HP of BSHC can be predicted to save the time and cost of laboratory tests, bacteria selection and healing mechanisms adoption. In this paper, three types of BSHC, including ureolytic bacterial healing concrete (UBHC), aerobic bacterial healing concrete (ABHC) and nitrifying bacterial healing concrete (NBHC), and ML models with five kinds of algorithms consisting of the support vector regression (SVR), decision tree regression (DTR), deep neural network (DNN), gradient boosting regression (GBR) and random forest (RF) are established. Most importantly, 22 influencing factors are first employed as variables in the ML models to predict the HP of BSHC. A total of 797 sets of BSHC tests available in the open literature between 2000 and 2021 are collected to verify the ML models. The grid search algorithm (GSA) is also utilised for tuning parameters of the algorithms. Moreover, the coefficient of determination (R2) and root mean square error (RMSE) are applied to evaluate the prediction ability, including the prediction performance and accuracy of the ML models. The results exhibit that the GBR model has better prediction ability (R2GBR = 0.956, RMSEGBR = 6.756%) than other ML models. Finally, the influence of the variables on the HP is investigated by employing the sensitivity analysis in the GBR model.
- Research Article
- 10.1149/ma2024-023335mtgabs
- Nov 22, 2024
- Electrochemical Society Meeting Abstracts
As battery electric vehicles (BEVs) become more widespread, collecting large amounts of time-series data on lithium-ion batteries in real-world settings has become possible. In recent years, there has been growing interest in methods that utilize collected battery data and machine learning (ML) models to predict battery performance accurately [1-4].However, there is a challenge in that prediction accuracy declines when the distribution of training data is uneven, and it has been a barrier to practical application. This challenge is due to real-world data often being unevenly distributed, and ML models generally have low extrapolation accuracy [5-6]. A specific challenge for BEVs is that the data in the low state of charge (SOC) region is insufficient because batteries are infrequently discharged to lower SOC due to users' anxieties about running out of electricity, resulting in less prediction accuracy.Here, we studied several representative ML models and their characteristics to achieve high extrapolation accuracy of voltage prediction. We artificially generated data for training and testing using an electrochemical simulation model. To evaluate the extrapolation accuracy of each model, we trained the ML models using data that excluded the low SOC region. We quantified voltage prediction accuracy for the entire SOC region using root mean square error (RMSE). Fig. 1 shows results for four ML models (“Random Forest,” “Gaussian Process,” “LSTM," and “MLP-ReLU"). The results confirmed that accuracy highly depends on the model and that MLP-ReLU (multilayer perceptron with ReLU activation function) has the highest accuracy. For further study, we set up several SOC ranges and comprehensively evaluated the RMSEs of eight representative ML models. Our results confirmed that the MLP-ReLU showed the best accuracy [7]. Furthermore, we examined four representative MLP activation functions (“ReLU,” “identity,” “tanh,” and “logistic”) and confirmed that the ReLU performed the best.This study shows that MLP-ReLU is suitable for building a model with high extrapolation accuracy even when training data is unevenly distributed. From these results, we concluded that we are close to realizing practical applications for predicting battery performance with high accuracy by utilizing real-world battery data and ML models.
- Research Article
2
- 10.3389/feart.2024.1344690
- May 31, 2024
- Frontiers in Earth Science
Extreme weather events and global climate change have exacerbated the problem of evaporation rates. Thus, accurately predicting soil moisture evaporation rates affecting soil cracking becomes crucial. However, less is known about how novel feature engineering techniques and machine-learning predictions may account for estimating the soil moisture evaporation rate. This research focuses on predicting the evaporation rate of soil using machine learning (ML) models. The dataset comprised twenty-one ground-based parameters, including temperature, humidity, and soil-related features, used as features to predict evaporation potential. To tackle the high number of features and potential uncorrelated features, a novel guided backpropagation-based feature selection technique was developed to rank the most relevant features. The top-10 features, highly correlated with evaporation rate, were selected for ML model input, alongside the top-5 and all features. Several ML models, including multiple regression (MR), K-nearest neighbor (KNN), multilayer perceptron (MLP), sequential minimal optimization regression (SMOreg), random forest (RF), and a novel K-Nearest Oracles (KNORA) ensemble, were constructed for the purpose of forecasting the evaporation rate. The average error of these models was assessed using the root mean squared error (RMSE). Experimental results showed that the KNORA ensemble model performed the best, achieving a 7.54 mg/h RMSE in testing with the top-10 features. MLP was followed closely by a 25.1 mg/h RMSE in the same testing. An empirical model using all features showed a higher RMSE of 1319.1 mg/h, indicating the superiority of the ML models for accurate evaporation rate predictions. We highlight the implications of our results for climate-induced soil cracking in the real world.
- Research Article
3
- 10.13031/jnrae.15647
- Jan 1, 2023
- Journal of Natural Resources and Agricultural Ecosystems
Highlights Machine Learning (ML) models are identified, reviewed, and analyzed for HAB predictions. Data preprocessing is vital for efficient ML model development. ML models for toxin production and monitoring are limited. Abstract. Harmful algal blooms (HABs) are detrimental to livestock, humans, pets, the environment, and the global economy, which calls for a robust approach to their management. While process-based models can inform practitioners about HAB enabling conditions, they have inherent limitations in accurately predicting harmful algal blooms. To address these limitations, Machine Learning (ML) models can potentially leverage large volumes of IoT data to aid in near real-time predictions. ML models have evolved as efficient tools for understanding patterns and relationships between water quality parameters and HAB expansion. This review describes ML models currently used for predicting and forecasting HABs in freshwater ecosystems and presents model structures and their application for predicting algal parameters and related toxins. The review revealed that regression trees, random forest, Artificial Neural Network (ANN), Support Vector Regression (SVR), Long Short-Term Memory (LSTM), and Gated Recurrent Unit (GRU) are the most frequently used models for HABs monitoring. This review shows ML models' prowess in identifying significant variables influencing algal growth, HAB drivers, and multistep HAB prediction. Hybrid models also improve the prediction of algal-related parameters through improved optimization techniques and variable selection algorithms. While ML models often focus on algal biomass prediction, few studies apply ML models for toxin monitoring and prediction. This limitation can be associated with a lack of high-frequency toxin datasets for model development, and exploring this domain is encouraged. This review serves as a guide for policymakers and researchers to implement ML models for HAB prediction and reveals the potential of ML models for decision support and early prediction for HAB management. Keywords: Cyanobacteria, Freshwater, Harmful algal blooms, Machine learning, Water quality.
- Research Article
5
- 10.1155/2023/5513446
- Nov 13, 2023
- International Journal of Energy Research
Battery performance prediction techniques based on machine learning (ML) models and lithium-ion battery (LIB) data collected in the real world have received much attention recently. However, poor extrapolation accuracy is a major challenge for ML models using real-world data, as the data frequency distribution can be uneven. Here, we have investigated the extrapolation accuracy of the ML models by using artificial data generated with an electrochemical simulation model. Specifically, we set a lower open circuit voltage (OCV) limit for the training data and generated data limited to the higher state of charge (SOC) region to train the voltage prediction model. We have validated the root mean squared error (RMSE) of the voltage for the test data at several lower OCV limit settings and defined the average + 3 standard deviations of them as an evaluation metric. Eight representative ML models were evaluated, and it was found that the multilayer perceptron (MLP) showed an accuracy of 92.7 mV, which was the best extrapolation accuracy. We also evaluated models with published experimental data and found that the MLP had an accuracy of 102.4 mV, reconfirming that it had the best extrapolation accuracy. We also found that MLP was robust to changes in the data of interest since the accuracy degradation when changing from simulation to experimental data was as small as a factor of 1.1. This result shows that MLP can achieve higher voltage prediction accuracy even when collecting data for comprehensive SOC conditions is difficult.
- Research Article
1
- 10.1016/j.atmosres.2023.107173
- Dec 13, 2023
- Atmospheric Research
Improving the hindcast of the northward shift of South Asian high in June with machine learning
- Research Article
- 10.18502/japh.v10i1.18093
- Mar 9, 2025
- Journal of Air Pollution and Health
Introduction: Air pollution is a significant global health challenge, contributing to the deaths of millions of people annually. Among these pollutants, Particulate Matter (PM2.5) is the most harmful to the respiratory system causing serious health problems. This study focused on predicting PM2.5 in the air of Islamabad, capital of Pakistan by using machine learning and deep learning models. Materials and methods: Two machine learning models (Decision Tree and Random Forest) and four deep learning models including Multi-Layer Neural Network (MLNN), Long Short-Term Memory (LSTM), Recurrent Neural Network (RNN), Gated Recurrent Unit (GRU) are used in the study. Each model's performance was assessed by using statistical indicators including coefficient of determination (R2), Mean Absolute Error (MAE), Root Mean Square Error (RMSE), and Relative Root Mean Square Error (RRMSE). These models are also ranked based on their performance by compromise programming technique. Results: Machine learning models performed better in the training phase by achieving higher R2 values of 0.98 and 0.97 but couldn’t maintain the same performance in the testing phase. Whereas the deep learning models performed best in both the training and testing phases. MLNN model attained higher R2 value of 0.98 in training and 0.88 in testing and is evaluated as top-ranked prediction model in predicting particulate matter PM2.5. Whereas,LSTM, GRU, RNN, Decision Tree, and Random Forest are placed at the 2nd,3rd, 4th, 5th, and 6th positions having R2 values of 0.86, 0.87, 0.82, 0.99, and0.97 during training and 0.71, 0.69, 0.69, 0.75, and 0.85 respectively during testing. Conclusion: Deep learning models, especially MLNN, showed strong performance in predicting PM2.5 as compared to the machine learning models.
- Research Article
20
- 10.2196/47833
- Nov 20, 2023
- JMIR Medical Informatics
Machine learning (ML) models provide more choices to patients with diabetes mellitus (DM) to more properly manage blood glucose (BG) levels. However, because of numerous types of ML algorithms, choosing an appropriate model is vitally important. In a systematic review and network meta-analysis, this study aimed to comprehensively assess the performance of ML models in predicting BG levels. In addition, we assessed ML models used to detect and predict adverse BG (hypoglycemia) events by calculating pooled estimates of sensitivity and specificity. PubMed, Embase, Web of Science, and Institute of Electrical and Electronics Engineers Explore databases were systematically searched for studies on predicting BG levels and predicting or detecting adverse BG events using ML models, from inception to November 2022. Studies that assessed the performance of different ML models in predicting or detecting BG levels or adverse BG events of patients with DM were included. Studies with no derivation or performance metrics of ML models were excluded. The Quality Assessment of Diagnostic Accuracy Studies tool was applied to assess the quality of included studies. Primary outcomes were the relative ranking of ML models for predicting BG levels in different prediction horizons (PHs) and pooled estimates of the sensitivity and specificity of ML models in detecting or predicting adverse BG events. In total, 46 eligible studies were included for meta-analysis. Regarding ML models for predicting BG levels, the means of the absolute root mean square error (RMSE) in a PH of 15, 30, 45, and 60 minutes were 18.88 (SD 19.71), 21.40 (SD 12.56), 21.27 (SD 5.17), and 30.01 (SD 7.23) mg/dL, respectively. The neural network model (NNM) showed the highest relative performance in different PHs. Furthermore, the pooled estimates of the positive likelihood ratio and the negative likelihood ratio of ML models were 8.3 (95% CI 5.7-12.0) and 0.31 (95% CI 0.22-0.44), respectively, for predicting hypoglycemia and 2.4 (95% CI 1.6-3.7) and 0.37 (95% CI 0.29-0.46), respectively, for detecting hypoglycemia. Statistically significant high heterogeneity was detected in all subgroups, with different sources of heterogeneity. For predicting precise BG levels, the RMSE increases with a rise in the PH, and the NNM shows the highest relative performance among all the ML models. Meanwhile, current ML models have sufficient ability to predict adverse BG events, while their ability to detect adverse BG events needs to be enhanced. PROSPERO CRD42022375250; https://www.crd.york.ac.uk/prospero/display_record.php?RecordID=375250.
- Research Article
20
- 10.3390/w14101666
- May 23, 2022
- Water
Accurate estimation of reference evapotranspiration (ETo) plays a vital role in irrigation and water resource planning. The Penman–Monteith method recommended by the Food and Agriculture Organization (FAO PM56) is widely used and considered a standard to calculate ETo. However, FAO PM56 cannot be used with limited meteorological variables, so it is compulsory to choose an alternative model for ETo estimation, which requires fewer variables. This study built ten machine learning (ML) models based on multi-function, neural network, and tree-based structure against the FAO PM56 method. For this purpose, time series temperature data on a monthly scale are only used to train ML models. The developed ML models were applied to estimate ETo at different test stations and the obtained results were compared with the FAO PM56 method to verify and validate their performance in ETo estimation for the selected stations. In addition, multiple statistical indicators, including root-mean-square error (RMSE), coefficient of determination (R2), mean absolute error (MAE), Nash–Sutcliffe efficiency (NSE), and correlation coefficient (r) were calculated to compare the performance of each ML model on ETo estimation. Among the applied ML models, the ETo tree boost (TB) ML model outperformed the other ML models in estimating ETo in diverse climatic conditions based on statistical indicators (R2, NSE, r, RMSE, and MAE). Moreover, the observed R2, NSE, and r were the highest for the TB ML model, while RMSE and MAE were found to be the lowest at the study sites compared to other applied ML models. Lastly, ETo point data yielded from the TB ML model was used in an interpolation process to create monthly and annual ETo maps. Based on the ETo maps, this study suggests mainly a focus on areas with high ETo values and proper irrigation scheduling of crops to ensure water sustainability.
- Conference Article
4
- 10.1109/globconpt57482.2022.9938166
- Sep 23, 2022
This paper presents solar photovoltaic (PV) energy prediction based on thin-film technology utilizing various machine learning (ML) models. Several ML models like Support Vector Machine (SVM), Extra Tree Regression (ETR), Decision Tree Regression (DTR), K-Nearest Neighbour (kNN) and Feed-Forward Neural Network (FFNN) were utilized to evaluate each model's performance according to performance metrics. The primary input parameters such as time, solar radiation, wind speed, ambient and PV module temperatures, and the actual power generated by the thin-film PV panel based on the 2018 data set were considered for predicting solar PV output power. The ETR is proposed to predict the PV power output in this work and compared with other ML models. The results showed that ETRs outperformed the different ML algorithms, whereas DTR performed the poorest. The ETR model had the best performance, with Root Mean Square Error (RMSE) and Mean Absolute Error (MAE) values of 59.17 and 39.07, respectively. On the other hand, the DTR model performed poorly, with an RMSE of 81.83 and an MAE of 52.9, respectively.
- Research Article
2
- 10.13052/jmm1550-4646.1837
- Jan 22, 2022
- Journal of Mobile Multimedia
Prediction and forecasting of crop yield recently plays a vital role in the field of Agriculture. Drastic changes in climatic conditions, changes in rainfall season, and lack of nutrients content in the soil etc., due to major factors such as rapid industrialisation, global warming and pollution. This leads to the farmers’ predictions based on their own agricultural experiences on various crop yields based on external factors gone wrong. This results in farmers not getting adequate yield and suffering from financial loss. Machine learning and time series models are involved in this research work to carry out prediction and forecast of corn and soybean crop production over time through mobile application and it consist of various regression algorithms of machine learning such as multiple linear regression (MLR), decision tree regression (DTR), random forest tree regression (RFTR), k-nearest neighbour (KNN) and gradient boosting regression (GBR) are used for crop yield prediction. Time series models such as auto regression (AR), moving average (MA), auto regression integrated moving average (ARIMA) and vector auto regression (VAR) used for forecast of crop production. Comparative analysis also made between machine learning and time series models, in which GBR of machine learning outperformed other machine learning models with 92.648% predicted yield accuracy and VAR of time series model outperformed other time series models with 94.367% forecasted yield accuracy. Regression metrics such as mean square error (MSE), root mean square error (RMSE) and mean absolute error (MAE) are also involved in predicting crop yields.
- Research Article
15
- 10.1109/access.2020.3016664
- Jan 1, 2020
- IEEE Access
There have been many efforts to detect rumors using various machine learning (ML) models, but there is still a lack of understanding of their performance against different rumor topics and available features, resulting in a significant performance degrade against completely new and unseen (unknown) rumors. To address this issue, we investigate the relationship between ML models, features, and rumor topics to select the best rumor detection model under specific conditions using 13 different ML models. Our experiment results demonstrate that there is no clear winner among the ML models in all different rumor topics with respect to the detection performance. To overcome this problem, a possible way is to use an ensemble of ML models. Although previous work presents an improved detection of rumors using ensemble solutions (ES), their evaluation did not consider detecting unknown rumors. Further, they did not present nor evaluate the configuration of the ES to ensure that it indeed performs better than using a single ML model. Based on these observations, we propose to evaluate the use of an ES by examining their unknown rumor detection performance compared with single ML models but as well as different configurations of the ESes. Our experimental results using real-world datasets found that an ES of Random Forest, XGBoost and Multilayer perceptron overall produced the best F1 score of 0.79 for detecting unknown rumors, a significant improvement compared with a single best ML model which only achieved a 0.58 F1 score. We also showed that not all ESes are the same, with significantly degraded detection and large variations in performance when different ML models are used to construct the ES. Hence, it is infeasible to rely on any single ML model-based rumor detector. Finally, our solution also performed better than other recent detectors, such as eventAI and NileTMRG that performed similar to using a single ML model - making it a much more attractive solution to detect unknown rumors in practice.
- Research Article
7
- 10.1186/s12302-025-01078-w
- Mar 3, 2025
- Environmental Sciences Europe
The pollution in Dhaka's navigable waterways, including the Buriganga, Balu, Tongi Khal, and Turag rivers, is a significant concern due to rapid industrial and urban expansion. Industrial discharges, domestic sewage and inadequate waste management are the primary sources of this pollution, degrading water quality and threatening aquatic ecosystems. This study aimed to predict the Water Quality Index (WQI) of these rivers using fourteen machine learning (ML) models: Decision Tree Regression, Linear Regression, Ridge Regression, Stochastic Gradient Descent (SGD) Regressor, Extreme Gradient Boosting (XGB) Regressor, Light Gradient Boosting Machine (GBM) Regressor, Elastic Net Regressor, Support Vector Regression (SVM), Random Forest Regression, Bayesian Ridge Regressor, Artificial Neural Network (ANN), AdaBoost Regressor, CatBoost Regressor and Extra Trees Regressor. The objective was to evaluate and compare these models to identify the most effective predictive method for WQI, enabling efficient environmental monitoring and management of urban waterways. Among the evaluated ML models, ANN and Random Forest Regressor performed the best. The ANN model demonstrated superior predictive capability, achieving a Root Mean Squared Error (RMSE) of 2.34, a Mean Absolute Error (MAE) of 1.24, a Nash–Sutcliffe Efficiency (NSE) of 0.97, and a Coefficient of Determination (R2) of 0.97. Furthermore, an Adjusted R2 value of 0.965 further confirmed its ability to capture complex patterns in water quality data with remarkable accuracy. These findings emphasize the importance of using AI modeling techniques, specifically ANN and Random Forest Regression, to improve the accuracy of WQI forecasts for the waterways. This study contributes to the field of environmental science by offering a novel integration of feature selection techniques with ML models to enhance efficiency and cost-effectiveness of water quality monitoring. Unlike previous studies, this research specifically addresses the challenges of urban waterways in Dhaka, Bangladesh, a region significantly impacted by industrial and urban pollution. To our knowledge, this is the first study to apply such a comprehensive range of ML models to predict the WQI of Dhaka’s four major rivers. By providing a reliable methodology for WQI estimation, this study supports informed decision-making and proactive measures to protect vital water resources.
- Research Article
25
- 10.1038/s41598-024-69544-8
- Aug 10, 2024
- Scientific Reports
Solar photovoltaic (PV) systems, integral for sustainable energy, face challenges in forecasting due to the unpredictable nature of environmental factors influencing energy output. This study explores five distinct machine learning (ML) models which are built and compared to predict energy production based on four independent weather variables: wind speed, relative humidity, ambient temperature, and solar irradiation. The evaluated models include multiple linear regression (MLR), decision tree regression (DTR), random forest regression (RFR), support vector regression (SVR), and multi-layer perceptron (MLP). These models were hyperparameter tuned using chimp optimization algorithm (ChOA) for a performance appraisal. The models are subsequently validated on the data from a 264 kWp PV system, installed at the Applied Science University (ASU) in Amman, Jordan. Of all 5 models, MLP shows best root mean square error (RMSE), with the corresponding value of 0.503, followed by mean absolute error (MAE) of 0.397 and a coefficient of determination (R2) value of 0.99 in predicting energy from the observed environmental parameters. Finally, the process highlights the fact that fine-tuning of ML models for improved prediction accuracy in energy production domain still involves the use of advanced optimization techniques like ChOA, compared with other widely used optimization algorithms from the literature.
- Ask R Discovery
- Chat PDF
AI summaries and top papers from 250M+ research sources.