An examination of daily CO2 emissions prediction through a comparative analysis of machine learning, deep learning, and statistical models

  • Abstract
  • Highlights & Summary
  • Literature Map
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

Human-induced global warming, primarily attributed to the rise in atmospheric CO2, poses a substantial risk to the survival of humanity. While most research focuses on predicting annual CO2 emissions, which are crucial for setting long-term emission mitigation targets, the precise prediction of daily CO2 emissions is equally vital for setting short-term targets. This study examines the performance of 14 models in predicting daily CO2 emissions data from 1/1/2022 to 30/9/2023 across the top four polluting regions (China, India, the USA, and the EU27&UK). The 14 models used in the study include four statistical models (ARMA, ARIMA, SARMA, and SARIMA), three machine learning models (support vector machine (SVM), random forest (RF), and gradient boosting (GB)), and seven deep learning models (artificial neural network (ANN), recurrent neural network variations such as gated recurrent unit (GRU), long short-term memory (LSTM), bidirectional-LSTM (BILSTM), and three hybrid combinations of CNN-RNN). Performance evaluation employs four metrics (R2, MAE, RMSE, and MAPE). The results show that the machine learning (ML) and deep learning (DL) models, with higher R2 (0.714–0.932) and lower RMSE (0.480–0.247) values, respectively, outperformed the statistical model, which had R2 (− 0.060–0.719) and RMSE (1.695–0.537) values, in predicting daily CO2 emissions across all four regions. The performance of the ML and DL models was further enhanced by differencing, a technique that improves accuracy by ensuring stationarity and creating additional features and patterns from which the model can learn. Additionally, applying ensemble techniques such as bagging and voting improved the performance of the ML models by approximately 9.6%, whereas hybrid combinations of CNN-RNN enhanced the performance of the RNN models. In summary, the performance of both the ML and DL models was relatively similar. However, due to the high computational requirements associated with DL models, the recommended models for daily CO2 emission prediction are ML models using the ensemble technique of voting and bagging. This model can assist in accurately forecasting daily emissions, aiding authorities in setting targets for CO2 emission reduction.

Similar Papers
  • Research Article
  • Cite Count Icon 16
  • 10.1038/s41598-024-82931-5
Explainable artificial intelligence for stroke prediction through comparison of deep learning and machine learning models
  • Dec 28, 2024
  • Scientific Reports
  • Khadijeh Moulaei + 5 more

Failure to predict stroke promptly may lead to delayed treatment, causing severe consequences like permanent neurological damage or death. Early detection using deep learning (DL) and machine learning (ML) models can enhance patient outcomes and mitigate the long-term effects of strokes. The aim of this study is to compare these models, exploring their efficacy in predicting stroke. This study analyzed a dataset comprising 663 records from patients hospitalized at Hazrat Rasool Akram Hospital in Tehran, Iran, including 401 healthy individuals and 262 stroke patients. A total of eight established ML (SVM, XGB, KNN, RF) and DL (DNN, FNN, LSTM, CNN) models were utilized to predict stroke. Techniques such as 10-fold cross-validation and hyperparameter tuning were implemented to prevent overfitting. The study also focused on interpretability through Shapley Additive Explanations (SHAP). The evaluation of model’s performance was based on accuracy, specificity, sensitivity, F1-score, and ROC curve metrics. Among DL models, LSTM showed superior sensitivity at 96.15%, while FNN exhibited better specificity (96.0%), accuracy (96.0%), F1-score (95.0%), and ROC (98.0%) among DL models. For ML models, RF displayed higher sensitivity (99.9%), accuracy (99.0%), specificity (100%), F1-score (99.0%), and ROC (99.9%). Overall, RF outperformed all models, while DL models surpassed ML models in most metrics except for RF. DL models (CNN, LSTM, DNN, FNN) achieved sensitivities from 93.0 to 96.15%, specificities from 80.0 to 96.0%, accuracies from 92.0 to 96.0%, F1-scores from 87.34 to 95.0%, and ROC scores from 95.0 to 98.0%. In contrast, ML models (KNN, XGB, SVM) showed sensitivities between 29.0% and 94.0%, specificities between 89.47% and 96.0%, accuracies between 71.0% and 95.0%, F1-scores between 44.0% and 95.0%, and ROC scores between 64.0% and 95.0%. This study demonstrates the efficacy of DL and ML models in predicting stroke, with the RF models outperforming all others in key metrics. While DL models generally surpassed ML models, RF’s exceptional performance highlights the potential of combining these technologies for early stroke detection, significantly improving patient outcomes by preventing severe consequences like permanent neurological damage or death.

  • Research Article
  • Cite Count Icon 11
  • 10.1111/exsy.13153
Deep learning‐based smishing message identification using regular expression feature generation
  • Oct 5, 2022
  • Expert Systems
  • Aakanksha Sharaff + 2 more

The increase in the number of undesired SMS termed smishing message and the data imbalance problem has generated a great demand for the development of more reliable anti‐spam filters. State of the art machine learning approaches are being employed to recognize and separate spam messages. Most recent studies target message classification by using numerous properties and features of the words but fail to consider the circumstantial features like long‐range dependencies between the words that are extremely important in identifying smishing messages. The idea is to develop an intelligent model that will distinguish between smishing messages and ham messages, by adopting a combined approach of regular expression (Regex), machine learning (ML) and deep learning (DL) models. Regex rules are generated using the dataset's spam messages for the purpose of refining the dataset. Support vector machine (SVM), Multinomial Naive Bayes and Random Forest are included under machine learning models and long short‐term memory (LSTM), bidirectional long short‐term memory (Bi‐LSTM), stacked LSTM and stacked Bi‐LSTM are included under deep learning models. The comparison between machine learning models and deep learning models is also carried out based on the performance evaluation parameters namely accuracy, precision, recall and F1 score of the models. It is observed that deep learning models perform better than machine learning models and the introduction of a regular expression to the dataset increases the efficiency of both the deep learning models and machine learning models.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 20
  • 10.1038/s41598-024-66481-4
Explainable artificial intelligence (XAI) for predicting the need for intubation in methanol-poisoned patients: a study comparing deep and machine learning models
  • Jul 8, 2024
  • Scientific Reports
  • Khadijeh Moulaei + 14 more

The need for intubation in methanol-poisoned patients, if not predicted in time, can lead to irreparable complications and even death. Artificial intelligence (AI) techniques like machine learning (ML) and deep learning (DL) greatly aid in accurately predicting intubation needs for methanol-poisoned patients. So, our study aims to assess Explainable Artificial Intelligence (XAI) for predicting intubation necessity in methanol-poisoned patients, comparing deep learning and machine learning models. This study analyzed a dataset of 897 patient records from Loghman Hakim Hospital in Tehran, Iran, encompassing cases of methanol poisoning, including those requiring intubation (202 cases) and those not requiring it (695 cases). Eight established ML (SVM, XGB, DT, RF) and DL (DNN, FNN, LSTM, CNN) models were used. Techniques such as tenfold cross-validation and hyperparameter tuning were applied to prevent overfitting. The study also focused on interpretability through SHAP and LIME methods. Model performance was evaluated based on accuracy, specificity, sensitivity, F1-score, and ROC curve metrics. Among DL models, LSTM showed superior performance in accuracy (94.0%), sensitivity (99.0%), specificity (94.0%), and F1-score (97.0%). CNN led in ROC with 78.0%. For ML models, RF excelled in accuracy (97.0%) and specificity (100%), followed by XGB with sensitivity (99.37%), F1-score (98.27%), and ROC (96.08%). Overall, RF and XGB outperformed other models, with accuracy (97.0%) and specificity (100%) for RF, and sensitivity (99.37%), F1-score (98.27%), and ROC (96.08%) for XGB. ML models surpassed DL models across all metrics, with accuracies from 93.0% to 97.0% for DL and 93.0% to 99.0% for ML. Sensitivities ranged from 98.0% to 99.37% for DL and 93.0% to 99.0% for ML. DL models achieved specificities from 78.0% to 94.0%, while ML models ranged from 93.0% to 100%. F1-scores for DL were between 93.0% and 97.0%, and for ML between 96.0% and 98.27%. DL models scored ROC between 68.0% and 78.0%, while ML models ranged from 84.0% to 96.08%. Key features for predicting intubation necessity include GCS at admission, ICU admission, age, longer folic acid therapy duration, elevated BUN and AST levels, VBG_HCO3 at initial record, and hemodialysis presence. This study as the showcases XAI's effectiveness in predicting intubation necessity in methanol-poisoned patients. ML models, particularly RF and XGB, outperform DL counterparts, underscoring their potential for clinical decision-making.

  • Research Article
  • Cite Count Icon 34
  • 10.1016/j.resourpol.2023.104216
A novel deep-learning technique for forecasting oil price volatility using historical prices of five precious metals in context of green financing – A comparison of deep learning, machine learning, and statistical models
  • Oct 1, 2023
  • Resources Policy
  • Muhammad Mohsin + 1 more

A novel deep-learning technique for forecasting oil price volatility using historical prices of five precious metals in context of green financing – A comparison of deep learning, machine learning, and statistical models

  • Research Article
  • Cite Count Icon 30
  • 10.1371/journal.pone.0317619
Evaluating Machine Learning and Deep Learning models for predicting Wind Turbine power output from environmental factors.
  • Jan 23, 2025
  • PloS one
  • Montaser Abdelsattar + 4 more

This study presents a comprehensive comparative analysis of Machine Learning (ML) and Deep Learning (DL) models for predicting Wind Turbine (WT) power output based on environmental variables such as temperature, humidity, wind speed, and wind direction. Along with Artificial Neural Network (ANN), Long Short-Term Memory (LSTM), Recurrent Neural Network (RNN), and Convolutional Neural Network (CNN), the following ML models were looked at: Linear Regression (LR), Support Vector Regressor (SVR), Random Forest (RF), Extra Trees (ET), Adaptive Boosting (AdaBoost), Categorical Boosting (CatBoost), Extreme Gradient Boosting (XGBoost), and Light Gradient Boosting Machine (LightGBM). Using a dataset of 40,000 observations, the models were assessed based on R-squared, Mean Absolute Error (MAE), and Root Mean Square Error (RMSE). ET achieved the highest performance among ML models, with an R-squared value of 0.7231 and a RMSE of 0.1512. Among DL models, ANN demonstrated the best performance, achieving an R-squared value of 0.7248 and a RMSE of 0.1516. The results show that DL models, especially ANN, did slightly better than the best ML models. This means that they are better at modeling non-linear dependencies in multivariate data. Preprocessing techniques, including feature scaling and parameter tuning, improved model performance by enhancing data consistency and optimizing hyperparameters. When compared to previous benchmarks, the performance of both ANN and ET demonstrates significant predictive accuracy gains in WT power output forecasting. This study's novelty lies in directly comparing a diverse range of ML and DL algorithms while highlighting the potential of advanced computational approaches for renewable energy optimization.

  • Preprint Article
  • 10.5194/ems2025-562
Hydrological modelling using machine and deep learning models across multiple case studies
  • Jul 16, 2025
  • Majid Niazkar + 3 more

Machine learning (ML) and deep learning (DL) models can play an important role when it comes to modelling complicated processes. Such capability is necessary for hydrological and climate-related applications. Generally, ML models utilize precipitation and temperature time series of a basin as input to develop a lumped rainfall-runoff model to simulate streamflow at the basin outlet. However, when it is divided into several sub-basins, Graph Neural Networks (GNN) can consider each sub-basin as a node and link them together using a connectivity matrix to account for spatial variations of hydroclimatic variables. In this study, GNN and various ML models with different types of architecture, ranging from neural networks, tree-based structure, and gradient boosting, were exploited for daily streamflow simulation over different case studies. For each case study, the basin was divided into a few sub-basins for which daily precipitation and temperature data were aggregated and used as input. For training GNN, the connection matrix of sub-basins was also used as input. Basically, 75% of historical records were utilized to train GNN and different ML models, e.g., artificial neural networks, support vector machine, decision tree, random forest, eXtreme Gradient Boosting (XGBoost), Light Gradient-Boosting Machine (LightGBM), and Category Boosting (CatBoost), while the rest was used for testing. Streamflow simulation was conducted with/without considering seasonality impact and lag times. The obtained results clearly demonstrate that considering seasonality and time lags can enhance accuracy of streamflow predictions based on Kling–Gupta efficiency (KGE). Furthermore, GNN with seasonality impact and time lags achieved promising results across different case studies with KGE>0.85 for training and KGE>0.59 for testing data, respectively. Among ML models, boosting models, e.g., LightGBM and XGBoost, performed slightly better than other ML models. for Finally, this comparative analysis provides valuable insights for ML/DL applications in climate change impact assessments.Acknowledgements: This research work was carried out as part of the TRANSCEND project with funding received from the European Union Horizon Europe Research and Innovation Programme under Grant Agreement No. 10108411.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 13
  • 10.1371/journal.pone.0282608
A hybrid CNN and ensemble model for COVID-19 lung infection detection on chest CT scans.
  • Mar 9, 2023
  • PLOS ONE
  • Ahmed A Akl + 3 more

COVID-19 is highly infectious and causes acute respiratory disease. Machine learning (ML) and deep learning (DL) models are vital in detecting disease from computerized chest tomography (CT) scans. The DL models outperformed the ML models. For COVID-19 detection from CT scan images, DL models are used as end-to-end models. Thus, the performance of the model is evaluated for the quality of the extracted feature and classification accuracy. There are four contributions included in this work. First, this research is motivated by studying the quality of the extracted feature from the DL by feeding these extracted to an ML model. In other words, we proposed comparing the end-to-end DL model performance against the approach of using DL for feature extraction and ML for the classification of COVID-19 CT scan images. Second, we proposed studying the effect of fusing extracted features from image descriptors, e.g., Scale-Invariant Feature Transform (SIFT), with extracted features from DL models. Third, we proposed a new Convolutional Neural Network (CNN) to be trained from scratch and then compared to the deep transfer learning on the same classification problem. Finally, we studied the performance gap between classic ML models against ensemble learning models. The proposed framework is evaluated using a CT dataset, where the obtained results are evaluated using five different metrics The obtained results revealed that using the proposed CNN model is better than using the well-known DL model for the purpose of feature extraction. Moreover, using a DL model for feature extraction and an ML model for the classification task achieved better results in comparison to using an end-to-end DL model for detecting COVID-19 CT scan images. Of note, the accuracy rate of the former method improved by using ensemble learning models instead of the classic ML models. The proposed method achieved the best accuracy rate of 99.39%.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 6
  • 10.54254/2755-2721/55/20241475
Enhancing customer behavior prediction in e-commerce: A comparative analysis of machine learning and deep learning models
  • Jul 25, 2024
  • Applied and Computational Engineering
  • Deming Liu + 4 more

The digital era has transformed the way businesses interact with their customers, with online platforms serving as crucial touchpoints for user engagement. Understanding customer behavior in this context is paramount for enhancing user experience, optimizing marketing strategies, and driving business growth. This study aims to explore the likelihood of customers making purchases based on their clickstream data by employing both machine learning and deep learning techniques. This research uses a machine learning model Random Forest (RF), Gradient Boosting Decision Trees (GBDT), Extreme Gradient Boosting (XGBOOST) and deep learning model Recurrent Neural Networks (RNN), Long Short-Term Memory (LSTM) to predict whether customers will purchase the items using 33,040,175 records in the file of the click and 1,177,769 records in the buys file from real e-commerce customers. The results show that both machine learning and deep learning can accurately forecast the purchasing behavior of customers with an accuracy of around 72 to 75 percent. For the machine learning model, attains the highest prediction accuracy when using a sliding window of 6 days. For the deep learning model, the LSTM model with 50 layers shows the highest prediction of customers willingness to purchase an item. Compared with previous studies, the three machine learning models narrow the range of days, give more accurate predictions, and also improve the model. Both RNN and LSTM show similar accuracy for customer behavior. The current research has asserted that both machine learning and deep learning models give profound results on whether customers will purchase a product, and there is not a significant difference between machine learning and deep learning in this classification topic.

  • Research Article
  • Cite Count Icon 23
  • 10.1007/s11356-021-13503-7
Spatial modelling of soil salinity: deep or shallow learning models?
  • Mar 23, 2021
  • Environmental Science and Pollution Research
  • Aliakbar Mohammadifar + 3 more

Understanding the spatial distribution of soil salinity is required to conserve land against degradation and desertification. Against this background, this study is the first attempt to predict soil salinity in the Jaghin basin, in southern Iran, by applying and comparing the performance of four deep learning (DL) models (deep convolutional neural networks-DCNNs, dense connected deep neural networks-DenseDNNs, recurrent neural networks-long short-term memory-RNN-LSTM and recurrent neural networks-gated recurrent unit-RNN-GRU) and six shallow machine learning (ML) models (bagged classification and regression tree-BCART, cforest, cubist, quantile regression with LASSO penalty-QR-LASSO, ridge regression-RR and support vectore machine-SVM). To do this, 49 environmental landsat8-derived variables including digital elevation model (DEM)-extracted covariates, soil-salinity indices, and other variables (e.g., soil order, lithology, land use) were mapped spatially. For assessing the relationships between soil salinity (EC) and factors controlling EC, we collected 319 surficial (0-5 cm depth) soil samples for measuring soil salinity on the basis of electrical conductivity (EC). We then selected the most important features (covariates) controlling soil salinity by applying a MARS model. The performance of the DL and shallow ML models for generating soil salinity spatial maps (SSSMs) was assessed using a Taylor diagram and the Nash Sutcliff coefficient (NSE). Among all 10 predictive models, DL models with NSE ≥ 0.9 (DCNNs was the most accurate model with NSE = 0.96) were selected as the four best models, and performed better than the six shallow ML models with NSE ≤ 0.83 (QR-LASSO was the weakest predictive model with NSE = 0.50). Based on DCNNs-, the values of the EC ranged between 0.67 and 14.73 dS/m, whereas for QR-LASSO the corresponding EC values were 0.37 to 19.6 dS/m. Overall, DL models performed better than shallow ML models for production of the SSSMs and therefore we recommend applying DL models for prediction purposes in environmental sciences.

  • Research Article
  • Cite Count Icon 10
  • 10.1038/s41598-025-99167-6
Trade-offs between machine learning and deep learning for mental illness detection on social media
  • Apr 25, 2025
  • Scientific Reports
  • Zhanyi Ding + 7 more

Social media platforms provide valuable insights into mental health trends by capturing user-generated discussions on conditions such as depression, anxiety, and suicidal ideation. Machine learning (ML) and deep learning (DL) models have been increasingly applied to classify mental health conditions from textual data, but selecting the most effective model involves trade-offs in accuracy, interpretability, and computational efficiency. This study evaluates multiple ML models, including logistic regression, random forest, and LightGBM, alongside DL architectures such as ALBERT and Gated Recurrent Units (GRUs), for both binary and multi-class classification of mental health conditions. Our findings indicate that ML and DL models achieve comparable classification performance on medium-sized datasets, with ML models offering greater interpretability through variable importance scores, while DL models are more robust to complex linguistic patterns. Additionally, ML models require explicit feature engineering, whereas DL models learn hierarchical representations directly from text. Logistic regression provides the advantage of capturing both positive and negative associations between features and mental health conditions, whereas tree-based models prioritize decision-making power through split-based feature selection. This study offers empirical insights into the advantages and limitations of different modeling approaches and provides recommendations for selecting appropriate methods based on dataset size, interpretability needs, and computational constraints.

  • Research Article
  • Cite Count Icon 5
  • 10.3390/modelling6020035
Advancement of Artificial Intelligence in Cost Estimation for Project Management Success: A Systematic Review of Machine Learning, Deep Learning, Regression, and Hybrid Models
  • Apr 24, 2025
  • Modelling
  • Md Mahfuzul Islam Shamim + 3 more

This systematic review investigates the integration of artificial intelligence (AI) in cost estimation within project management, focusing on its impact on accuracy and efficiency compared to traditional methods. This study synthesizes findings from 39 high-quality articles published between 2016 and 2024, evaluating various machine learning (ML), deep learning (DL), regression, and hybrid models in sectors such as construction, healthcare, manufacturing, and real estate. The results show that AI-powered approaches, particularly artificial neural networks (ANNs)—which constitute 26.33% of the studies—, enhance predictive accuracy and adaptability to complex, dynamic project environments. Key AI techniques, including support vector machines (SVMs) (7.90% of studies), decision trees, and gradient-boosting models, offer substantial improvements in cost prediction and resource optimization. ML models, including ANNs and deep learning models, represent approximately 70% of the reviewed studies, demonstrating a clear trend toward the adoption of advanced AI techniques. On average, deep learning models perform with 85–90% accuracy in cost estimation, making them highly effective for handling complex, nonlinear relationships and large datasets. Machine learning models achieve an average accuracy of 75–80%, providing strong performance, particularly in industries like road construction and healthcare. Regression models typically deliver 70–80% accuracy, being more suitable for simpler cost estimations where the relationships between variables are linear. Hybrid models combine the strengths of different algorithms, achieving 80–90% accuracy on average, and are particularly effective in complex, multi-faceted projects. Overall, deep learning and hybrid models offer the highest accuracy in cost estimation, while machine learning and regression models still provide reliable results for specific applications.

  • Research Article
  • 10.25147/ijcsr.2017.001.1.224
Empirical Analysis of the State-of-the-Art Models for Handling Polarity Shifts Due to Implicit Negation in Mobile Phone Reviews
  • Jan 1, 2025
  • International Journal of Computing Sciences Research
  • Millicent Murithi + 2 more

Purpose–This paper presents a comprehensive empirical analysis focusing on sentiment flux within state-of-the-art models designed for handling polarity shifts due to implicit negation in Amazon mobile phones' reviews. Method–The research evaluates diverse models across five categories: traditional machine learning (ML), deep learning (DL), and hybrid models combining both approaches. Various feature extraction, feature selection, and data augmentation techniques are tested on Amazon mobile phone reviews dataset. BERT and LSTM are used for deep learning while SVM and Naive Bayes are used for traditional ML. ANOVA is used to identify the presence or absence of significant differences and interactions among these entities. Results –DL shows superior performance compared to traditional ML models. ANOVA analysis shows significant performance differences between conventional ML and DL models. Traditional ML models interact significantly with feature extraction and selection techniques while DL models do not. Traditional ML models do not interact significantly with data augmentation methods while DL models do. FastText extraction outperforms word2vec; Back translation outperforms synonym replacement while recursive feature selection (RFE) surpasses TF-IDF (Term Frequency-Inverse Document Frequency). The BERT and LSTM exhibit one of the strongest performances. Conclusion –The study concludes that DL models are more effective. Data augmentation techniques significantly impact the performance of DL models, with back translation showing superior performance over synonym replacement. This provides a leverage point in developing an improved model in the future. Recommendations –Future research should focus on developing a hybrid model for Enhanced Polarity Shift Management of Mobile Phone Reviews using Contextual Back Translation Augmented by Seq2seq Perturbations. This aims at leveraging contextual back translation and Seq2seq perturbations to generate a diverse interpretation that consequently improves the model's ability to handle nuanced expressions of sentiments due to implicit negation with enhanced accuracy, generalizability, robustness to polarity shifts, and contextual understanding. Research Implications –The findings provide valuable insights into the development of state-of-the-art models, offering a promising direction for further research in sentiment analysis. Keywords –empirical analysis, hybrid, perturbations, implicit negation, sentiment flux

  • Research Article
  • 10.18502/japh.v10i1.18093
Prediction of particulate matter PM2.5 level in the air of Islamabad, Pakistan by using machine learning and deep learning approaches
  • Mar 9, 2025
  • Journal of Air Pollution and Health
  • Muhammad Waqas + 4 more

Introduction: Air pollution is a significant global health challenge, contributing to the deaths of millions of people annually. Among these pollutants, Particulate Matter (PM2.5) is the most harmful to the respiratory system causing serious health problems. This study focused on predicting PM2.5 in the air of Islamabad, capital of Pakistan by using machine learning and deep learning models. Materials and methods: Two machine learning models (Decision Tree and Random Forest) and four deep learning models including Multi-Layer Neural Network (MLNN), Long Short-Term Memory (LSTM), Recurrent Neural Network (RNN), Gated Recurrent Unit (GRU) are used in the study. Each model's performance was assessed by using statistical indicators including coefficient of determination (R2), Mean Absolute Error (MAE), Root Mean Square Error (RMSE), and Relative Root Mean Square Error (RRMSE). These models are also ranked based on their performance by compromise programming technique. Results: Machine learning models performed better in the training phase by achieving higher R2 values of 0.98 and 0.97 but couldn’t maintain the same performance in the testing phase. Whereas the deep learning models performed best in both the training and testing phases. MLNN model attained higher R2 value of 0.98 in training and 0.88 in testing and is evaluated as top-ranked prediction model in predicting particulate matter PM2.5. Whereas,LSTM, GRU, RNN, Decision Tree, and Random Forest are placed at the 2nd,3rd, 4th, 5th, and 6th positions having R2 values of 0.86, 0.87, 0.82, 0.99, and0.97 during training and 0.71, 0.69, 0.69, 0.75, and 0.85 respectively during testing. Conclusion: Deep learning models, especially MLNN, showed strong performance in predicting PM2.5 as compared to the machine learning models.

  • Research Article
  • 10.37547/marketing-fmmej-05-02-02
Enhancing Stock Price Prediction through Sentiment Analysis: A Comparative Study of Machine Learning and Deep Learning Models Using Financial News Data
  • Feb 3, 2025
  • Frontline Marketing, Management and Economics Journal
  • Paresh Chandra Nath + 6 more

This study explores the use of machine learning (ML) and deep learning (DL) models for predicting stock price movements through sentiment analysis of financial news articles. Four models were evaluated: Random Forest (RF), Gradient Boosting (GB), Long Short-Term Memory (LSTM), and Bidirectional Encoder Representations from Transformers (BERT). The results showed that deep learning models, particularly BERT, outperformed traditional ML models, achieving higher accuracy, precision, recall, and F1 scores. BERT’s ability to capture contextual relationships in text proved superior in handling the complexities of financial news. This research highlights the effectiveness of sentiment analysis in stock market prediction and suggests that advanced ML and DL techniques can enhance forecasting accuracy. Future work could focus on refining these models by integrating more data sources and exploring hybrid approaches.

  • Research Article
  • 10.1186/s12885-025-15121-9
Comparing radiomics, deep learning, and fusion models for predicting occult pleural dissemination in patients with non-small cell lung cancer: a retrospective multicenter study
  • Oct 29, 2025
  • BMC Cancer
  • Tao Bao + 11 more

BackgroundOccult pleural dissemination (PD) in non-small cell lung cancer (NSCLC) patients is likely to be missed on computed tomography (CT) scans, associated with poor survival, and generally contraindicated for radical surgery. This study aimed to develop and compare the performance of radiomics-based machine learning (ML), deep learning (DL), and fusion models to preoperatively identify occult PD in NSCLC patients.Materials and methodsA total of 326 NSCLC patients from three Chinese high-volume medical centers (2016–2023) were retrospectively collected and divided into training (n = 216), internal test (n = 54), and external test (n = 56) cohorts. Ten radiomics-based ML models and eight DL models were trained using CT images at the maximum cross-sectional slice of the primary tumor. Moreover, another two fusion models (prefusion and postfusion) were developed using feature-based and decision-based methods. The receiver operating characteristic curve (ROC) and area under the curve (AUC) were mainly used to compare the predictive performance of the models.ResultsThe GBM (AUC: 0.821) and DenseNet121 (AUC: 0.764) models achieved the highest AUC among ML and DL models in the external test cohorts, respectively. The postfusion model, integrating the output probabilities from GBM and DenseNet121 models, showed superior performance (AUC: 0.828–0.978) compared to the prefusion model (AUC: 0.817–0.877). Moreover, the postfusion model demonstrated the highest degree of sensitivity (82.1–97.2%) among all models across the three cohorts.ConclusionsThe postfusion model, which integrates radiomics-based ML and DL models, can serve as a sensitive diagnostic tool to predict occult PD in NSCLC patients, thereby helping to avoid unnecessary surgeries.Supplementary InformationThe online version contains supplementary material available at 10.1186/s12885-025-15121-9.

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.