Abstract

With the availability of data and computational technologies in the modern world, machine learning (ML) has emerged as a preferred methodology for data analysis and prediction. While ML holds great promise, the results from such models are not fully unreliable due to the challenges introduced by uncertainty. An ML model generates an optimal solution based on its training data. However, if the uncertainty in the data and the model parameters are not considered, such optimal solutions have a high risk of failure in actual world deployment. This paper surveys the different approaches used in ML to quantify uncertainty. The paper also exhibits the implications of quantifying uncertainty when using ML by performing two case studies with space physics in focus. The first case study consists of the classification of auroral images in predefined labels. In the second case study, the horizontal component of the perturbed magnetic field measured at the Earth’s surface was predicted for the study of Geomagnetically Induced Currents (GICs) by training the model using time series data. In both cases, a Bayesian Neural Network (BNN) was trained to generate predictions, along with epistemic and aleatoric uncertainties. Finally, the pros and cons of both Gaussian Process Regression (GPR) models and Bayesian Deep Learning (DL) are weighed. The paper also provides recommendations for the models that need exploration, focusing on space weather prediction.

Highlights

  • IntroductionData are generated and collected through the web and sensors in the modern world due to the increased usage and advancements in electronic devices

  • Publisher’s Note: MDPI stays neutralData are generated and collected through the web and sensors in the modern world due to the increased usage and advancements in electronic devices

  • A more sophisticated architecture for the Neural Networks (NN), along with the Bayesian inference, could lead to greater accuracy, this study aims to present a motivation for uncertainty quantification in the domain of space weather forecasting

Read more

Summary

Introduction

Data are generated and collected through the web and sensors in the modern world due to the increased usage and advancements in electronic devices. Corporation (IDC) reported that, in 2011, the overall data volume generated in the world was 1.8 Zeta Bytes (ZB), and, within five years, it increased by nearly nine times [1]. Due to this availability and usage of data, the term Big Data has emerged, and a variant of its definition is as such: high volume, velocity, and variety of data that demand cost-effective, innovative forms of processing for enhanced insight and decision-making [2]. With the advantage of increased data availability in such complex domains and the advancement in computational hardware, there is a growing with regard to jurisdictional claims in published maps and institutional affiliations

Objectives
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.