Abstract
Accurate forecasting of F10.7 index is important on short, medium and long-term timescales since F10.7 is an excellent proxy of solar activity and it plays an important role within the Space Weather framework. The analysis of the signatures of transient solar radio emission and its prediction are a challenging task as the underpinning physical processes are typically nonlinear, non-stationary and chaotic. In this paper we want to present three Deep Learning approaches for the daily forecasting of the adjusted F10.7 solar radio flux up to 45 days, using a family of Long Short Term Memory (LSTM) based models. We investigated two novel hybrid architectures: the LSTM model used in combination with Fast Iterative Filtering as decomposition algorithm (FIF-LSTM) and a method based on Multi-Head-Attention architecture (FIF-LSTM-MHA). FIF is a robust decomposition signal method very suitable for analyzing non-linear and non-stationary time series and it is used to separate the original time series into different oscillation components according to frequency, derived without leaving the time domain before to be fed into the neural network. The Attention mechanism is able to keep track of long-term dependencies in data sequences and improve the computational efficiency of the prediction model by reducing the effect of irrelevant information, mimicking human attention and selecting the most critical input. Our comparative analysis evaluated the models’ performance for different time lags and solar activity levels. The results indicated that the hybrid models achieve better performance than the LSTM model for mid-range F10.7 predictions while the LSTM achieves better performance within the first few time lags. FIF-LSTM-MHA gives more promising output for longer forecasts since it tends to smooth the prediction curve due to the peculiarity of the Attention module to discard less relevant features of the time series and highlight the global trend.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.