Abstract

Long short-term memory (LSTM) models based on specialized deep neural network-based architecture have emerged as an important model for forecasting time-series. However, the literature does not provide clear guidelines for design choices, which affect forecasting performance. Such choices include the need for pre-processing techniques such as deseasonalization, ordering of the input data, network size, batch size, and forecasting horizon. We detail this in the context of short-term forecasting of global horizontal irradiance, an accepted proxy for solar energy. Particularly, short-term forecasting is critical because the cloud conditions change at a sub-hourly having large impacts on incident solar radiation. We conduct an empirical investigation based on data from three solar stations from two climatic zones of India over two seasons. From an application perspective, it may be noted that despite the thrust given to solar energy generation in India, the literature contains few instances of robust studies across climatic zones and seasons. The model thus obtained subsequently outperformed three recent benchmark methods based on random forest, recurrent neural network, and LSTM, respectively, in terms of forecasting accuracy. Our findings underscore the importance of considering the temporal order of the data, lack of any discernible benefit from data pre-processing, the effect of making the LSTM model stateful. It is also found that the number of nodes in an LSTM network, as well as batch size, is influenced by the variability of the input data.

Highlights

  • Solar radiation is one of the most important components of alternative sources of energy [32, 37]

  • It is observed that DSS-Long short-term memory (LSTM) has a better normalized root mean square error (nRMSE) score if it is dealing with raw time-series

  • A stable short-term forecasting model for solar energy generation is critical as there is a lot of variance due to the subhourly cloud phenomenon

Read more

Summary

Introduction

Solar radiation is one of the most important components of alternative sources of energy [32, 37]. In paper, [14] and [50], the authors have treated the input features as independent of time Another important design issue is data pre-processing such as identification and removal of trend and seasonality. There is a general disagreement between design choices for an LSTM, such as preserving the temporal order of the data and the need for pre-processing. Apart from the above-mentioned two design issues, it is perceived that a few other issues like the batch size, the prediction horizon, adjustments for inherent input data variability can impact model performance. The design questions enlisted have been empirically evaluated, and important recommendations like considering the temporal order of the data (Non-Supervised setup), no pre-processing, and preserving dependency between batches have been made.

Related work
Statistical and machine learning models
Deep learning‐based models
Deep learning sequence model and LSTM
Materials and methods
Data collection
Data pre‐processing
Supervised or non‐supervised learning
LSTM architectures
Evaluation of forecasting model
Importance of data pre‐processing
Supervised or non‐supervised?
Effect of batch size
Input variability vs Network complexity
Comparison to other prediction approaches
Conclusion
Compliance with ethical standards
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call