{"title":"Deep Learning LSTM-based approaches for 10.7 cm solar radio flux forecasting up to 45-days","authors":"G. Jerse , A. Marcucci","doi":"10.1016/j.ascom.2024.100786","DOIUrl":null,"url":null,"abstract":"<div><p>Accurate forecasting of F10.7 index is important on short, medium and long-term timescales since F10.7 is an excellent proxy of solar activity and it plays an important role within the Space Weather framework. The analysis of the signatures of transient solar radio emission and its prediction are a challenging task as the underpinning physical processes are typically nonlinear, non-stationary and chaotic. In this paper we want to present three Deep Learning approaches for the daily forecasting of the adjusted F10.7 solar radio flux up to 45 days, using a family of Long Short Term Memory (LSTM) based models. We investigated two novel hybrid architectures: the LSTM model used in combination with Fast Iterative Filtering as decomposition algorithm (FIF-LSTM) and a method based on Multi-Head-Attention architecture (FIF-LSTM-MHA). FIF is a robust decomposition signal method very suitable for analyzing non-linear and non-stationary time series and it is used to separate the original time series into different oscillation components according to frequency, derived without leaving the time domain before to be fed into the neural network. The Attention mechanism is able to keep track of long-term dependencies in data sequences and improve the computational efficiency of the prediction model by reducing the effect of irrelevant information, mimicking human attention and selecting the most critical input. Our comparative analysis evaluated the models’ performance for different time lags and solar activity levels. The results indicated that the hybrid models achieve better performance than the LSTM model for mid-range F10.7 predictions while the LSTM achieves better performance within the first few time lags. FIF-LSTM-MHA gives more promising output for longer forecasts since it tends to smooth the prediction curve due to the peculiarity of the Attention module to discard less relevant features of the time series and highlight the global trend.</p></div>","PeriodicalId":48757,"journal":{"name":"Astronomy and Computing","volume":"46 ","pages":"Article 100786"},"PeriodicalIF":1.9000,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2213133724000015/pdfft?md5=702c624fbdced03b7eb08d858ec2dc15&pid=1-s2.0-S2213133724000015-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Astronomy and Computing","FirstCategoryId":"101","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2213133724000015","RegionNum":4,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ASTRONOMY & ASTROPHYSICS","Score":null,"Total":0}
引用次数: 0
Abstract
Accurate forecasting of F10.7 index is important on short, medium and long-term timescales since F10.7 is an excellent proxy of solar activity and it plays an important role within the Space Weather framework. The analysis of the signatures of transient solar radio emission and its prediction are a challenging task as the underpinning physical processes are typically nonlinear, non-stationary and chaotic. In this paper we want to present three Deep Learning approaches for the daily forecasting of the adjusted F10.7 solar radio flux up to 45 days, using a family of Long Short Term Memory (LSTM) based models. We investigated two novel hybrid architectures: the LSTM model used in combination with Fast Iterative Filtering as decomposition algorithm (FIF-LSTM) and a method based on Multi-Head-Attention architecture (FIF-LSTM-MHA). FIF is a robust decomposition signal method very suitable for analyzing non-linear and non-stationary time series and it is used to separate the original time series into different oscillation components according to frequency, derived without leaving the time domain before to be fed into the neural network. The Attention mechanism is able to keep track of long-term dependencies in data sequences and improve the computational efficiency of the prediction model by reducing the effect of irrelevant information, mimicking human attention and selecting the most critical input. Our comparative analysis evaluated the models’ performance for different time lags and solar activity levels. The results indicated that the hybrid models achieve better performance than the LSTM model for mid-range F10.7 predictions while the LSTM achieves better performance within the first few time lags. FIF-LSTM-MHA gives more promising output for longer forecasts since it tends to smooth the prediction curve due to the peculiarity of the Attention module to discard less relevant features of the time series and highlight the global trend.
Astronomy and ComputingASTRONOMY & ASTROPHYSICSCOMPUTER SCIENCE,-COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS
CiteScore
4.10
自引率
8.00%
发文量
67
期刊介绍:
Astronomy and Computing is a peer-reviewed journal that focuses on the broad area between astronomy, computer science and information technology. The journal aims to publish the work of scientists and (software) engineers in all aspects of astronomical computing, including the collection, analysis, reduction, visualisation, preservation and dissemination of data, and the development of astronomical software and simulations. The journal covers applications for academic computer science techniques to astronomy, as well as novel applications of information technologies within astronomy.