{"title":"Astroconformer:利用基于转换器的深度学习模型分析恒星光变曲线的前景","authors":"Jia-Shu Pan, Yuan-Sen Ting, Jie Yu","doi":"10.1093/mnras/stae068","DOIUrl":null,"url":null,"abstract":"Stellar light curves contain valuable information about oscillations and granulation, offering insights into stars’ internal structures and evolutionary states. Traditional asteroseismic techniques, primarily focused on power spectral analysis, often overlook the crucial phase information in these light curves. Addressing this gap, recent machine learning applications, particularly those using Convolutional Neural Networks (CNNs), have made strides in inferring stellar properties from light curves. However, CNNs are limited by their localized feature extraction capabilities. In response, we introduce Astroconformer, a Transformer-based deep learning framework, specifically designed to capture long-range dependencies in stellar light curves. Our empirical analysis centers on estimating surface gravity (log g), using a dataset derived from single-quarter Kepler light curves with log g values ranging from 0.2 to 4.4. Astroconformer demonstrates superior performance, achieving a root-mean-square-error (RMSE) of 0.017 dex at log g ≈ 3 in data-rich regimes and up to 0.1 dex in sparser areas. This performance surpasses both K-nearest neighbor models and advanced CNNs. Ablation studies highlight the influence of receptive field size on model effectiveness, with larger fields correlating to improved results. Astroconformer also excels in extracting νmax with high precision. It achieves less than 2 % relative median absolute error for 90-day red giant light curves. Notably, the error remains under 3 % for 30-day light curves, whose oscillations are undetectable by a conventional pipeline in 30 % cases. Furthermore, the attention mechanisms in Astroconformer align closely with the characteristics of stellar oscillations and granulation observed in light curves.","PeriodicalId":18930,"journal":{"name":"Monthly Notices of the Royal Astronomical Society","volume":"26 1","pages":""},"PeriodicalIF":4.7000,"publicationDate":"2024-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Astroconformer: The prospects of analyzing stellar light curves with transformer-based deep learning models\",\"authors\":\"Jia-Shu Pan, Yuan-Sen Ting, Jie Yu\",\"doi\":\"10.1093/mnras/stae068\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Stellar light curves contain valuable information about oscillations and granulation, offering insights into stars’ internal structures and evolutionary states. Traditional asteroseismic techniques, primarily focused on power spectral analysis, often overlook the crucial phase information in these light curves. Addressing this gap, recent machine learning applications, particularly those using Convolutional Neural Networks (CNNs), have made strides in inferring stellar properties from light curves. However, CNNs are limited by their localized feature extraction capabilities. In response, we introduce Astroconformer, a Transformer-based deep learning framework, specifically designed to capture long-range dependencies in stellar light curves. Our empirical analysis centers on estimating surface gravity (log g), using a dataset derived from single-quarter Kepler light curves with log g values ranging from 0.2 to 4.4. Astroconformer demonstrates superior performance, achieving a root-mean-square-error (RMSE) of 0.017 dex at log g ≈ 3 in data-rich regimes and up to 0.1 dex in sparser areas. This performance surpasses both K-nearest neighbor models and advanced CNNs. Ablation studies highlight the influence of receptive field size on model effectiveness, with larger fields correlating to improved results. Astroconformer also excels in extracting νmax with high precision. It achieves less than 2 % relative median absolute error for 90-day red giant light curves. Notably, the error remains under 3 % for 30-day light curves, whose oscillations are undetectable by a conventional pipeline in 30 % cases. Furthermore, the attention mechanisms in Astroconformer align closely with the characteristics of stellar oscillations and granulation observed in light curves.\",\"PeriodicalId\":18930,\"journal\":{\"name\":\"Monthly Notices of the Royal Astronomical Society\",\"volume\":\"26 1\",\"pages\":\"\"},\"PeriodicalIF\":4.7000,\"publicationDate\":\"2024-01-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Monthly Notices of the Royal Astronomical Society\",\"FirstCategoryId\":\"101\",\"ListUrlMain\":\"https://doi.org/10.1093/mnras/stae068\",\"RegionNum\":3,\"RegionCategory\":\"物理与天体物理\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ASTRONOMY & ASTROPHYSICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Monthly Notices of the Royal Astronomical Society","FirstCategoryId":"101","ListUrlMain":"https://doi.org/10.1093/mnras/stae068","RegionNum":3,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ASTRONOMY & ASTROPHYSICS","Score":null,"Total":0}
Astroconformer: The prospects of analyzing stellar light curves with transformer-based deep learning models
Stellar light curves contain valuable information about oscillations and granulation, offering insights into stars’ internal structures and evolutionary states. Traditional asteroseismic techniques, primarily focused on power spectral analysis, often overlook the crucial phase information in these light curves. Addressing this gap, recent machine learning applications, particularly those using Convolutional Neural Networks (CNNs), have made strides in inferring stellar properties from light curves. However, CNNs are limited by their localized feature extraction capabilities. In response, we introduce Astroconformer, a Transformer-based deep learning framework, specifically designed to capture long-range dependencies in stellar light curves. Our empirical analysis centers on estimating surface gravity (log g), using a dataset derived from single-quarter Kepler light curves with log g values ranging from 0.2 to 4.4. Astroconformer demonstrates superior performance, achieving a root-mean-square-error (RMSE) of 0.017 dex at log g ≈ 3 in data-rich regimes and up to 0.1 dex in sparser areas. This performance surpasses both K-nearest neighbor models and advanced CNNs. Ablation studies highlight the influence of receptive field size on model effectiveness, with larger fields correlating to improved results. Astroconformer also excels in extracting νmax with high precision. It achieves less than 2 % relative median absolute error for 90-day red giant light curves. Notably, the error remains under 3 % for 30-day light curves, whose oscillations are undetectable by a conventional pipeline in 30 % cases. Furthermore, the attention mechanisms in Astroconformer align closely with the characteristics of stellar oscillations and granulation observed in light curves.
期刊介绍:
Monthly Notices of the Royal Astronomical Society is one of the world''s leading primary research journals in astronomy and astrophysics, as well as one of the longest established. It publishes the results of original research in positional and dynamical astronomy, astrophysics, radio astronomy, cosmology, space research and the design of astronomical instruments.