Willa Potosnak, Cristian Challu, Mononito Goswami, Michał Wiliński, Nina Żukowska
{"title":"深度时间序列预测中的隐含推理","authors":"Willa Potosnak, Cristian Challu, Mononito Goswami, Michał Wiliński, Nina Żukowska","doi":"arxiv-2409.10840","DOIUrl":null,"url":null,"abstract":"Recently, time series foundation models have shown promising zero-shot\nforecasting performance on time series from a wide range of domains. However,\nit remains unclear whether their success stems from a true understanding of\ntemporal dynamics or simply from memorizing the training data. While implicit\nreasoning in language models has been studied, similar evaluations for time\nseries models have been largely unexplored. This work takes an initial step\ntoward assessing the reasoning abilities of deep time series forecasting\nmodels. We find that certain linear, MLP-based, and patch-based Transformer\nmodels generalize effectively in systematically orchestrated\nout-of-distribution scenarios, suggesting underexplored reasoning capabilities\nbeyond simple pattern memorization.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Implicit Reasoning in Deep Time Series Forecasting\",\"authors\":\"Willa Potosnak, Cristian Challu, Mononito Goswami, Michał Wiliński, Nina Żukowska\",\"doi\":\"arxiv-2409.10840\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recently, time series foundation models have shown promising zero-shot\\nforecasting performance on time series from a wide range of domains. However,\\nit remains unclear whether their success stems from a true understanding of\\ntemporal dynamics or simply from memorizing the training data. While implicit\\nreasoning in language models has been studied, similar evaluations for time\\nseries models have been largely unexplored. This work takes an initial step\\ntoward assessing the reasoning abilities of deep time series forecasting\\nmodels. We find that certain linear, MLP-based, and patch-based Transformer\\nmodels generalize effectively in systematically orchestrated\\nout-of-distribution scenarios, suggesting underexplored reasoning capabilities\\nbeyond simple pattern memorization.\",\"PeriodicalId\":501301,\"journal\":{\"name\":\"arXiv - CS - Machine Learning\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Machine Learning\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.10840\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Machine Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10840","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Implicit Reasoning in Deep Time Series Forecasting
Recently, time series foundation models have shown promising zero-shot
forecasting performance on time series from a wide range of domains. However,
it remains unclear whether their success stems from a true understanding of
temporal dynamics or simply from memorizing the training data. While implicit
reasoning in language models has been studied, similar evaluations for time
series models have been largely unexplored. This work takes an initial step
toward assessing the reasoning abilities of deep time series forecasting
models. We find that certain linear, MLP-based, and patch-based Transformer
models generalize effectively in systematically orchestrated
out-of-distribution scenarios, suggesting underexplored reasoning capabilities
beyond simple pattern memorization.