{"title":"设计通用因果深度学习模型:几何(超级)变压器","authors":"Beatrice Acciaio, Anastasis Kratsios, Gudmund Pammer","doi":"10.1111/mafi.12389","DOIUrl":null,"url":null,"abstract":"<p>Several problems in stochastic analysis are defined through their geometry, and preserving that geometric structure is essential to generating meaningful predictions. Nevertheless, how to design principled deep learning (DL) models capable of encoding these geometric structures remains largely unknown. We address this open problem by introducing a universal causal geometric DL framework in which the user specifies a suitable pair of metric spaces <math>\n <semantics>\n <mi>X</mi>\n <annotation>$\\mathcal {X}$</annotation>\n </semantics></math> and <math>\n <semantics>\n <mi>Y</mi>\n <annotation>$\\mathcal {Y}$</annotation>\n </semantics></math> and our framework returns a DL model capable of causally approximating any “regular” map sending time series in <math>\n <semantics>\n <msup>\n <mi>X</mi>\n <mi>Z</mi>\n </msup>\n <annotation>$\\mathcal {X}^{\\mathbb {Z}}$</annotation>\n </semantics></math> to time series in <math>\n <semantics>\n <msup>\n <mi>Y</mi>\n <mi>Z</mi>\n </msup>\n <annotation>$\\mathcal {Y}^{\\mathbb {Z}}$</annotation>\n </semantics></math> while respecting their forward flow of information throughout time. Suitable geometries on <math>\n <semantics>\n <mi>Y</mi>\n <annotation>$\\mathcal {Y}$</annotation>\n </semantics></math> include various (adapted) Wasserstein spaces arising in optimal stopping problems, a variety of statistical manifolds describing the conditional distribution of continuous-time finite state Markov chains, and all Fréchet spaces admitting a Schauder basis, for example, as in classical finance. Suitable spaces <math>\n <semantics>\n <mi>X</mi>\n <annotation>$\\mathcal {X}$</annotation>\n </semantics></math> are compact subsets of any Euclidean space. Our results all quantitatively express the number of parameters needed for our DL model to achieve a given approximation error as a function of the target map's regularity and the geometric structure both of <math>\n <semantics>\n <mi>X</mi>\n <annotation>$\\mathcal {X}$</annotation>\n </semantics></math> and of <math>\n <semantics>\n <mi>Y</mi>\n <annotation>$\\mathcal {Y}$</annotation>\n </semantics></math>. Even when omitting any temporal structure, our universal approximation theorems are the first guarantees that Hölder functions, defined between such <math>\n <semantics>\n <mi>X</mi>\n <annotation>$\\mathcal {X}$</annotation>\n </semantics></math> and <math>\n <semantics>\n <mi>Y</mi>\n <annotation>$\\mathcal {Y}$</annotation>\n </semantics></math> can be approximated by DL models.</p>","PeriodicalId":49867,"journal":{"name":"Mathematical Finance","volume":null,"pages":null},"PeriodicalIF":1.6000,"publicationDate":"2023-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/mafi.12389","citationCount":"0","resultStr":"{\"title\":\"Designing universal causal deep learning models: The geometric (Hyper)transformer\",\"authors\":\"Beatrice Acciaio, Anastasis Kratsios, Gudmund Pammer\",\"doi\":\"10.1111/mafi.12389\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Several problems in stochastic analysis are defined through their geometry, and preserving that geometric structure is essential to generating meaningful predictions. Nevertheless, how to design principled deep learning (DL) models capable of encoding these geometric structures remains largely unknown. We address this open problem by introducing a universal causal geometric DL framework in which the user specifies a suitable pair of metric spaces <math>\\n <semantics>\\n <mi>X</mi>\\n <annotation>$\\\\mathcal {X}$</annotation>\\n </semantics></math> and <math>\\n <semantics>\\n <mi>Y</mi>\\n <annotation>$\\\\mathcal {Y}$</annotation>\\n </semantics></math> and our framework returns a DL model capable of causally approximating any “regular” map sending time series in <math>\\n <semantics>\\n <msup>\\n <mi>X</mi>\\n <mi>Z</mi>\\n </msup>\\n <annotation>$\\\\mathcal {X}^{\\\\mathbb {Z}}$</annotation>\\n </semantics></math> to time series in <math>\\n <semantics>\\n <msup>\\n <mi>Y</mi>\\n <mi>Z</mi>\\n </msup>\\n <annotation>$\\\\mathcal {Y}^{\\\\mathbb {Z}}$</annotation>\\n </semantics></math> while respecting their forward flow of information throughout time. Suitable geometries on <math>\\n <semantics>\\n <mi>Y</mi>\\n <annotation>$\\\\mathcal {Y}$</annotation>\\n </semantics></math> include various (adapted) Wasserstein spaces arising in optimal stopping problems, a variety of statistical manifolds describing the conditional distribution of continuous-time finite state Markov chains, and all Fréchet spaces admitting a Schauder basis, for example, as in classical finance. Suitable spaces <math>\\n <semantics>\\n <mi>X</mi>\\n <annotation>$\\\\mathcal {X}$</annotation>\\n </semantics></math> are compact subsets of any Euclidean space. Our results all quantitatively express the number of parameters needed for our DL model to achieve a given approximation error as a function of the target map's regularity and the geometric structure both of <math>\\n <semantics>\\n <mi>X</mi>\\n <annotation>$\\\\mathcal {X}$</annotation>\\n </semantics></math> and of <math>\\n <semantics>\\n <mi>Y</mi>\\n <annotation>$\\\\mathcal {Y}$</annotation>\\n </semantics></math>. Even when omitting any temporal structure, our universal approximation theorems are the first guarantees that Hölder functions, defined between such <math>\\n <semantics>\\n <mi>X</mi>\\n <annotation>$\\\\mathcal {X}$</annotation>\\n </semantics></math> and <math>\\n <semantics>\\n <mi>Y</mi>\\n <annotation>$\\\\mathcal {Y}$</annotation>\\n </semantics></math> can be approximated by DL models.</p>\",\"PeriodicalId\":49867,\"journal\":{\"name\":\"Mathematical Finance\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":1.6000,\"publicationDate\":\"2023-04-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1111/mafi.12389\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Mathematical Finance\",\"FirstCategoryId\":\"96\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1111/mafi.12389\",\"RegionNum\":3,\"RegionCategory\":\"经济学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"BUSINESS, FINANCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Mathematical Finance","FirstCategoryId":"96","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/mafi.12389","RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"BUSINESS, FINANCE","Score":null,"Total":0}
Designing universal causal deep learning models: The geometric (Hyper)transformer
Several problems in stochastic analysis are defined through their geometry, and preserving that geometric structure is essential to generating meaningful predictions. Nevertheless, how to design principled deep learning (DL) models capable of encoding these geometric structures remains largely unknown. We address this open problem by introducing a universal causal geometric DL framework in which the user specifies a suitable pair of metric spaces and and our framework returns a DL model capable of causally approximating any “regular” map sending time series in to time series in while respecting their forward flow of information throughout time. Suitable geometries on include various (adapted) Wasserstein spaces arising in optimal stopping problems, a variety of statistical manifolds describing the conditional distribution of continuous-time finite state Markov chains, and all Fréchet spaces admitting a Schauder basis, for example, as in classical finance. Suitable spaces are compact subsets of any Euclidean space. Our results all quantitatively express the number of parameters needed for our DL model to achieve a given approximation error as a function of the target map's regularity and the geometric structure both of and of . Even when omitting any temporal structure, our universal approximation theorems are the first guarantees that Hölder functions, defined between such and can be approximated by DL models.
期刊介绍:
Mathematical Finance seeks to publish original research articles focused on the development and application of novel mathematical and statistical methods for the analysis of financial problems.
The journal welcomes contributions on new statistical methods for the analysis of financial problems. Empirical results will be appropriate to the extent that they illustrate a statistical technique, validate a model or provide insight into a financial problem. Papers whose main contribution rests on empirical results derived with standard approaches will not be considered.