Pub Date : 2024-04-22DOI: 10.1016/j.jco.2024.101857
Erich Novak
{"title":"David Krieg is the winner of the 2024 Joseph F. Traub Prize for Achievement in Information-Based Complexity","authors":"Erich Novak","doi":"10.1016/j.jco.2024.101857","DOIUrl":"https://doi.org/10.1016/j.jco.2024.101857","url":null,"abstract":"","PeriodicalId":50227,"journal":{"name":"Journal of Complexity","volume":"83 ","pages":"Article 101857"},"PeriodicalIF":1.7,"publicationDate":"2024-04-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0885064X24000347/pdfft?md5=2ff01990fe6f5661ee24005dadc016f8&pid=1-s2.0-S0885064X24000347-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140633006","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-20DOI: 10.1016/j.jco.2024.101856
Onyekachi Emenike , Fred J. Hickernell , Peter Kritzer
A large literature specifies conditions under which the information complexity for a sequence of numerical problems defined for dimensions grows at a moderate rate, i.e., the sequence of problems is tractable. Here, we focus on the situation where the space of available information consists of all linear functionals, and the problems are defined as linear operator mappings between Hilbert spaces. We unify the proofs of known tractability results and generalize a number of existing results. These generalizations are expressed as five theorems that provide equivalent conditions for (strong) tractability in terms of sums of functions of the singular values of the solution operators.
{"title":"A unified treatment of tractability for approximation problems defined on Hilbert spaces","authors":"Onyekachi Emenike , Fred J. Hickernell , Peter Kritzer","doi":"10.1016/j.jco.2024.101856","DOIUrl":"10.1016/j.jco.2024.101856","url":null,"abstract":"<div><p>A large literature specifies conditions under which the information complexity for a sequence of numerical problems defined for dimensions <span><math><mn>1</mn><mo>,</mo><mn>2</mn><mo>,</mo><mo>…</mo></math></span> grows at a moderate rate, i.e., the sequence of problems is <em>tractable</em>. Here, we focus on the situation where the space of available information consists of all linear functionals, and the problems are defined as linear operator mappings between Hilbert spaces. We unify the proofs of known tractability results and generalize a number of existing results. These generalizations are expressed as five theorems that provide equivalent conditions for (strong) tractability in terms of sums of functions of the singular values of the solution operators.</p></div>","PeriodicalId":50227,"journal":{"name":"Journal of Complexity","volume":"84 ","pages":"Article 101856"},"PeriodicalIF":1.7,"publicationDate":"2024-04-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140799561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-17DOI: 10.1016/j.jco.2024.101854
Santhosh George, Indra Bate, Muniyasamy M, Chandhini G, Kedarnath Senapati
Ezquerro and Hernandez (2009) studied a modified Chebyshev's method to solve a nonlinear equation approximately in the Banach space setting where the convergence analysis utilizes Taylor series expansion and hence requires the existence of at least fourth-order Fréchet derivative of the involved operator. No error estimate on the error distance was given in their work. In this paper, we obtained the convergence order and error estimate of the error distance without Taylor series expansion. We have made assumptions only on the involved operator and its first and second Fréchet derivative. So, we extend the applicability of the modified Chebyshev's method. Further, we compare the modified Chebyshev method's efficiency index and dynamics with other similar methods. Numerical examples validate the theoretical results.
{"title":"Enhancing the applicability of Chebyshev-like method","authors":"Santhosh George, Indra Bate, Muniyasamy M, Chandhini G, Kedarnath Senapati","doi":"10.1016/j.jco.2024.101854","DOIUrl":"https://doi.org/10.1016/j.jco.2024.101854","url":null,"abstract":"<div><p>Ezquerro and Hernandez (2009) studied a modified Chebyshev's method to solve a nonlinear equation approximately in the Banach space setting where the convergence analysis utilizes Taylor series expansion and hence requires the existence of at least fourth-order Fréchet derivative of the involved operator. No error estimate on the error distance was given in their work. In this paper, we obtained the convergence order and error estimate of the error distance without Taylor series expansion. We have made assumptions only on the involved operator and its first and second Fréchet derivative. So, we extend the applicability of the modified Chebyshev's method. Further, we compare the modified Chebyshev method's efficiency index and dynamics with other similar methods. Numerical examples validate the theoretical results.</p></div>","PeriodicalId":50227,"journal":{"name":"Journal of Complexity","volume":"83 ","pages":"Article 101854"},"PeriodicalIF":1.7,"publicationDate":"2024-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140619311","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-16DOI: 10.1016/j.jco.2024.101855
Michael Gnewuch
We improve the best known upper bound for the bracketing number of d-dimensional axis-parallel boxes anchored in 0 (or, put differently, of lower left orthants intersected with the d-dimensional unit cube ). More precisely, we provide a better estimate for the cardinality of an algorithmic bracketing cover construction due to Eric Thiémard, which forms the core of his algorithm to approximate the star discrepancy of arbitrary point sets from Thiémard (2001) [22]. Moreover, the new upper bound for the bracketing number of anchored axis-parallel boxes yields an improved upper estimate for the bracketing number of arbitrary axis-parallel boxes in . In our upper bounds all constants are fully explicit.
我们改进了锚定在 0 的 d 维轴平行盒(或者换句话说,与 d 维单位立方体 [0,1]d 相交的左下正交)的已知括弧数上限。更确切地说,我们为埃里克-蒂埃玛尔(Eric Thiémard)提出的括号盖构造的心数提供了一个更好的估计,该构造构成了蒂埃玛尔(Thiémard)(2001)[22]中近似任意点集星形差异算法的核心。此外,锚定轴平行盒的括弧数的新上界可以改进[0,1]d 中任意轴平行盒的括弧数的上界估计。在我们的上界中,所有常数都是完全明确的。
{"title":"Improved bounds for the bracketing number of orthants or revisiting an algorithm of Thiémard to compute bounds for the star discrepancy","authors":"Michael Gnewuch","doi":"10.1016/j.jco.2024.101855","DOIUrl":"https://doi.org/10.1016/j.jco.2024.101855","url":null,"abstract":"<div><p>We improve the best known upper bound for the bracketing number of <em>d</em>-dimensional axis-parallel boxes anchored in 0 (or, put differently, of lower left orthants intersected with the <em>d</em>-dimensional unit cube <span><math><msup><mrow><mo>[</mo><mn>0</mn><mo>,</mo><mn>1</mn><mo>]</mo></mrow><mrow><mi>d</mi></mrow></msup></math></span>). More precisely, we provide a better estimate for the cardinality of an algorithmic bracketing cover construction due to Eric Thiémard, which forms the core of his algorithm to approximate the star discrepancy of arbitrary point sets from Thiémard (2001) <span>[22]</span>. Moreover, the new upper bound for the bracketing number of anchored axis-parallel boxes yields an improved upper estimate for the bracketing number of arbitrary axis-parallel boxes in <span><math><msup><mrow><mo>[</mo><mn>0</mn><mo>,</mo><mn>1</mn><mo>]</mo></mrow><mrow><mi>d</mi></mrow></msup></math></span>. In our upper bounds all constants are fully explicit.</p></div>","PeriodicalId":50227,"journal":{"name":"Journal of Complexity","volume":"83 ","pages":"Article 101855"},"PeriodicalIF":1.7,"publicationDate":"2024-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0885064X24000323/pdfft?md5=23c928b5ffffc6732ad1f4739311a07b&pid=1-s2.0-S0885064X24000323-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140619310","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-24DOI: 10.1016/j.jco.2024.101853
Markus Holzleitner , Sergei V. Pereverzyev
This article offers a comprehensive treatment of polynomial functional regression, culminating in the establishment of a novel finite sample bound. This bound encompasses various aspects, including general smoothness conditions, capacity conditions, and regularization techniques. In doing so, it extends and generalizes several findings from the context of linear functional regression as well. We also provide numerical evidence that using higher order polynomial terms can lead to an improved performance.
{"title":"On regularized polynomial functional regression","authors":"Markus Holzleitner , Sergei V. Pereverzyev","doi":"10.1016/j.jco.2024.101853","DOIUrl":"https://doi.org/10.1016/j.jco.2024.101853","url":null,"abstract":"<div><p>This article offers a comprehensive treatment of polynomial functional regression, culminating in the establishment of a novel finite sample bound. This bound encompasses various aspects, including general smoothness conditions, capacity conditions, and regularization techniques. In doing so, it extends and generalizes several findings from the context of linear functional regression as well. We also provide numerical evidence that using higher order polynomial terms can lead to an improved performance.</p></div>","PeriodicalId":50227,"journal":{"name":"Journal of Complexity","volume":"83 ","pages":"Article 101853"},"PeriodicalIF":1.7,"publicationDate":"2024-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140309915","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-18DOI: 10.1016/j.jco.2024.101851
Robert J. Kunsch
We study the numerical integration of functions from isotropic Sobolev spaces using finitely many function evaluations within randomized algorithms, aiming for the smallest possible probabilistic error guarantee at confidence level . For spaces consisting of continuous functions, non-linear Monte Carlo methods with optimal confidence properties have already been known, in few cases even linear methods that succeed in that respect. In this paper we promote a method called stratified control variates (SCV) and by it show that already linear methods achieve optimal probabilistic error rates in the high smoothness regime without the need to adjust algorithmic parameters to the uncertainty δ. We also analyse a version of SCV in the low smoothness regime where may contain functions with singularities. Here, we observe a polynomial dependence of the error on in contrast to the logarithmic dependence in the high smoothness regime.
{"title":"Linear Monte Carlo quadrature with optimal confidence intervals","authors":"Robert J. Kunsch","doi":"10.1016/j.jco.2024.101851","DOIUrl":"https://doi.org/10.1016/j.jco.2024.101851","url":null,"abstract":"<div><p>We study the numerical integration of functions from isotropic Sobolev spaces <span><math><msubsup><mrow><mi>W</mi></mrow><mrow><mi>p</mi></mrow><mrow><mi>s</mi></mrow></msubsup><mo>(</mo><msup><mrow><mo>[</mo><mn>0</mn><mo>,</mo><mn>1</mn><mo>]</mo></mrow><mrow><mi>d</mi></mrow></msup><mo>)</mo></math></span> using finitely many function evaluations within randomized algorithms, aiming for the smallest possible probabilistic error guarantee <span><math><mi>ε</mi><mo>></mo><mn>0</mn></math></span> at confidence level <span><math><mn>1</mn><mo>−</mo><mi>δ</mi><mo>∈</mo><mo>(</mo><mn>0</mn><mo>,</mo><mn>1</mn><mo>)</mo></math></span>. For spaces consisting of continuous functions, non-linear Monte Carlo methods with optimal confidence properties have already been known, in few cases even linear methods that succeed in that respect. In this paper we promote a method called <em>stratified control variates</em> (SCV) and by it show that already linear methods achieve optimal probabilistic error rates in the high smoothness regime without the need to adjust algorithmic parameters to the uncertainty <em>δ</em>. We also analyse a version of SCV in the low smoothness regime where <span><math><msubsup><mrow><mi>W</mi></mrow><mrow><mi>p</mi></mrow><mrow><mi>s</mi></mrow></msubsup><mo>(</mo><msup><mrow><mo>[</mo><mn>0</mn><mo>,</mo><mn>1</mn><mo>]</mo></mrow><mrow><mi>d</mi></mrow></msup><mo>)</mo></math></span> may contain functions with singularities. Here, we observe a polynomial dependence of the error on <span><math><msup><mrow><mi>δ</mi></mrow><mrow><mo>−</mo><mn>1</mn></mrow></msup></math></span> in contrast to the logarithmic dependence in the high smoothness regime.</p></div>","PeriodicalId":50227,"journal":{"name":"Journal of Complexity","volume":"83 ","pages":"Article 101851"},"PeriodicalIF":1.7,"publicationDate":"2024-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0885064X24000281/pdfft?md5=6b29dfcc17f60ddbf6ee19d289e21700&pid=1-s2.0-S0885064X24000281-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140180612","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-16DOI: 10.1016/j.jco.2024.101852
François Clément , Carola Doerr , Luís Paquete
Building upon the exact methods presented in our earlier work (2022) [5], we introduce a heuristic approach for the star discrepancy subset selection problem. The heuristic gradually improves the current-best subset by replacing one of its elements at a time. While it does not necessarily return an optimal solution, we obtain promising results for all tested dimensions. For example, for moderate sizes , we obtain point sets in dimension 6 with star discrepancy up to 35% better than that of the first n points of the Sobol' sequence. Our heuristic works in all dimensions, the main limitation being the precision of the discrepancy calculation algorithms. We provide a comparison with an energy functional introduced by Steinerberger (2019) [31], showing that our heuristic performs better on all tested instances. Finally, our results give further empirical information on inverse star discrepancy conjectures.
{"title":"Heuristic approaches to obtain low-discrepancy point sets via subset selection","authors":"François Clément , Carola Doerr , Luís Paquete","doi":"10.1016/j.jco.2024.101852","DOIUrl":"https://doi.org/10.1016/j.jco.2024.101852","url":null,"abstract":"<div><p>Building upon the exact methods presented in our earlier work (2022) <span>[5]</span>, we introduce a heuristic approach for the star discrepancy subset selection problem. The heuristic gradually improves the current-best subset by replacing one of its elements at a time. While it does not necessarily return an optimal solution, we obtain promising results for all tested dimensions. For example, for moderate sizes <span><math><mn>30</mn><mo>≤</mo><mi>n</mi><mo>≤</mo><mn>240</mn></math></span>, we obtain point sets in dimension 6 with <span><math><msub><mrow><mi>L</mi></mrow><mrow><mo>∞</mo></mrow></msub></math></span> star discrepancy up to 35% better than that of the first <em>n</em> points of the Sobol' sequence. Our heuristic works in all dimensions, the main limitation being the precision of the discrepancy calculation algorithms. We provide a comparison with an energy functional introduced by Steinerberger (2019) <span>[31]</span>, showing that our heuristic performs better on all tested instances. Finally, our results give further empirical information on inverse star discrepancy conjectures.</p></div>","PeriodicalId":50227,"journal":{"name":"Journal of Complexity","volume":"83 ","pages":"Article 101852"},"PeriodicalIF":1.7,"publicationDate":"2024-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0885064X24000293/pdfft?md5=026f36f25d20579c91a0fc64a95356e5&pid=1-s2.0-S0885064X24000293-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140190621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-13DOI: 10.1016/j.jco.2024.101842
Chenxu Pang , Xiaojie Wang , Yue Wu
This article investigates the weak approximation towards the invariant measure of semi-linear stochastic differential equations (SDEs) under non-globally Lipschitz coefficients. For this purpose, we propose a linear-theta-projected Euler (LTPE) scheme, which also admits an invariant measure, to handle the potential influence of the linear stiffness. Under certain assumptions, both the SDE and the corresponding LTPE method are shown to converge exponentially to the underlying invariant measures, respectively. Moreover, with time-independent regularity estimates for the corresponding Kolmogorov equation, the weak error between the numerical invariant measure and the original one can be guaranteed with convergence of order one. In terms of computational complexity, the proposed ergodicity preserving scheme with the nonlinearity explicitly treated has a significant advantage over the ergodicity preserving implicit Euler method in the literature. Numerical experiments are provided to verify our theoretical findings.
{"title":"Linear implicit approximations of invariant measures of semi-linear SDEs with non-globally Lipschitz coefficients","authors":"Chenxu Pang , Xiaojie Wang , Yue Wu","doi":"10.1016/j.jco.2024.101842","DOIUrl":"https://doi.org/10.1016/j.jco.2024.101842","url":null,"abstract":"<div><p>This article investigates the weak approximation towards the invariant measure of semi-linear stochastic differential equations (SDEs) under non-globally Lipschitz coefficients. For this purpose, we propose a linear-theta-projected Euler (LTPE) scheme, which also admits an invariant measure, to handle the potential influence of the linear stiffness. Under certain assumptions, both the SDE and the corresponding LTPE method are shown to converge exponentially to the underlying invariant measures, respectively. Moreover, with time-independent regularity estimates for the corresponding Kolmogorov equation, the weak error between the numerical invariant measure and the original one can be guaranteed with convergence of order one. In terms of computational complexity, the proposed ergodicity preserving scheme with the nonlinearity explicitly treated has a significant advantage over the ergodicity preserving implicit Euler method in the literature. Numerical experiments are provided to verify our theoretical findings.</p></div>","PeriodicalId":50227,"journal":{"name":"Journal of Complexity","volume":"83 ","pages":"Article 101842"},"PeriodicalIF":1.7,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0885064X24000190/pdfft?md5=196a33f1ce0b753c885d6d05ad1d70a4&pid=1-s2.0-S0885064X24000190-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140188091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-13DOI: 10.1016/j.jco.2024.101840
David Krieg , Peter Kritzer
We consider linear problems in the worst-case setting. That is, given a linear operator and a pool of admissible linear measurements, we want to approximate the operator uniformly on a convex and balanced set by means of algorithms using at most n such measurements. It is known that, in general, linear algorithms do not yield an optimal approximation. However, as we show here, an optimal approximation can always be obtained with a homogeneous algorithm. This is of interest for two reasons. First, the homogeneity allows us to extend any error bound on the unit ball to the full input space. Second, homogeneous algorithms are better suited to tackle problems on cones, a scenario far less understood than the classical situation of balls. We use the optimality of homogeneous algorithms to prove solvability for a family of problems defined on cones. We illustrate our results by several examples.
我们考虑的是最坏情况下的线性问题。也就是说,给定一个线性算子和一组可接受的线性测量值,我们希望通过最多使用 n 个此类测量值的算法,在一个凸平衡集合上均匀地近似算子。众所周知,一般来说,线性算法不会产生最佳近似值。然而,正如我们在此所展示的,同质算法总能获得最佳近似值。我们之所以对此感兴趣,有两个原因。首先,同质算法允许我们将单位球上的任何误差约束扩展到整个输入空间。其次,同质算法更适合解决锥体上的问题,而对锥体问题的理解远不如对球的经典理解。我们利用同构算法的最优性来证明定义在圆锥上的一系列问题的可解性。我们通过几个例子来说明我们的结果。
{"title":"Homogeneous algorithms and solvable problems on cones","authors":"David Krieg , Peter Kritzer","doi":"10.1016/j.jco.2024.101840","DOIUrl":"https://doi.org/10.1016/j.jco.2024.101840","url":null,"abstract":"<div><p>We consider linear problems in the worst-case setting. That is, given a linear operator and a pool of admissible linear measurements, we want to approximate the operator uniformly on a convex and balanced set by means of algorithms using at most <em>n</em> such measurements. It is known that, in general, linear algorithms do not yield an optimal approximation. However, as we show here, an optimal approximation can always be obtained with a homogeneous algorithm. This is of interest for two reasons. First, the homogeneity allows us to extend any error bound on the unit ball to the full input space. Second, homogeneous algorithms are better suited to tackle problems on cones, a scenario far less understood than the classical situation of balls. We use the optimality of homogeneous algorithms to prove solvability for a family of problems defined on cones. We illustrate our results by several examples.</p></div>","PeriodicalId":50227,"journal":{"name":"Journal of Complexity","volume":"83 ","pages":"Article 101840"},"PeriodicalIF":1.7,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0885064X24000177/pdfft?md5=a93de3c8d5e250c4bbebc0c932ec7f46&pid=1-s2.0-S0885064X24000177-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140180611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-08DOI: 10.1016/j.jco.2024.101841
Simon Foucart , Chunyang Liao
For objects belonging to a known model set and observed through a prescribed linear process, we aim at determining methods to recover linear quantities of these objects that are optimal from a worst-case perspective. Working in a Hilbert setting, we show that, if the model set is the intersection of two hyperellipsoids centered at the origin, then there is an optimal recovery method which is linear. It is specifically given by a constrained regularization procedure whose parameters can be precomputed by semidefinite programming. This general framework can be applied to several scenarios, including the two-space problem and problems involving -inaccurate data. It can also be applied to the problem of recovery from -inaccurate data. For the latter, we reach the conclusion of existence of an optimal recovery method which is linear, again given by constrained regularization, under a computationally verifiable sufficient condition.
{"title":"Radius of information for two intersected centered hyperellipsoids and implications in optimal recovery from inaccurate data","authors":"Simon Foucart , Chunyang Liao","doi":"10.1016/j.jco.2024.101841","DOIUrl":"https://doi.org/10.1016/j.jco.2024.101841","url":null,"abstract":"<div><p>For objects belonging to a known model set and observed through a prescribed linear process, we aim at determining methods to recover linear quantities of these objects that are optimal from a worst-case perspective. Working in a Hilbert setting, we show that, if the model set is the intersection of two hyperellipsoids centered at the origin, then there is an optimal recovery method which is linear. It is specifically given by a constrained regularization procedure whose parameters can be precomputed by semidefinite programming. This general framework can be applied to several scenarios, including the two-space problem and problems involving <span><math><msub><mrow><mi>ℓ</mi></mrow><mrow><mn>2</mn></mrow></msub></math></span>-inaccurate data. It can also be applied to the problem of recovery from <span><math><msub><mrow><mi>ℓ</mi></mrow><mrow><mn>1</mn></mrow></msub></math></span>-inaccurate data. For the latter, we reach the conclusion of existence of an optimal recovery method which is linear, again given by constrained regularization, under a computationally verifiable sufficient condition.</p></div>","PeriodicalId":50227,"journal":{"name":"Journal of Complexity","volume":"83 ","pages":"Article 101841"},"PeriodicalIF":1.7,"publicationDate":"2024-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140123263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}