首页 > 最新文献

Foundations of Computational Mathematics最新文献

英文 中文
Non-parametric Learning of Stochastic Differential Equations with Non-asymptotic Fast Rates of Convergence
IF 3 1区 数学 Q2 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2025-03-06 DOI: 10.1007/s10208-025-09705-x
Riccardo Bonalli, Alessandro Rudi

We propose a novel non-parametric learning paradigm for the identification of drift and diffusion coefficients of multi-dimensional non-linear stochastic differential equations, which relies upon discrete-time observations of the state. The key idea essentially consists of fitting a RKHS-based approximation of the corresponding Fokker–Planck equation to such observations, yielding theoretical estimates of non-asymptotic learning rates which, unlike previous works, become increasingly tighter when the regularity of the unknown drift and diffusion coefficients becomes higher. Our method being kernel-based, offline pre-processing may be profitably leveraged to enable efficient numerical implementation, offering excellent balance between precision and computational complexity.

{"title":"Non-parametric Learning of Stochastic Differential Equations with Non-asymptotic Fast Rates of Convergence","authors":"Riccardo Bonalli, Alessandro Rudi","doi":"10.1007/s10208-025-09705-x","DOIUrl":"https://doi.org/10.1007/s10208-025-09705-x","url":null,"abstract":"<p>We propose a novel non-parametric learning paradigm for the identification of drift and diffusion coefficients of multi-dimensional non-linear stochastic differential equations, which relies upon discrete-time observations of the state. The key idea essentially consists of fitting a RKHS-based approximation of the corresponding Fokker–Planck equation to such observations, yielding theoretical estimates of non-asymptotic learning rates which, unlike previous works, become increasingly tighter when the regularity of the unknown drift and diffusion coefficients becomes higher. Our method being kernel-based, offline pre-processing may be profitably leveraged to enable efficient numerical implementation, offering excellent balance between precision and computational complexity.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"11 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2025-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143569769","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sums of Squares Certificates for Polynomial Moment Inequalities
IF 3 1区 数学 Q2 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2025-03-04 DOI: 10.1007/s10208-025-09703-z
Igor Klep, Victor Magron, Jurij Volčič

This paper introduces and develops the algebraic framework of moment polynomials, which are polynomial expressions in commuting variables and their formal mixed moments. Their positivity and optimization over probability measures supported on semialgebraic sets and subject to moment polynomial constraints is investigated. On the one hand, a positive solution to Hilbert’s 17th problem for pseudo-moments is given. On the other hand, moment polynomials positive on actual measures are shown to be sums of squares and formal moments of squares up to arbitrarily small perturbation of their coefficients. When only measures supported on a bounded semialgebraic set are considered, a stronger algebraic certificate for moment polynomial positivity is derived. This result gives rise to a converging hierarchy of semidefinite programs for moment polynomial optimization. Finally, as an application, two open nonlinear Bell inequalities from quantum physics are settled.

{"title":"Sums of Squares Certificates for Polynomial Moment Inequalities","authors":"Igor Klep, Victor Magron, Jurij Volčič","doi":"10.1007/s10208-025-09703-z","DOIUrl":"https://doi.org/10.1007/s10208-025-09703-z","url":null,"abstract":"<p>This paper introduces and develops the algebraic framework of moment polynomials, which are polynomial expressions in commuting variables and their formal mixed moments. Their positivity and optimization over probability measures supported on semialgebraic sets and subject to moment polynomial constraints is investigated. On the one hand, a positive solution to Hilbert’s 17th problem for pseudo-moments is given. On the other hand, moment polynomials positive on actual measures are shown to be sums of squares and formal moments of squares up to arbitrarily small perturbation of their coefficients. When only measures supported on a bounded semialgebraic set are considered, a stronger algebraic certificate for moment polynomial positivity is derived. This result gives rise to a converging hierarchy of semidefinite programs for moment polynomial optimization. Finally, as an application, two open nonlinear Bell inequalities from quantum physics are settled.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"35 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2025-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143546173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Unfiltered Low-Regularity Integrator for the KdV Equation with Solutions Below $$mathbf{H^1}$$
IF 3 1区 数学 Q2 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2025-03-04 DOI: 10.1007/s10208-025-09702-0
Buyang Li, Yifei Wu

This article is concerned with the construction and analysis of new time discretizations for the KdV equation on a torus for low-regularity solutions below (H^1). New harmonic analysis tools, including averaging approximations to the exponential phase functions and trilinear estimates of the KdV operator, are established for the construction and analysis of time discretizations with higher convergence orders under low-regularity conditions. In addition, new perturbation techniques are introduced to establish stability estimates of time discretizations under low-regularity conditions without using filters when the energy techniques fail. The proposed method is proved to be convergent with order (gamma ) (up to a logarithmic factor) in (L^2) under the regularity condition (uin C([0,T];H^gamma )) for (gamma in (0,1]).

本文关注环上 KdV 方程对于低于 (H^1)的低规则解的新时间离散的构造和分析。本文建立了新的谐波分析工具,包括指数相位函数的平均近似和 KdV 算子的三线性估计,用于构建和分析低规则性条件下具有更高收敛阶数的时间离散。此外,还引入了新的扰动技术,当能量技术失效时,无需使用滤波器即可建立低规则性条件下时间离散的稳定性估计。在 (gamma in C([0,T];H^gamma )) 为 (gamma in (0,1])的规则性条件下,所提出的方法被证明在 (L^2) 中以 (gamma )阶收敛(达到对数因子)。
{"title":"An Unfiltered Low-Regularity Integrator for the KdV Equation with Solutions Below $$mathbf{H^1}$$","authors":"Buyang Li, Yifei Wu","doi":"10.1007/s10208-025-09702-0","DOIUrl":"https://doi.org/10.1007/s10208-025-09702-0","url":null,"abstract":"<p>This article is concerned with the construction and analysis of new time discretizations for the KdV equation on a torus for low-regularity solutions below <span>(H^1)</span>. New harmonic analysis tools, including averaging approximations to the exponential phase functions and trilinear estimates of the KdV operator, are established for the construction and analysis of time discretizations with higher convergence orders under low-regularity conditions. In addition, new perturbation techniques are introduced to establish stability estimates of time discretizations under low-regularity conditions without using filters when the energy techniques fail. The proposed method is proved to be convergent with order <span>(gamma )</span> (up to a logarithmic factor) in <span>(L^2)</span> under the regularity condition <span>(uin C([0,T];H^gamma ))</span> for <span>(gamma in (0,1])</span>.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"59 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2025-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143546174","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Restarts Subject to Approximate Sharpness: A Parameter-Free and Optimal Scheme For First-Order Methods
IF 3 1区 数学 Q2 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2025-02-18 DOI: 10.1007/s10208-024-09673-8
Ben Adcock, Matthew J. Colbrook, Maksym Neyra-Nesterenko

Sharpness is an almost generic assumption in continuous optimization that bounds the distance from minima by objective function suboptimality. It facilitates the acceleration of first-order methods through restarts. However, sharpness involves problem-specific constants that are typically unknown, and restart schemes typically reduce convergence rates. Moreover, these schemes are challenging to apply in the presence of noise or with approximate model classes (e.g., in compressive imaging or learning problems), and they generally assume that the first-order method used produces feasible iterates. We consider the assumption of approximate sharpness, a generalization of sharpness that incorporates an unknown constant perturbation to the objective function error. This constant offers greater robustness (e.g., with respect to noise or relaxation of model classes) for finding approximate minimizers. By employing a new type of search over the unknown constants, we design a restart scheme that applies to general first-order methods and does not require the first-order method to produce feasible iterates. Our scheme maintains the same convergence rate as when the constants are known. The convergence rates we achieve for various first-order methods match the optimal rates or improve on previously established rates for a wide range of problems. We showcase our restart scheme in several examples and highlight potential future applications and developments of our framework and theory.

{"title":"Restarts Subject to Approximate Sharpness: A Parameter-Free and Optimal Scheme For First-Order Methods","authors":"Ben Adcock, Matthew J. Colbrook, Maksym Neyra-Nesterenko","doi":"10.1007/s10208-024-09673-8","DOIUrl":"https://doi.org/10.1007/s10208-024-09673-8","url":null,"abstract":"<p>Sharpness is an almost generic assumption in continuous optimization that bounds the distance from minima by objective function suboptimality. It facilitates the acceleration of first-order methods through <i>restarts</i>. However, sharpness involves problem-specific constants that are typically unknown, and restart schemes typically reduce convergence rates. Moreover, these schemes are challenging to apply in the presence of noise or with approximate model classes (e.g., in compressive imaging or learning problems), and they generally assume that the first-order method used produces feasible iterates. We consider the assumption of <i>approximate sharpness</i>, a generalization of sharpness that incorporates an unknown constant perturbation to the objective function error. This constant offers greater robustness (e.g., with respect to noise or relaxation of model classes) for finding approximate minimizers. By employing a new type of search over the unknown constants, we design a restart scheme that applies to general first-order methods and does not require the first-order method to produce feasible iterates. Our scheme maintains the same convergence rate as when the constants are known. The convergence rates we achieve for various first-order methods match the optimal rates or improve on previously established rates for a wide range of problems. We showcase our restart scheme in several examples and highlight potential future applications and developments of our framework and theory.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"13 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2025-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143443283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multilinear Hyperquiver Representations
IF 3 1区 数学 Q2 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2025-02-14 DOI: 10.1007/s10208-025-09692-z
Tommi Muller, Vidit Nanda, Anna Seigal

We count singular vector tuples of a system of tensors assigned to the edges of a directed hypergraph. To do so, we study the generalisation of quivers to directed hypergraphs. Assigning vector spaces to the nodes of a hypergraph and multilinear maps to its hyperedges gives a hyperquiver representation. Hyperquiver representations generalise quiver representations (where all hyperedges are edges) and tensors (where there is only one multilinear map). The singular vectors of a hyperquiver representation are a compatible assignment of vectors to the nodes. We compute the dimension and degree of the variety of singular vectors of a sufficiently generic hyperquiver representation. Our formula specialises to known results that count the singular vectors and eigenvectors of a generic tensor. Lastly, we study a hypergraph generalisation of the inverse tensor eigenvalue problem and solve it algorithmically.

{"title":"Multilinear Hyperquiver Representations","authors":"Tommi Muller, Vidit Nanda, Anna Seigal","doi":"10.1007/s10208-025-09692-z","DOIUrl":"https://doi.org/10.1007/s10208-025-09692-z","url":null,"abstract":"<p>We count singular vector tuples of a system of tensors assigned to the edges of a directed hypergraph. To do so, we study the generalisation of quivers to directed hypergraphs. Assigning vector spaces to the nodes of a hypergraph and multilinear maps to its hyperedges gives a hyperquiver representation. Hyperquiver representations generalise quiver representations (where all hyperedges are edges) and tensors (where there is only one multilinear map). The singular vectors of a hyperquiver representation are a compatible assignment of vectors to the nodes. We compute the dimension and degree of the variety of singular vectors of a sufficiently generic hyperquiver representation. Our formula specialises to known results that count the singular vectors and eigenvectors of a generic tensor. Lastly, we study a hypergraph generalisation of the inverse tensor eigenvalue problem and solve it algorithmically.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"9 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2025-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143417487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Representations of the Symmetric Group are Decomposable in Polynomial Time
IF 3 1区 数学 Q2 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2025-02-10 DOI: 10.1007/s10208-025-09697-8
Sheehan Olver

We introduce an algorithm to decompose matrix representations of the symmetric group over the reals into irreducible representations, which as a by-product also computes the multiplicities of the irreducible representations. The algorithm applied to a d-dimensional representation of (S_n) is shown to have a complexity of ({mathcal {O}}(n^2 d^3)) operations for determining which irreducible representations are present and their corresponding multiplicities and a further ({mathcal {O}}(n d^4)) operations to fully decompose representations with non-trivial multiplicities. These complexity bounds are pessimistic and in a practical implementation using floating point arithmetic and exploiting sparsity we observe better complexity. We demonstrate this algorithm on the problem of computing multiplicities of two tensor products of irreducible representations (the Kronecker coefficients problem) as well as higher order tensor products. For hook and hook-like irreducible representations the algorithm has polynomial complexity as n increases. We also demonstrate an application to constructing a basis of homogeneous polynomials so that applying a permutation of variables induces an irreducible representation.

{"title":"Representations of the Symmetric Group are Decomposable in Polynomial Time","authors":"Sheehan Olver","doi":"10.1007/s10208-025-09697-8","DOIUrl":"https://doi.org/10.1007/s10208-025-09697-8","url":null,"abstract":"<p>We introduce an algorithm to decompose matrix representations of the symmetric group over the reals into irreducible representations, which as a by-product also computes the multiplicities of the irreducible representations. The algorithm applied to a <i>d</i>-dimensional representation of <span>(S_n)</span> is shown to have a complexity of <span>({mathcal {O}}(n^2 d^3))</span> operations for determining which irreducible representations are present and their corresponding multiplicities and a further <span>({mathcal {O}}(n d^4))</span> operations to fully decompose representations with non-trivial multiplicities. These complexity bounds are pessimistic and in a practical implementation using floating point arithmetic and exploiting sparsity we observe better complexity. We demonstrate this algorithm on the problem of computing multiplicities of two tensor products of irreducible representations (the Kronecker coefficients problem) as well as higher order tensor products. For hook and hook-like irreducible representations the algorithm has polynomial complexity as <i>n</i> increases. We also demonstrate an application to constructing a basis of homogeneous polynomials so that applying a permutation of variables induces an irreducible representation.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"62 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2025-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143385014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Safely Learning Dynamical Systems
IF 3 1区 数学 Q2 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2025-02-04 DOI: 10.1007/s10208-025-09689-8
Amir Ali Ahmadi, Abraar Chaudhry, Vikas Sindhwani, Stephen Tu

A fundamental challenge in learning an unknown dynamical system is to reduce model uncertainty by making measurements while maintaining safety. In this work, we formulate a mathematical definition of what it means to safely learn a dynamical system by sequentially deciding where to initialize the next trajectory. In our framework, the state of the system is required to stay within a safety region for a horizon of T time steps under the action of all dynamical systems that (i) belong to a given initial uncertainty set, and (ii) are consistent with the information gathered so far. For our first set of results, we consider the setting of safely learning a linear dynamical system involving n states. For the case (T=1), we present a linear programming-based algorithm that either safely recovers the true dynamics from at most n trajectories, or certifies that safe learning is impossible. For (T=2), we give a semidefinite representation of the set of safe initial conditions and show that (lceil n/2 rceil ) trajectories generically suffice for safe learning. For (T = infty ), we provide semidefinite representable inner approximations of the set of safe initial conditions and show that one trajectory generically suffices for safe learning. Finally, we extend a number of our results to the cases where the initial uncertainty set contains sparse, low-rank, or permutation matrices, or when the dynamical system involves a control input. Our second set of results concerns the problem of safely learning a general class of nonlinear dynamical systems. For the case (T=1), we give a second-order cone programming based representation of the set of safe initial conditions. For (T=infty ), we provide semidefinite representable inner approximations to the set of safe initial conditions. We show how one can safely collect trajectories and fit a polynomial model of the nonlinear dynamics that is consistent with the initial uncertainty set and best agrees with the observations. We also present extensions of some of our results to the cases where the measurements are noisy or the dynamical system involves disturbances.

{"title":"Safely Learning Dynamical Systems","authors":"Amir Ali Ahmadi, Abraar Chaudhry, Vikas Sindhwani, Stephen Tu","doi":"10.1007/s10208-025-09689-8","DOIUrl":"https://doi.org/10.1007/s10208-025-09689-8","url":null,"abstract":"<p>A fundamental challenge in learning an unknown dynamical system is to reduce model uncertainty by making measurements while maintaining safety. In this work, we formulate a mathematical definition of what it means to safely learn a dynamical system by sequentially deciding where to initialize the next trajectory. In our framework, the state of the system is required to stay within a safety region for a horizon of <i>T</i> time steps under the action of all dynamical systems that (i) belong to a given initial uncertainty set, and (ii) are consistent with the information gathered so far. For our first set of results, we consider the setting of safely learning a linear dynamical system involving <i>n</i> states. For the case <span>(T=1)</span>, we present a linear programming-based algorithm that either safely recovers the true dynamics from at most <i>n</i> trajectories, or certifies that safe learning is impossible. For <span>(T=2)</span>, we give a semidefinite representation of the set of safe initial conditions and show that <span>(lceil n/2 rceil )</span> trajectories generically suffice for safe learning. For <span>(T = infty )</span>, we provide semidefinite representable inner approximations of the set of safe initial conditions and show that one trajectory generically suffices for safe learning. Finally, we extend a number of our results to the cases where the initial uncertainty set contains sparse, low-rank, or permutation matrices, or when the dynamical system involves a control input. Our second set of results concerns the problem of safely learning a general class of nonlinear dynamical systems. For the case <span>(T=1)</span>, we give a second-order cone programming based representation of the set of safe initial conditions. For <span>(T=infty )</span>, we provide semidefinite representable inner approximations to the set of safe initial conditions. We show how one can safely collect trajectories and fit a polynomial model of the nonlinear dynamics that is consistent with the initial uncertainty set and best agrees with the observations. We also present extensions of some of our results to the cases where the measurements are noisy or the dynamical system involves disturbances.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"28 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2025-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143125227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stabilizing Decomposition of Multiparameter Persistence Modules
IF 3 1区 数学 Q2 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2025-01-27 DOI: 10.1007/s10208-025-09695-w
Håvard Bakke Bjerkevik

While decomposition of one-parameter persistence modules behaves nicely, as demonstrated by the algebraic stability theorem, decomposition of multiparameter modules is known to be unstable in a certain precise sense. Until now, it has not been clear that there is any way to get around this and build a meaningful stability theory for multiparameter module decomposition. We introduce new tools, in particular (epsilon )-refinements and (epsilon )-erosion neighborhoods, to start building such a theory. We then define the (epsilon )-pruning of a module, which is a new invariant acting like a “refined barcode” that shows great promise to extract features from a module by approximately decomposing it. Our main theorem can be interpreted as a generalization of the algebraic stability theorem to multiparameter modules up to a factor of 2r, where r is the maximal pointwise dimension of one of the modules. Furthermore, we show that the factor 2r is close to optimal. Finally, we discuss the possibility of strengthening the stability theorem for modules that decompose into pointwise low-dimensional summands, and pose a conjecture phrased purely in terms of basic linear algebra and graph theory that seems to capture the difficulty of doing this. We also show that this conjecture is relevant for other areas of multipersistence, like the computational complexity of approximating the interleaving distance, and recent applications of relative homological algebra to multipersistence.

{"title":"Stabilizing Decomposition of Multiparameter Persistence Modules","authors":"Håvard Bakke Bjerkevik","doi":"10.1007/s10208-025-09695-w","DOIUrl":"https://doi.org/10.1007/s10208-025-09695-w","url":null,"abstract":"<p>While decomposition of one-parameter persistence modules behaves nicely, as demonstrated by the algebraic stability theorem, decomposition of multiparameter modules is known to be unstable in a certain precise sense. Until now, it has not been clear that there is any way to get around this and build a meaningful stability theory for multiparameter module decomposition. We introduce new tools, in particular <span>(epsilon )</span>-refinements and <span>(epsilon )</span>-erosion neighborhoods, to start building such a theory. We then define the <span>(epsilon )</span>-pruning of a module, which is a new invariant acting like a “refined barcode” that shows great promise to extract features from a module by approximately decomposing it. Our main theorem can be interpreted as a generalization of the algebraic stability theorem to multiparameter modules up to a factor of 2<i>r</i>, where <i>r</i> is the maximal pointwise dimension of one of the modules. Furthermore, we show that the factor 2<i>r</i> is close to optimal. Finally, we discuss the possibility of strengthening the stability theorem for modules that decompose into pointwise low-dimensional summands, and pose a conjecture phrased purely in terms of basic linear algebra and graph theory that seems to capture the difficulty of doing this. We also show that this conjecture is relevant for other areas of multipersistence, like the computational complexity of approximating the interleaving distance, and recent applications of relative homological algebra to multipersistence.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"48 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2025-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143049647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimal Regularization for a Data Source
IF 3 1区 数学 Q2 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2025-01-27 DOI: 10.1007/s10208-025-09693-y
Oscar Leong, Eliza O’ Reilly, Yong Sheng Soh, Venkat Chandrasekaran

In optimization-based approaches to inverse problems and to statistical estimation, it is common to augment criteria that enforce data fidelity with a regularizer that promotes desired structural properties in the solution. The choice of a suitable regularizer is typically driven by a combination of prior domain information and computational considerations. Convex regularizers are attractive computationally but they are limited in the types of structure they can promote. On the other hand, nonconvex regularizers are more flexible in the forms of structure they can promote and they have showcased strong empirical performance in some applications, but they come with the computational challenge of solving the associated optimization problems. In this paper, we seek a systematic understanding of the power and the limitations of convex regularization by investigating the following questions: Given a distribution, what is the optimal regularizer for data drawn from the distribution? What properties of a data source govern whether the optimal regularizer is convex? We address these questions for the class of regularizers specified by functionals that are continuous, positively homogeneous, and positive away from the origin. We say that a regularizer is optimal for a data distribution if the Gibbs density with energy given by the regularizer maximizes the population likelihood (or equivalently, minimizes cross-entropy loss) over all regularizer-induced Gibbs densities. As the regularizers we consider are in one-to-one correspondence with star bodies, we leverage dual Brunn-Minkowski theory to show that a radial function derived from a data distribution is akin to a “computational sufficient statistic” as it is the key quantity for identifying optimal regularizers and for assessing the amenability of a data source to convex regularization. Using tools such as (Gamma )-convergence from variational analysis, we show that our results are robust in the sense that the optimal regularizers for a sample drawn from a distribution converge to their population counterparts as the sample size grows large. Finally, we give generalization guarantees for various families of star bodies that recover previous results for polyhedral regularizers (i.e., dictionary learning) and lead to new ones for a variety of classes of star bodies.

{"title":"Optimal Regularization for a Data Source","authors":"Oscar Leong, Eliza O’ Reilly, Yong Sheng Soh, Venkat Chandrasekaran","doi":"10.1007/s10208-025-09693-y","DOIUrl":"https://doi.org/10.1007/s10208-025-09693-y","url":null,"abstract":"<p>In optimization-based approaches to inverse problems and to statistical estimation, it is common to augment criteria that enforce data fidelity with a regularizer that promotes desired structural properties in the solution. The choice of a suitable regularizer is typically driven by a combination of prior domain information and computational considerations. Convex regularizers are attractive computationally but they are limited in the types of structure they can promote. On the other hand, nonconvex regularizers are more flexible in the forms of structure they can promote and they have showcased strong empirical performance in some applications, but they come with the computational challenge of solving the associated optimization problems. In this paper, we seek a systematic understanding of the power and the limitations of convex regularization by investigating the following questions: Given a distribution, what is the optimal regularizer for data drawn from the distribution? What properties of a data source govern whether the optimal regularizer is convex? We address these questions for the class of regularizers specified by functionals that are continuous, positively homogeneous, and positive away from the origin. We say that a regularizer is optimal for a data distribution if the Gibbs density with energy given by the regularizer maximizes the population likelihood (or equivalently, minimizes cross-entropy loss) over all regularizer-induced Gibbs densities. As the regularizers we consider are in one-to-one correspondence with star bodies, we leverage dual Brunn-Minkowski theory to show that a radial function derived from a data distribution is akin to a “computational sufficient statistic” as it is the key quantity for identifying optimal regularizers and for assessing the amenability of a data source to convex regularization. Using tools such as <span>(Gamma )</span>-convergence from variational analysis, we show that our results are robust in the sense that the optimal regularizers for a sample drawn from a distribution converge to their population counterparts as the sample size grows large. Finally, we give generalization guarantees for various families of star bodies that recover previous results for polyhedral regularizers (i.e., dictionary learning) and lead to new ones for a variety of classes of star bodies. </p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"48 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2025-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143049667","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sharp Bounds for Max-sliced Wasserstein Distances 最大切片Wasserstein距离的尖锐界
IF 3 1区 数学 Q2 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2025-01-22 DOI: 10.1007/s10208-025-09690-1
March T. Boedihardjo

We obtain essentially matching upper and lower bounds for the expected max-sliced 1-Wasserstein distance between a probability measure on a separable Hilbert space and its empirical distribution from n samples. By proving a Banach space version of this result, we also obtain an upper bound, that is sharp up to a log factor, for the expected max-sliced 2-Wasserstein distance between a symmetric probability measure (mu ) on a Euclidean space and its symmetrized empirical distribution in terms of the operator norm of the covariance matrix of (mu ) and the diameter of the support of (mu ).

我们从n个样本中获得可分离Hilbert空间上的概率测度与其经验分布之间的期望最大切片1-Wasserstein距离的基本匹配上界和下界。通过证明这一结果的Banach空间版本,我们也得到了欧几里德空间上对称概率测度(mu )与其对称经验分布(以协方差矩阵(mu )的算子范数和(mu )的支撑直径表示)之间的期望最大切2-Wasserstein距离的上界,该上界精确到一个对数因子。
{"title":"Sharp Bounds for Max-sliced Wasserstein Distances","authors":"March T. Boedihardjo","doi":"10.1007/s10208-025-09690-1","DOIUrl":"https://doi.org/10.1007/s10208-025-09690-1","url":null,"abstract":"<p>We obtain essentially matching upper and lower bounds for the expected max-sliced 1-Wasserstein distance between a probability measure on a separable Hilbert space and its empirical distribution from <i>n</i> samples. By proving a Banach space version of this result, we also obtain an upper bound, that is sharp up to a log factor, for the expected max-sliced 2-Wasserstein distance between a symmetric probability measure <span>(mu )</span> on a Euclidean space and its symmetrized empirical distribution in terms of the operator norm of the covariance matrix of <span>(mu )</span> and the diameter of the support of <span>(mu )</span>.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"22 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2025-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143020600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Foundations of Computational Mathematics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1