Pub Date : 2024-12-04DOI: 10.1007/s10208-024-09659-6
Kendrick Shepherd, Deepesh Toshniwal
Given a domain (Omega subset mathbb {R}^n), the de Rham complex of differential forms arises naturally in the study of problems in electromagnetism and fluid mechanics defined on (Omega ), and its discretization helps build stable numerical methods for such problems. For constructing such stable methods, one critical requirement is ensuring that the discrete subcomplex is cohomologically equivalent to the continuous complex. When (Omega ) is a hypercube, we thus require that the discrete subcomplex be exact. Focusing on such (Omega ), we theoretically analyze the discrete de Rham complex built from hierarchical B-spline differential forms, i.e., the discrete differential forms are smooth splines and support adaptive refinements—these properties are key to enabling accurate and efficient numerical simulations. We provide locally-verifiable sufficient conditions that ensure that the discrete spline complex is exact. Numerical tests are presented to support the theoretical results, and the examples discussed include complexes that satisfy our prescribed conditions as well as those that violate them.
{"title":"Locally-Verifiable Sufficient Conditions for Exactness of the Hierarchical B-spline Discrete de Rham Complex in $$mathbb {R}^n$$","authors":"Kendrick Shepherd, Deepesh Toshniwal","doi":"10.1007/s10208-024-09659-6","DOIUrl":"https://doi.org/10.1007/s10208-024-09659-6","url":null,"abstract":"<p>Given a domain <span>(Omega subset mathbb {R}^n)</span>, the de Rham complex of differential forms arises naturally in the study of problems in electromagnetism and fluid mechanics defined on <span>(Omega )</span>, and its discretization helps build stable numerical methods for such problems. For constructing such stable methods, one critical requirement is ensuring that the discrete subcomplex is cohomologically equivalent to the continuous complex. When <span>(Omega )</span> is a hypercube, we thus require that the discrete subcomplex be exact. Focusing on such <span>(Omega )</span>, we theoretically analyze the discrete de Rham complex built from hierarchical B-spline differential forms, i.e., the discrete differential forms are smooth splines and support adaptive refinements—these properties are key to enabling accurate and efficient numerical simulations. We provide locally-verifiable sufficient conditions that ensure that the discrete spline complex is exact. Numerical tests are presented to support the theoretical results, and the examples discussed include complexes that satisfy our prescribed conditions as well as those that violate them.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"82 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142776456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-25DOI: 10.1007/s10208-024-09674-7
Théophile Chaumont-Frelet, Martin Vohralík
We analyze constrained and unconstrained minimization problems on patches of tetrahedra sharing a common vertex with discontinuous piecewise polynomial data of degree p. We show that the discrete minimizers in the spaces of piecewise polynomials of degree p conforming in the (H^1), ({varvec{H}}(textbf{curl})), or ({varvec{H}}({text {div}})) spaces are as good as the minimizers in these entire (infinite-dimensional) Sobolev spaces, up to a constant that is independent of p. These results are useful in the analysis and design of finite element methods, namely for devising stable local commuting projectors and establishing local-best–global-best equivalences in a priori analysis and in the context of a posteriori error estimation. Unconstrained minimization in (H^1) and constrained minimization in ({varvec{H}}({text {div}})) have been previously treated in the literature. Along with improvement of the results in the (H^1) and ({varvec{H}}({text {div}})) cases, our key contribution is the treatment of the ({varvec{H}}(textbf{curl})) framework. This enables us to cover the whole de Rham diagram in three space dimensions in a single setting.
我们分析了共享一个共同顶点的四面体斑块上的有约束和无约束最小化问题,这些斑块具有度数为 p 的不连续片断多项式数据。我们证明了在符合 (H^1)、({varvec{H}}(textbf{curl}))或({varvec{H}}({text {div}}))空间的 p 度分片多项式空间中的离散最小化与这些整个(无限维)Sobolev 空间中的最小化一样好,直到一个与 p 无关的常数。这些结果在有限元方法的分析和设计中非常有用,即在先验分析和后验误差估计中设计稳定的局部换向投影器和建立局部最优-全局最优等价。以前的文献已经讨论过 (H^1) 中的无约束最小化和 ({varvec{H}}({text {div}})) 中的有约束最小化。在改进了(H^1)和({varvec{H}}({text {div}}))情况下的结果的同时,我们的主要贡献在于对({varvec{H}}(textbf{curl}))框架的处理。这使我们能够在一个单一的环境中涵盖三维空间中的整个德拉姆图。
{"title":"Constrained and Unconstrained Stable Discrete Minimizations for p-Robust Local Reconstructions in Vertex Patches in the de Rham Complex","authors":"Théophile Chaumont-Frelet, Martin Vohralík","doi":"10.1007/s10208-024-09674-7","DOIUrl":"https://doi.org/10.1007/s10208-024-09674-7","url":null,"abstract":"<p>We analyze constrained and unconstrained minimization problems on patches of tetrahedra sharing a common vertex with discontinuous piecewise polynomial data of degree <i>p</i>. We show that the discrete minimizers in the spaces of piecewise polynomials of degree <i>p</i> conforming in the <span>(H^1)</span>, <span>({varvec{H}}(textbf{curl}))</span>, or <span>({varvec{H}}({text {div}}))</span> spaces are as good as the minimizers in these entire (infinite-dimensional) Sobolev spaces, up to a constant that is independent of <i>p</i>. These results are useful in the analysis and design of finite element methods, namely for devising stable local commuting projectors and establishing local-best–global-best equivalences in a priori analysis and in the context of a posteriori error estimation. Unconstrained minimization in <span>(H^1)</span> and constrained minimization in <span>({varvec{H}}({text {div}}))</span> have been previously treated in the literature. Along with improvement of the results in the <span>(H^1)</span> and <span>({varvec{H}}({text {div}}))</span> cases, our key contribution is the treatment of the <span>({varvec{H}}(textbf{curl}))</span> framework. This enables us to cover the whole de Rham diagram in three space dimensions in a single setting.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"113 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142713198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-22DOI: 10.1007/s10208-024-09680-9
Alan Edelman
{"title":"Tribute to Nick Higham","authors":"Alan Edelman","doi":"10.1007/s10208-024-09680-9","DOIUrl":"https://doi.org/10.1007/s10208-024-09680-9","url":null,"abstract":"","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"255 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142690541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-20DOI: 10.1007/s10208-024-09681-8
Brendan Keith, Thomas M. Surowiec
The proximal Galerkin finite element method is a high-order, low iteration complexity, nonlinear numerical method that preserves the geometric and algebraic structure of pointwise bound constraints in infinite-dimensional function spaces. This paper introduces the proximal Galerkin method and applies it to solve free boundary problems, enforce discrete maximum principles, and develop a scalable, mesh-independent algorithm for optimal design with pointwise bound constraints. This paper also introduces the latent variable proximal point (LVPP) algorithm, from which the proximal Galerkin method derives. When analyzing the classical obstacle problem, we discover that the underlying variational inequality can be replaced by a sequence of second-order partial differential equations (PDEs) that are readily discretized and solved with, e.g., the proximal Galerkin method. Throughout this work, we arrive at several contributions that may be of independent interest. These include (1) a semilinear PDE we refer to as the entropic Poisson equation; (2) an algebraic/geometric connection between high-order positivity-preserving discretizations and certain infinite-dimensional Lie groups; and (3) a gradient-based, bound-preserving algorithm for two-field, density-based topology optimization. The complete proximal Galerkin methodology combines ideas from nonlinear programming, functional analysis, tropical algebra, and differential geometry and can potentially lead to new synergies among these areas as well as within variational and numerical analysis. Open-source implementations of our methods accompany this work to facilitate reproduction and broader adoption.
{"title":"Proximal Galerkin: A Structure-Preserving Finite Element Method for Pointwise Bound Constraints","authors":"Brendan Keith, Thomas M. Surowiec","doi":"10.1007/s10208-024-09681-8","DOIUrl":"https://doi.org/10.1007/s10208-024-09681-8","url":null,"abstract":"<p>The proximal Galerkin finite element method is a high-order, low iteration complexity, nonlinear numerical method that preserves the geometric and algebraic structure of pointwise bound constraints in infinite-dimensional function spaces. This paper introduces the proximal Galerkin method and applies it to solve free boundary problems, enforce discrete maximum principles, and develop a scalable, mesh-independent algorithm for optimal design with pointwise bound constraints. This paper also introduces the latent variable proximal point (LVPP) algorithm, from which the proximal Galerkin method derives. When analyzing the classical obstacle problem, we discover that the underlying variational <i>inequality</i> can be replaced by a sequence of second-order partial differential <i>equations</i> (PDEs) that are readily discretized and solved with, e.g., the proximal Galerkin method. Throughout this work, we arrive at several contributions that may be of independent interest. These include (1) a semilinear PDE we refer to as the <i>entropic Poisson equation</i>; (2) an algebraic/geometric connection between high-order positivity-preserving discretizations and certain infinite-dimensional Lie groups; and (3) a gradient-based, bound-preserving algorithm for two-field, density-based topology optimization. The complete proximal Galerkin methodology combines ideas from nonlinear programming, functional analysis, tropical algebra, and differential geometry and can potentially lead to new synergies among these areas as well as within variational and numerical analysis. Open-source implementations of our methods accompany this work to facilitate reproduction and broader adoption.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"99 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142678312","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-12DOI: 10.1007/s10208-024-09688-1
Bettina Eick
The theory of group classifications has undergone significant changes in the past 25 years. New methods have been introduced, some difficult problems have been solved and group classifications have become widely available through computer algebra systems. This survey describes the state of the art of the group classification problem, its history, its recent advances and some open problems.
{"title":"Classification of Finite Groups: Recent Developements and Open Problems","authors":"Bettina Eick","doi":"10.1007/s10208-024-09688-1","DOIUrl":"https://doi.org/10.1007/s10208-024-09688-1","url":null,"abstract":"<p>The theory of group classifications has undergone significant changes in the past 25 years. New methods have been introduced, some difficult problems have been solved and group classifications have become widely available through computer algebra systems. This survey describes the state of the art of the group classification problem, its history, its recent advances and some open problems.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"153 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142601447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-11DOI: 10.1007/s10208-024-09686-3
Sadashige Ishida, Hugo Lavenant
We present a discretization of the dynamic optimal transport problem for which we can obtain the convergence rate for the value of the transport cost to its continuous value when the temporal and spatial stepsize vanish. This convergence result does not require any regularity assumption on the measures, though experiments suggest that the rate is not sharp. Via an analysis of the duality gap we also obtain the convergence rates for the gradient of the optimal potentials and the velocity field under mild regularity assumptions. To obtain such rates, we discretize the dual formulation of the dynamic optimal transport problem and use the mature literature related to the error due to discretizing the Hamilton–Jacobi equation.
{"title":"Quantitative Convergence of a Discretization of Dynamic Optimal Transport Using the Dual Formulation","authors":"Sadashige Ishida, Hugo Lavenant","doi":"10.1007/s10208-024-09686-3","DOIUrl":"https://doi.org/10.1007/s10208-024-09686-3","url":null,"abstract":"<p>We present a discretization of the dynamic optimal transport problem for which we can obtain the convergence rate for the value of the transport cost to its continuous value when the temporal and spatial stepsize vanish. This convergence result does not require any regularity assumption on the measures, though experiments suggest that the rate is not sharp. Via an analysis of the duality gap we also obtain the convergence rates for the gradient of the optimal potentials and the velocity field under mild regularity assumptions. To obtain such rates, we discretize the dual formulation of the dynamic optimal transport problem and use the mature literature related to the error due to discretizing the Hamilton–Jacobi equation.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"4 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142599531","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-11DOI: 10.1007/s10208-024-09684-5
Johannes Hoffmann, Tobias Mai, Roland Speicher
We address the noncommutative version of the Edmonds’ problem, which asks to determine the inner rank of a matrix in noncommuting variables. We provide an algorithm for the calculation of this inner rank by relating the problem with the distribution of a basic object in free probability theory, namely operator-valued semicircular elements. We have to solve a matrix-valued quadratic equation, for which we provide precise analytical and numerical control on the fixed point algorithm for solving the equation. Numerical examples show the efficiency of the algorithm.
{"title":"Computing the Noncommutative Inner Rank by Means of Operator-Valued Free Probability Theory","authors":"Johannes Hoffmann, Tobias Mai, Roland Speicher","doi":"10.1007/s10208-024-09684-5","DOIUrl":"https://doi.org/10.1007/s10208-024-09684-5","url":null,"abstract":"<p>We address the noncommutative version of the Edmonds’ problem, which asks to determine the inner rank of a matrix in noncommuting variables. We provide an algorithm for the calculation of this inner rank by relating the problem with the distribution of a basic object in free probability theory, namely operator-valued semicircular elements. We have to solve a matrix-valued quadratic equation, for which we provide precise analytical and numerical control on the fixed point algorithm for solving the equation. Numerical examples show the efficiency of the algorithm.\u0000</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"34 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142599530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-07DOI: 10.1007/s10208-024-09683-6
Philippe Jaming, Martin Rathmair
We consider the problem of reconstructing a function (fin L^2({mathbb R})) given phase-less samples of its Gabor transform, which is defined by
$$begin{aligned} {mathcal {G}}f(x,y) :=2^{frac{1}{4}} int _{mathbb R}f(t) e^{-pi (t-x)^2} e^{-2pi i y t},text{ d }t,quad (x,y)in {mathbb R}^2. end{aligned}$$
More precisely, given sampling positions (Omega subseteq {mathbb R}^2) the task is to reconstruct f (up to global phase) from measurements ({|{mathcal {G}}f(omega )|: ,omega in Omega }). This non-linear inverse problem is known to suffer from severe ill-posedness. As for any other phase retrieval problem, constructive recovery is a notoriously delicate affair due to the lack of convexity. One of the fundamental insights in this line of research is that the connectivity of the measurements is both necessary and sufficient for reconstruction of phase information to be theoretically possible. In this article we propose a reconstruction algorithm which is based on solving two convex problems and, as such, amenable to numerical analysis. We show, empirically as well as analytically, that the scheme accurately reconstructs from noisy data within the connected regime. Moreover, to emphasize the practicability of the algorithm we argue that both convex problems can actually be reformulated as semi-definite programs for which efficient solvers are readily available. The approach is based on ideas from complex analysis, Gabor frame theory as well as matrix completion. As a byproduct, we also obtain improved truncation error for Gabor expensions with Gaussian generators.
我们考虑的问题是,在给定函数 Gabor 变换的无相采样的情况下,重构该函数(f/in L^2({mathbb R})),其定义为:$$begin{aligned} {mathcal {G}}f(x,y) :=2^{frac{1}{4}}}int _{mathbb R}f(t) e^{-pi (t-x)^2} e^{-2pi i y t},text{ d }t,quad (x,y)in {mathbb R}^2.end{aligned}$ 更确切地说,给定采样位置(Omega subseteq {mathbb R}^2)的任务是根据测量结果重建 f(直到全局相位)({|{mathcal {G}}f(omega )|:,omega in Omega })。众所周知,这个非线性逆问题存在严重的问题。与其他任何相位检索问题一样,由于缺乏凸性,构造恢复是一个众所周知的棘手问题。这一研究方向的基本观点之一是,测量的连通性是理论上重建相位信息的必要条件和充分条件。在这篇文章中,我们提出了一种基于求解两个凸问题的重建算法,因此可以进行数值分析。我们通过实证和分析表明,该方案能准确地从连接状态下的噪声数据中进行重建。此外,为了强调算法的实用性,我们认为这两个凸问题实际上都可以重新表述为半定式程序,而半定式程序的高效求解器是现成的。这种方法基于复杂分析、Gabor 框架理论以及矩阵补全的思想。作为副产品,我们还改进了高斯发生器 Gabor 展开的截断误差。
{"title":"Gabor Phase Retrieval via Semidefinite Programming","authors":"Philippe Jaming, Martin Rathmair","doi":"10.1007/s10208-024-09683-6","DOIUrl":"https://doi.org/10.1007/s10208-024-09683-6","url":null,"abstract":"<p>We consider the problem of reconstructing a function <span>(fin L^2({mathbb R}))</span> given phase-less samples of its Gabor transform, which is defined by </p><span>$$begin{aligned} {mathcal {G}}f(x,y) :=2^{frac{1}{4}} int _{mathbb R}f(t) e^{-pi (t-x)^2} e^{-2pi i y t},text{ d }t,quad (x,y)in {mathbb R}^2. end{aligned}$$</span><p>More precisely, given sampling positions <span>(Omega subseteq {mathbb R}^2)</span> the task is to reconstruct <i>f</i> (up to global phase) from measurements <span>({|{mathcal {G}}f(omega )|: ,omega in Omega })</span>. This non-linear inverse problem is known to suffer from severe ill-posedness. As for any other phase retrieval problem, constructive recovery is a notoriously delicate affair due to the lack of convexity. One of the fundamental insights in this line of research is that the connectivity of the measurements is both necessary and sufficient for reconstruction of phase information to be theoretically possible. In this article we propose a reconstruction algorithm which is based on solving two convex problems and, as such, amenable to numerical analysis. We show, empirically as well as analytically, that the scheme accurately reconstructs from noisy data within the connected regime. Moreover, to emphasize the practicability of the algorithm we argue that both convex problems can actually be reformulated as semi-definite programs for which efficient solvers are readily available. The approach is based on ideas from complex analysis, Gabor frame theory as well as matrix completion. As a byproduct, we also obtain improved truncation error for Gabor expensions with Gaussian generators.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"61 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142597481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-31DOI: 10.1007/s10208-024-09687-2
Ren-Cang Li
The NEPv approach has been increasingly used lately for optimization on the Stiefel manifold arising from machine learning. General speaking, the approach first turns the first order optimality condition into a nonlinear eigenvalue problem with eigenvector dependency (NEPv) and then solve the nonlinear problem via some variations of the self-consistent-field (SCF) iteration. The difficulty, however, lies in designing a proper SCF iteration so that a maximizer is found at the end. Currently, each use of the approach is very much individualized, especially in its convergence analysis phase to show that the approach does work or otherwise. Related, the NPDo approach is recently proposed for the sum of coupled traces and it seeks to turn the first order optimality condition into a nonlinear polar decomposition with orthogonal factor dependency (NPDo). In this paper, two unifying frameworks are established, one for each approach. Each framework is built upon a basic assumption, under which globally convergence to a stationary point is guaranteed and during the SCF iterative process that leads to the stationary point, the objective function increases monotonically. Also the notion of atomic function for each approach is proposed, and the atomic functions include commonly used matrix traces of linear and quadratic forms as special ones. It is shown that the basic assumptions of the approaches are satisfied by their respective atomic functions and, more importantly, by convex compositions of their respective atomic functions. Together they provide a large collection of objectives for which either one of approaches or both are guaranteed to work, respectively.
{"title":"A Theory of the NEPv Approach for Optimization on the Stiefel Manifold","authors":"Ren-Cang Li","doi":"10.1007/s10208-024-09687-2","DOIUrl":"https://doi.org/10.1007/s10208-024-09687-2","url":null,"abstract":"<p>The NEPv approach has been increasingly used lately for optimization on the Stiefel manifold arising from machine learning. General speaking, the approach first turns the first order optimality condition into a nonlinear eigenvalue problem with eigenvector dependency (NEPv) and then solve the nonlinear problem via some variations of the self-consistent-field (SCF) iteration. The difficulty, however, lies in designing a proper SCF iteration so that a maximizer is found at the end. Currently, each use of the approach is very much individualized, especially in its convergence analysis phase to show that the approach does work or otherwise. Related, the NPDo approach is recently proposed for the sum of coupled traces and it seeks to turn the first order optimality condition into a nonlinear polar decomposition with orthogonal factor dependency (NPDo). In this paper, two unifying frameworks are established, one for each approach. Each framework is built upon a basic assumption, under which globally convergence to a stationary point is guaranteed and during the SCF iterative process that leads to the stationary point, the objective function increases monotonically. Also the notion of atomic function for each approach is proposed, and the atomic functions include commonly used matrix traces of linear and quadratic forms as special ones. It is shown that the basic assumptions of the approaches are satisfied by their respective atomic functions and, more importantly, by convex compositions of their respective atomic functions. Together they provide a large collection of objectives for which either one of approaches or both are guaranteed to work, respectively.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"67 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142562122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-18DOI: 10.1007/s10208-024-09676-5
Sören Bartels, Alex Kaltenbach
In this paper, we propose a general approach for explicit a posteriori error representation for convex minimization problems using basic convex duality relations. Exploiting discrete orthogonality relations in the space of element-wise constant vector fields as well as a discrete integration-by-parts formula between the Crouzeix–Raviart and the Raviart–Thomas element, all convex duality relations are transferred to a discrete level, making the explicit a posteriori error representation –initially based on continuous arguments only– practicable from a numerical point of view. In addition, we provide a generalized Marini formula that determines a discrete primal solution in terms of a given discrete dual solution. We benchmark all these concepts via the Rudin–Osher–Fatemi model. This leads to an adaptive algorithm that yields a (quasi-optimal) linear convergence rate.
{"title":"Explicit A Posteriori Error Representation for Variational Problems and Application to TV-Minimization","authors":"Sören Bartels, Alex Kaltenbach","doi":"10.1007/s10208-024-09676-5","DOIUrl":"https://doi.org/10.1007/s10208-024-09676-5","url":null,"abstract":"<p>In this paper, we propose a general approach for explicit <i>a posteriori</i> error representation for convex minimization problems using basic convex duality relations. Exploiting discrete orthogonality relations in the space of element-wise constant vector fields as well as a discrete integration-by-parts formula between the Crouzeix–Raviart and the Raviart–Thomas element, all convex duality relations are transferred to a discrete level, making the explicit <i>a posteriori</i> error representation –initially based on continuous arguments only– practicable from a numerical point of view. In addition, we provide a generalized Marini formula that determines a discrete primal solution in terms of a given discrete dual solution. We benchmark all these concepts via the Rudin–Osher–Fatemi model. This leads to an adaptive algorithm that yields a (quasi-optimal) linear convergence rate.\u0000</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"20 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142449554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}