首页 > 最新文献

SIAM Journal on Optimization最新文献

英文 中文
High Probability Complexity Bounds for Adaptive Step Search Based on Stochastic Oracles 基于随机字典的自适应阶跃搜索的高概率复杂性边界
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-07-02 DOI: 10.1137/22m1512764
Billy Jin, Katya Scheinberg, Miaolan Xie
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2411-2439, September 2024.
Abstract. We consider a step search method for continuous optimization under a stochastic setting where the function values and gradients are available only through inexact probabilistic zeroth- and first-order oracles. (We introduce the term step search for a class of methods, similar to line search, but where step direction can change during the back-tracking procedure.) Unlike the stochastic gradient method and its many variants, the algorithm does not use a prespecified sequence of step sizes but increases or decreases the step size adaptively according to the estimated progress of the algorithm. These oracles capture multiple standard settings including expected loss minimization and zeroth-order optimization. Moreover, our framework is very general and allows the function and gradient estimates to be biased. The proposed algorithm is simple to describe and easy to implement. Under fairly general conditions on the oracles, we derive a high probability tail bound on the iteration complexity of the algorithm when it is applied to nonconvex, convex, and strongly convex (more generally, those satisfying the Polyak-Łojasiewicz (PL) condition) functions. Our analysis strengthens and extends prior results for stochastic step and line search methods.
SIAM 优化期刊》,第 34 卷第 3 期,第 2411-2439 页,2024 年 9 月。 摘要我们考虑了一种随机环境下连续优化的阶跃搜索方法,在这种环境下,函数值和梯度只能通过不精确的概率零阶和一阶奥拉茨获得。(我们引入阶跃搜索这一术语来表示一类方法,与直线搜索类似,但在回溯过程中阶跃方向会发生变化)。与随机梯度法及其多种变体不同,该算法不使用预先设定的步长序列,而是根据算法的估计进度自适应地增减步长。这些算法能捕捉到多种标准设置,包括预期损失最小化和零阶优化。此外,我们的框架非常通用,允许函数和梯度估计有偏差。所提出的算法描述简单,易于实现。在相当宽泛的算法条件下,当算法应用于非凸、凸和强凸(更一般地说,满足 Polyak-Łojasiewicz (PL) 条件的函数)函数时,我们得出了算法迭代复杂度的高概率尾界。我们的分析加强并扩展了先前关于随机步长和直线搜索方法的结果。
{"title":"High Probability Complexity Bounds for Adaptive Step Search Based on Stochastic Oracles","authors":"Billy Jin, Katya Scheinberg, Miaolan Xie","doi":"10.1137/22m1512764","DOIUrl":"https://doi.org/10.1137/22m1512764","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2411-2439, September 2024. <br/> Abstract. We consider a step search method for continuous optimization under a stochastic setting where the function values and gradients are available only through inexact probabilistic zeroth- and first-order oracles. (We introduce the term step search for a class of methods, similar to line search, but where step direction can change during the back-tracking procedure.) Unlike the stochastic gradient method and its many variants, the algorithm does not use a prespecified sequence of step sizes but increases or decreases the step size adaptively according to the estimated progress of the algorithm. These oracles capture multiple standard settings including expected loss minimization and zeroth-order optimization. Moreover, our framework is very general and allows the function and gradient estimates to be biased. The proposed algorithm is simple to describe and easy to implement. Under fairly general conditions on the oracles, we derive a high probability tail bound on the iteration complexity of the algorithm when it is applied to nonconvex, convex, and strongly convex (more generally, those satisfying the Polyak-Łojasiewicz (PL) condition) functions. Our analysis strengthens and extends prior results for stochastic step and line search methods.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141519786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fast Optimization of Charged Particle Dynamics with Damping 带阻尼的带电粒子动力学快速优化
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-07-02 DOI: 10.1137/23m1599045
Weiping Yan, Yu Tang, Gonglin Yuan
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2287-2313, September 2024.
Abstract. In this paper, the convergence analysis of accelerated second-order methods for convex optimization problems is developed from the point of view of autonomous dissipative inertial continuous dynamics in the magnetic field. Different from the classical heavy ball model with damping, we consider the motion of a charged particle in a magnetic field model involving the linear asymptotic vanishing damping. It is a coupled ordinary differential system by adding the magnetic coupled term [math] to the heavy ball system with [math]. In order to develop fast optimization methods, our first contribution is to prove the global existence and uniqueness of a smooth solution under certain regularity conditions of this system via the Banach fixed point theorem. Our second contribution is to establish the convergence rate of corresponding algorithms involving inertial features via discrete time versions of inertial dynamics under the magnetic field. Meanwhile, the connection of algorithms between the heavy ball model and the motion of a charged particle in a magnetic field model is established.
SIAM 优化期刊》,第 34 卷第 3 期,第 2287-2313 页,2024 年 9 月。 摘要本文从磁场中自主耗散惯性连续动力学的角度出发,对凸优化问题的加速二阶方法进行了收敛性分析。与带阻尼的经典重球模型不同,我们考虑的是带电粒子在磁场中的运动模型,涉及线性渐近消失阻尼。它是一个耦合常微分系统,在重球系统中加入了磁耦合项 [math]。为了开发快速优化方法,我们的第一个贡献是通过巴拿赫定点定理证明了该系统在某些正则性条件下光滑解的全局存在性和唯一性。我们的第二个贡献是通过磁场下惯性动力学的离散时间版本,建立了涉及惯性特征的相应算法的收敛率。同时,建立了重球模型与磁场中带电粒子运动模型之间算法的联系。
{"title":"Fast Optimization of Charged Particle Dynamics with Damping","authors":"Weiping Yan, Yu Tang, Gonglin Yuan","doi":"10.1137/23m1599045","DOIUrl":"https://doi.org/10.1137/23m1599045","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2287-2313, September 2024. <br/> Abstract. In this paper, the convergence analysis of accelerated second-order methods for convex optimization problems is developed from the point of view of autonomous dissipative inertial continuous dynamics in the magnetic field. Different from the classical heavy ball model with damping, we consider the motion of a charged particle in a magnetic field model involving the linear asymptotic vanishing damping. It is a coupled ordinary differential system by adding the magnetic coupled term [math] to the heavy ball system with [math]. In order to develop fast optimization methods, our first contribution is to prove the global existence and uniqueness of a smooth solution under certain regularity conditions of this system via the Banach fixed point theorem. Our second contribution is to establish the convergence rate of corresponding algorithms involving inertial features via discrete time versions of inertial dynamics under the magnetic field. Meanwhile, the connection of algorithms between the heavy ball model and the motion of a charged particle in a magnetic field model is established.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141519780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Nonconvex Relaxation Approach to Low-Rank Matrix Completion of Inexact Observed Data 一种新颖的非凸松弛法,用于完成非精确观测数据的低库矩阵
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-07-02 DOI: 10.1137/22m1543653
Yan Li, Liping Zhang
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2378-2410, September 2024.
Abstract. In recent years, matrix completion has become one of the main concepts in data science. In the process of data acquisition in real applications, in addition to missing data, observed data may be inaccurate. This paper is concerned with such matrix completion of inexact observed data which can be modeled as a rank minimization problem. We adopt the difference of the nuclear norm and the Frobenius norm as an approximation of the rank function, employ Tikhonov-type regularization to preserve the inherent characteristics of original data and control oscillation arising from inexact observations, and then establish a new nonsmooth and nonconvex relaxation model for such low-rank matrix completion. We propose a new accelerated proximal gradient–type algorithm to solve the nonsmooth and nonconvex minimization problem and show that the generated sequence is bounded and globally converges to a critical point of our model. Furthermore, the rate of convergence is given via the Kurdyka–Łojasiewicz property. We evaluate our model and method on visual images and received signal strength fingerprint data in an indoor positioning system. Numerical experiments illustrate that our approach outperforms some state-of-the-art methods, and also verify the efficacy of the Tikhonov-type regularization.
SIAM 优化期刊》,第 34 卷第 3 期,第 2378-2410 页,2024 年 9 月。 摘要近年来,矩阵补全已成为数据科学的主要概念之一。在实际应用的数据采集过程中,除了数据缺失外,观测到的数据也可能不准确。本文关注的就是这种不精确观测数据的矩阵补全问题,它可以建模为一个秩最小化问题。我们采用核规范和 Frobenius 规范之差作为秩函数的近似值,采用 Tikhonov 型正则化来保留原始数据的固有特征并控制非精确观测产生的振荡,然后为这种低秩矩阵补全建立了一个新的非光滑和非凸松弛模型。我们提出了一种新的加速近似梯度型算法来解决非光滑和非凸最小化问题,并证明所生成的序列是有界的,且全局收敛于我们模型的临界点。此外,收敛率是通过 Kurdyka-Łojasiewicz 属性给出的。我们在室内定位系统的视觉图像和接收信号强度指纹数据上评估了我们的模型和方法。数值实验表明,我们的方法优于一些最先进的方法,同时也验证了 Tikhonov 型正则化的有效性。
{"title":"A Novel Nonconvex Relaxation Approach to Low-Rank Matrix Completion of Inexact Observed Data","authors":"Yan Li, Liping Zhang","doi":"10.1137/22m1543653","DOIUrl":"https://doi.org/10.1137/22m1543653","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2378-2410, September 2024. <br/> Abstract. In recent years, matrix completion has become one of the main concepts in data science. In the process of data acquisition in real applications, in addition to missing data, observed data may be inaccurate. This paper is concerned with such matrix completion of inexact observed data which can be modeled as a rank minimization problem. We adopt the difference of the nuclear norm and the Frobenius norm as an approximation of the rank function, employ Tikhonov-type regularization to preserve the inherent characteristics of original data and control oscillation arising from inexact observations, and then establish a new nonsmooth and nonconvex relaxation model for such low-rank matrix completion. We propose a new accelerated proximal gradient–type algorithm to solve the nonsmooth and nonconvex minimization problem and show that the generated sequence is bounded and globally converges to a critical point of our model. Furthermore, the rate of convergence is given via the Kurdyka–Łojasiewicz property. We evaluate our model and method on visual images and received signal strength fingerprint data in an indoor positioning system. Numerical experiments illustrate that our approach outperforms some state-of-the-art methods, and also verify the efficacy of the Tikhonov-type regularization.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141519787","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Path-Following Methods for Maximum a Posteriori Estimators in Bayesian Hierarchical Models: How Estimates Depend on Hyperparameters 贝叶斯层次模型中最大后验估计器的路径跟踪方法:估计值如何取决于超参数
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-07-01 DOI: 10.1137/22m153330x
Zilai Si, Yucong Liu, Alexander Strang
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2201-2230, September 2024.
Abstract. Maximum a posteriori (MAP) estimation, like all Bayesian methods, depends on prior assumptions. These assumptions are often chosen to promote specific features in the recovered estimate. The form of the chosen prior determines the shape of the posterior distribution, thus the behavior of the estimator and complexity of the associated optimization problem. Here, we consider a family of Gaussian hierarchical models with generalized gamma hyperpriors designed to promote sparsity in linear inverse problems. By varying the hyperparameters, we move continuously between priors that act as smoothed [math] penalties with flexible [math], smoothing, and scale. We then introduce a predictor-corrector method that tracks MAP solution paths as the hyperparameters vary. Path following allows a user to explore the space of possible MAP solutions and to test the sensitivity of solutions to changes in the prior assumptions. By tracing paths from a convex region to a nonconvex region, the user could find local minimizers in strongly sparsity promoting regimes that are consistent with a convex relaxation derived using related prior assumptions. We show experimentally that these solutions are less error prone than direct optimization of the nonconvex problem.
SIAM 优化期刊》,第 34 卷第 3 期,第 2201-2230 页,2024 年 9 月。 摘要。最大后验(MAP)估计与所有贝叶斯方法一样,依赖于先验假设。选择这些假设通常是为了促进恢复估计中的特定特征。所选先验的形式决定了后验分布的形状,从而决定了估计器的行为和相关优化问题的复杂性。在这里,我们考虑了一系列具有广义伽马超先验的高斯层次模型,其目的是促进线性逆问题中的稀疏性。通过改变超参数,我们可以在作为平滑[数学]惩罚的先验之间连续移动,这些先验具有灵活的[数学]、平滑和规模。然后,我们引入一种预测器-校正器方法,随着超参数的变化跟踪 MAP 求解路径。路径跟踪允许用户探索可能的 MAP 解的空间,并测试解对先验假设变化的敏感性。通过追踪从凸区域到非凸区域的路径,用户可以在强稀疏性促进状态下找到局部最小值,这些最小值与使用相关先验假设得出的凸松弛一致。我们通过实验证明,这些解决方案比直接优化非凸问题更不易出错。
{"title":"Path-Following Methods for Maximum a Posteriori Estimators in Bayesian Hierarchical Models: How Estimates Depend on Hyperparameters","authors":"Zilai Si, Yucong Liu, Alexander Strang","doi":"10.1137/22m153330x","DOIUrl":"https://doi.org/10.1137/22m153330x","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2201-2230, September 2024. <br/> Abstract. Maximum a posteriori (MAP) estimation, like all Bayesian methods, depends on prior assumptions. These assumptions are often chosen to promote specific features in the recovered estimate. The form of the chosen prior determines the shape of the posterior distribution, thus the behavior of the estimator and complexity of the associated optimization problem. Here, we consider a family of Gaussian hierarchical models with generalized gamma hyperpriors designed to promote sparsity in linear inverse problems. By varying the hyperparameters, we move continuously between priors that act as smoothed [math] penalties with flexible [math], smoothing, and scale. We then introduce a predictor-corrector method that tracks MAP solution paths as the hyperparameters vary. Path following allows a user to explore the space of possible MAP solutions and to test the sensitivity of solutions to changes in the prior assumptions. By tracing paths from a convex region to a nonconvex region, the user could find local minimizers in strongly sparsity promoting regimes that are consistent with a convex relaxation derived using related prior assumptions. We show experimentally that these solutions are less error prone than direct optimization of the nonconvex problem.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141500617","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Feasible Method for General Convex Low-Rank SDP Problems 一般凸低域 SDP 问题的可行方法
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-07-01 DOI: 10.1137/23m1561464
Tianyun Tang, Kim-Chuan Toh
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2169-2200, September 2024.
Abstract. In this work, we consider the low-rank decomposition (SDPR) of general convex semidefinite programming (SDP) problems that contain both a positive semidefinite matrix and a nonnegative vector as variables. We develop a rank-support-adaptive feasible method to solve (SDPR) based on Riemannian optimization. The method is able to escape from a saddle point to ensure its convergence to a global optimal solution for generic constraint vectors. We prove its global convergence and local linear convergence without assuming that the objective function is twice differentiable. Due to the special structure of the low-rank SDP problem, our algorithm can achieve better iteration complexity than existing results for more general smooth nonconvex problems. In order to overcome the degeneracy issues of SDP problems, we develop two strategies based on random perturbation and dual refinement. These techniques enable us to solve some primal degenerate SDP problems efficiently, for example, Lovász theta SDPs. Our work is a step forward in extending the application range of Riemannian optimization approaches for solving SDP problems. Numerical experiments are conducted to verify the efficiency and robustness of our method.
SIAM 优化期刊》,第 34 卷第 3 期,第 2169-2200 页,2024 年 9 月。 摘要在这项工作中,我们考虑了同时包含正半有限矩阵和非负向量作为变量的一般凸半有限编程(SDP)问题的低阶分解(SDPR)。我们开发了一种基于黎曼优化的秩支持自适应可行方法来求解(SDPR)。该方法能够摆脱鞍点,确保收敛到一般约束向量的全局最优解。我们证明了它的全局收敛性和局部线性收敛性,而无需假设目标函数是二次微分的。由于低阶 SDP 问题的特殊结构,我们的算法比现有的更一般的平滑非凸问题的迭代复杂度更高。为了克服 SDP 问题的退化问题,我们开发了基于随机扰动和对偶细化的两种策略。这些技术使我们能够高效地解决一些原始退化 SDP 问题,例如 Lovász theta SDP。我们的工作在扩展黎曼优化方法解决 SDP 问题的应用范围方面向前迈出了一步。我们进行了数值实验来验证我们方法的效率和稳健性。
{"title":"A Feasible Method for General Convex Low-Rank SDP Problems","authors":"Tianyun Tang, Kim-Chuan Toh","doi":"10.1137/23m1561464","DOIUrl":"https://doi.org/10.1137/23m1561464","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2169-2200, September 2024. <br/> Abstract. In this work, we consider the low-rank decomposition (SDPR) of general convex semidefinite programming (SDP) problems that contain both a positive semidefinite matrix and a nonnegative vector as variables. We develop a rank-support-adaptive feasible method to solve (SDPR) based on Riemannian optimization. The method is able to escape from a saddle point to ensure its convergence to a global optimal solution for generic constraint vectors. We prove its global convergence and local linear convergence without assuming that the objective function is twice differentiable. Due to the special structure of the low-rank SDP problem, our algorithm can achieve better iteration complexity than existing results for more general smooth nonconvex problems. In order to overcome the degeneracy issues of SDP problems, we develop two strategies based on random perturbation and dual refinement. These techniques enable us to solve some primal degenerate SDP problems efficiently, for example, Lovász theta SDPs. Our work is a step forward in extending the application range of Riemannian optimization approaches for solving SDP problems. Numerical experiments are conducted to verify the efficiency and robustness of our method.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141519783","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Linear Convergence of Forward-Backward Accelerated Algorithms without Knowledge of the Modulus of Strong Convexity 不知道强凸模的正向-反向加速算法的线性收敛性
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-06-19 DOI: 10.1137/23m158111x
Bowen Li, Bin Shi, Ya-xiang Yuan
SIAM Journal on Optimization, Volume 34, Issue 2, Page 2150-2168, June 2024.
Abstract. A significant milestone in modern gradient-based optimization was achieved with the development of Nesterov’s accelerated gradient descent (NAG) method. This forward-backward technique has been further advanced with the introduction of its proximal generalization, commonly known as the fast iterative shrinkage-thresholding algorithm (FISTA), which enjoys widespread application in image science and engineering. Nonetheless, it remains unclear whether both NAG and FISTA exhibit linear convergence for strongly convex functions. Remarkably, these algorithms demonstrate convergence without requiring any prior knowledge of strongly convex modulus, and this intriguing characteristic has been acknowledged as an open problem in the comprehensive review [A. Chambolle and T. Pock, Acta Numer., 25 (2016), pp. 161–319]. In this paper, we address this question by utilizing the high-resolution ordinary differential equation (ODE) framework. Expanding upon the established phase-space representation, we emphasize the distinctive approach employed in crafting the Lyapunov function, which involves a dynamically adapting coefficient of kinetic energy that evolves throughout the iterations. Furthermore, we highlight that the linear convergence of both NAG and FISTA is independent of the parameter [math]. Additionally, we demonstrate that the square of the proximal subgradient norm likewise advances toward linear convergence.
SIAM 优化期刊》,第 34 卷第 2 期,第 2150-2168 页,2024 年 6 月。 摘要随着涅斯捷罗夫加速梯度下降(NAG)方法的发展,现代基于梯度的优化技术实现了一个重要的里程碑。随着其近似广义方法(通常称为快速迭代收缩阈值算法(FISTA))的引入,这种前向后向技术得到了进一步发展,并在图像科学与工程领域得到了广泛应用。然而,对于强凸函数,NAG 和 FISTA 是否表现出线性收敛性仍不清楚。值得注意的是,这些算法不需要任何强凸模的先验知识就能表现出收敛性,而这一引人入胜的特性在综合评论中被认为是一个开放性问题[A. Chambolle and T. Pock, Acta Numer., 25 (2016), pp.161-319]。在本文中,我们利用高分辨率常微分方程(ODE)框架来解决这一问题。在已建立的相空间表示法的基础上,我们强调了在制作 Lyapunov 函数时采用的独特方法,其中涉及在整个迭代过程中动态适应的动能系数。此外,我们还强调,NAG 和 FISTA 的线性收敛与参数 [math] 无关。此外,我们还证明了近似子梯度规范的平方同样会向线性收敛方向发展。
{"title":"Linear Convergence of Forward-Backward Accelerated Algorithms without Knowledge of the Modulus of Strong Convexity","authors":"Bowen Li, Bin Shi, Ya-xiang Yuan","doi":"10.1137/23m158111x","DOIUrl":"https://doi.org/10.1137/23m158111x","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 2, Page 2150-2168, June 2024. <br/> Abstract. A significant milestone in modern gradient-based optimization was achieved with the development of Nesterov’s accelerated gradient descent (NAG) method. This forward-backward technique has been further advanced with the introduction of its proximal generalization, commonly known as the fast iterative shrinkage-thresholding algorithm (FISTA), which enjoys widespread application in image science and engineering. Nonetheless, it remains unclear whether both NAG and FISTA exhibit linear convergence for strongly convex functions. Remarkably, these algorithms demonstrate convergence without requiring any prior knowledge of strongly convex modulus, and this intriguing characteristic has been acknowledged as an open problem in the comprehensive review [A. Chambolle and T. Pock, Acta Numer., 25 (2016), pp. 161–319]. In this paper, we address this question by utilizing the high-resolution ordinary differential equation (ODE) framework. Expanding upon the established phase-space representation, we emphasize the distinctive approach employed in crafting the Lyapunov function, which involves a dynamically adapting coefficient of kinetic energy that evolves throughout the iterations. Furthermore, we highlight that the linear convergence of both NAG and FISTA is independent of the parameter [math]. Additionally, we demonstrate that the square of the proximal subgradient norm likewise advances toward linear convergence.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141500616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Approximation Guarantees for Min-Max-Min Robust Optimization and [math]-Adaptability Under Objective Uncertainty 最小-最大-最小稳健优化的近似保证和目标不确定性下的[math]-适应性
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-06-17 DOI: 10.1137/23m1595084
Jannis Kurtz
SIAM Journal on Optimization, Volume 34, Issue 2, Page 2121-2149, June 2024.
Abstract. In this work we investigate the min-max-min robust optimization problem and the k-adaptability robust optimization problem for binary problems with uncertain costs. The idea of the first approach is to calculate a set of k feasible solutions which are worst-case optimal if in each possible scenario the best of the k solutions is implemented. It is known that the min-max-min robust problem can be solved efficiently if k is at least the dimension of the problem, while it is theoretically and computationally hard if k is small. However, nothing is known about the intermediate case, i.e., k lies between one and the dimension of the problem. We approach this open question and present an approximation algorithm which achieves good problem-specific approximation guarantees for the cases where k is close to or a fraction of the dimension. The derived bounds can be used to show that the min-max-min robust problem is solvable in oracle-polynomial time under certain conditions even if k is smaller than the dimension. We extend the previous results to the robust k-adaptability problem. As a consequence we can provide bounds on the number of necessary second-stage policies to approximate the exact two-stage robust problem. We derive an approximation algorithm for the k-adaptability problem which has similar guarantees as for the min-max-min problem. Finally, we test both algorithms on knapsack and shortest path problems. The experiments show that both algorithms calculate solutions with relatively small optimality gap in seconds.
SIAM 优化期刊》,第 34 卷第 2 期,第 2121-2149 页,2024 年 6 月。 摘要在这项工作中,我们研究了具有不确定成本的二元问题的最小-最大-最小鲁棒优化问题和 k-适应性鲁棒优化问题。第一种方法的思路是计算一组 k 个可行解,如果在每种可能的情况下都实施了 k 个解中的最佳解,则这些解都是最坏情况下的最优解。众所周知,如果 k 至少是问题的维度,则最小-最大-最小稳健问题可以高效求解,而如果 k 较小,则理论上和计算上都很困难。然而,对于中间情况,即 k 介于 1 和问题维度之间,我们却一无所知。我们从这一悬而未决的问题入手,提出了一种近似算法,它能在 k 接近维数或维数的几分之一的情况下,实现针对具体问题的良好近似保证。推导出的边界可以用来证明,即使 k 小于维数,最小-最大-最小鲁棒问题在某些条件下也可以在oracle-polynomial 时间内求解。我们将前面的结果扩展到鲁棒 k 适应性问题。因此,我们可以提供近似精确两阶段鲁棒问题所需的第二阶段策略数量的边界。我们为 k 适应性问题推导出了一种近似算法,该算法具有与最小-最大-最小问题类似的保证。最后,我们在knapsack和最短路径问题上测试了这两种算法。实验结果表明,这两种算法都能在几秒钟内计算出最优差距相对较小的解决方案。
{"title":"Approximation Guarantees for Min-Max-Min Robust Optimization and [math]-Adaptability Under Objective Uncertainty","authors":"Jannis Kurtz","doi":"10.1137/23m1595084","DOIUrl":"https://doi.org/10.1137/23m1595084","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 2, Page 2121-2149, June 2024. <br/> Abstract. In this work we investigate the min-max-min robust optimization problem and the k-adaptability robust optimization problem for binary problems with uncertain costs. The idea of the first approach is to calculate a set of k feasible solutions which are worst-case optimal if in each possible scenario the best of the k solutions is implemented. It is known that the min-max-min robust problem can be solved efficiently if k is at least the dimension of the problem, while it is theoretically and computationally hard if k is small. However, nothing is known about the intermediate case, i.e., k lies between one and the dimension of the problem. We approach this open question and present an approximation algorithm which achieves good problem-specific approximation guarantees for the cases where k is close to or a fraction of the dimension. The derived bounds can be used to show that the min-max-min robust problem is solvable in oracle-polynomial time under certain conditions even if k is smaller than the dimension. We extend the previous results to the robust k-adaptability problem. As a consequence we can provide bounds on the number of necessary second-stage policies to approximate the exact two-stage robust problem. We derive an approximation algorithm for the k-adaptability problem which has similar guarantees as for the min-max-min problem. Finally, we test both algorithms on knapsack and shortest path problems. The experiments show that both algorithms calculate solutions with relatively small optimality gap in seconds.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141500631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Parameter-Free Accelerated Gradient Descent for Nonconvex Minimization 用于非凸最小化的无参数加速梯度下降算法
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-06-17 DOI: 10.1137/22m1540934
Naoki Marumo, Akiko Takeda
SIAM Journal on Optimization, Volume 34, Issue 2, Page 2093-2120, June 2024.
Abstract. We propose a new first-order method for minimizing nonconvex functions with a Lipschitz continuous gradient and Hessian. The proposed method is an accelerated gradient descent with two restart mechanisms and finds a solution where the gradient norm is less than [math] in [math] function and gradient evaluations. Unlike existing first-order methods with similar complexity bounds, our algorithm is parameter-free because it requires no prior knowledge of problem-dependent parameters, e.g., the Lipschitz constants and the target accuracy [math]. The main challenge in achieving this advantage is estimating the Lipschitz constant of the Hessian using only first-order information. To this end, we develop a new Hessian-free analysis based on two technical inequalities: a Jensen-type inequality for gradients and an error bound for the trapezoidal rule. Several numerical results illustrate that the proposed method performs comparably to existing algorithms with similar complexity bounds, even without parameter tuning.
SIAM 优化期刊》,第 34 卷第 2 期,第 2093-2120 页,2024 年 6 月。 摘要。我们提出了一种新的一阶方法,用于最小化具有 Lipschitz 连续梯度和 Hessian 的非凸函数。所提出的方法是一种加速梯度下降法,具有两种重启机制,能在[math]函数和梯度评估中找到梯度规范小于[math]的解。与复杂度界限相似的现有一阶方法不同,我们的算法是无参数的,因为它不需要事先知道与问题相关的参数,如 Lipschitz 常量和目标精度[math]。实现这一优势的主要挑战在于仅使用一阶信息来估计赫塞斯的 Lipschitz 常量。为此,我们基于两个技术不等式:梯度的詹森式不等式和梯形法则的误差约束,开发了一种新的无 Hessian 分析方法。几个数值结果表明,即使不调整参数,所提出的方法也能与复杂度界限相似的现有算法相媲美。
{"title":"Parameter-Free Accelerated Gradient Descent for Nonconvex Minimization","authors":"Naoki Marumo, Akiko Takeda","doi":"10.1137/22m1540934","DOIUrl":"https://doi.org/10.1137/22m1540934","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 2, Page 2093-2120, June 2024. <br/> Abstract. We propose a new first-order method for minimizing nonconvex functions with a Lipschitz continuous gradient and Hessian. The proposed method is an accelerated gradient descent with two restart mechanisms and finds a solution where the gradient norm is less than [math] in [math] function and gradient evaluations. Unlike existing first-order methods with similar complexity bounds, our algorithm is parameter-free because it requires no prior knowledge of problem-dependent parameters, e.g., the Lipschitz constants and the target accuracy [math]. The main challenge in achieving this advantage is estimating the Lipschitz constant of the Hessian using only first-order information. To this end, we develop a new Hessian-free analysis based on two technical inequalities: a Jensen-type inequality for gradients and an error bound for the trapezoidal rule. Several numerical results illustrate that the proposed method performs comparably to existing algorithms with similar complexity bounds, even without parameter tuning.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141500632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stochastic Trust-Region and Direct-Search Methods: A Weak Tail Bound Condition and Reduced Sample Sizing 随机信任区域和直接搜索方法:弱尾边界条件和减少样本大小
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-06-14 DOI: 10.1137/22m1543446
F. Rinaldi, L. N. Vicente, D. Zeffiro
SIAM Journal on Optimization, Volume 34, Issue 2, Page 2067-2092, June 2024.
Abstract. Using tail bounds, we introduce a new probabilistic condition for function estimation in stochastic derivative-free optimization (SDFO) which leads to a reduction in the number of samples and eases algorithmic analyses. Moreover, we develop simple stochastic direct-search and trust-region methods for the optimization of a potentially nonsmooth function whose values can only be estimated via stochastic observations. For trial points to be accepted, these algorithms require the estimated function values to yield a sufficient decrease measured in terms of a power larger than 1 of the algoritmic stepsize. Our new tail bound condition is precisely imposed on the reduction estimate used to achieve such a sufficient decrease. This condition allows us to select the stepsize power used for sufficient decrease in such a way that the number of samples needed per iteration is reduced. In previous works, the number of samples necessary for global convergence at every iteration [math] of this type of algorithm was [math], where [math] is the stepsize or trust-region radius. However, using the new tail bound condition, and under mild assumptions on the noise, one can prove that such a number of samples is only [math], where [math] can be made arbitrarily small by selecting the power of the stepsize in the sufficient decrease test arbitrarily close to 1. In the common random number generator setting, a further improvement by a factor of [math] can be obtained. The global convergence properties of the stochastic direct-search and trust-region algorithms are established under the new tail bound condition.
SIAM 优化期刊》,第 34 卷第 2 期,第 2067-2092 页,2024 年 6 月。 摘要。利用尾边界,我们为随机无导数优化(SDFO)中的函数估计引入了一个新的概率条件,从而减少了样本数量并简化了算法分析。此外,我们还开发了简单的随机直接搜索和信任区域方法,用于优化潜在的非光滑函数,该函数的值只能通过随机观测进行估计。要接受试验点,这些算法要求估计的函数值产生足够的下降,下降的幂大于算法步长的 1。我们的新尾部约束条件正是强加在用于实现这种充分下降的缩减估计值上的。通过这一条件,我们可以选择用于充分减小的步长幂,从而减少每次迭代所需的样本数量。在以前的研究中,这类算法每次迭代[math]时全局收敛所需的样本数为[math],其中[math]为步长或信任区域半径。然而,利用新的尾界条件,并在对噪声的温和假设下,我们可以证明这样的样本数仅为[math],其中[math]可以通过在充分减小检验中选择任意接近 1 的步长幂来任意变小。在新的尾界条件下,建立了随机直接搜索算法和信任区域算法的全局收敛特性。
{"title":"Stochastic Trust-Region and Direct-Search Methods: A Weak Tail Bound Condition and Reduced Sample Sizing","authors":"F. Rinaldi, L. N. Vicente, D. Zeffiro","doi":"10.1137/22m1543446","DOIUrl":"https://doi.org/10.1137/22m1543446","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 2, Page 2067-2092, June 2024. <br/> Abstract. Using tail bounds, we introduce a new probabilistic condition for function estimation in stochastic derivative-free optimization (SDFO) which leads to a reduction in the number of samples and eases algorithmic analyses. Moreover, we develop simple stochastic direct-search and trust-region methods for the optimization of a potentially nonsmooth function whose values can only be estimated via stochastic observations. For trial points to be accepted, these algorithms require the estimated function values to yield a sufficient decrease measured in terms of a power larger than 1 of the algoritmic stepsize. Our new tail bound condition is precisely imposed on the reduction estimate used to achieve such a sufficient decrease. This condition allows us to select the stepsize power used for sufficient decrease in such a way that the number of samples needed per iteration is reduced. In previous works, the number of samples necessary for global convergence at every iteration [math] of this type of algorithm was [math], where [math] is the stepsize or trust-region radius. However, using the new tail bound condition, and under mild assumptions on the noise, one can prove that such a number of samples is only [math], where [math] can be made arbitrarily small by selecting the power of the stepsize in the sufficient decrease test arbitrarily close to 1. In the common random number generator setting, a further improvement by a factor of [math] can be obtained. The global convergence properties of the stochastic direct-search and trust-region algorithms are established under the new tail bound condition.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141496177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Efficient Sieving-Based Secant Method for Sparse Optimization Problems with Least-Squares Constraints 基于筛分的 Secant 高效方法,适用于具有最小二乘约束的稀疏优化问题
IF 3.1 1区 数学 Q1 Mathematics Pub Date : 2024-06-13 DOI: 10.1137/23m1594443
Qian Li, Defeng Sun, Yancheng Yuan
{"title":"An Efficient Sieving-Based Secant Method for Sparse Optimization Problems with Least-Squares Constraints","authors":"Qian Li, Defeng Sun, Yancheng Yuan","doi":"10.1137/23m1594443","DOIUrl":"https://doi.org/10.1137/23m1594443","url":null,"abstract":"","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141348630","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
SIAM Journal on Optimization
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1