首页 > 最新文献

SIAM Journal on Optimization最新文献

英文 中文
Data-Driven Distributionally Robust Multiproduct Pricing Problems under Pure Characteristics Demand Models 纯特征需求模型下数据驱动的分布稳健型多产品定价问题
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-09-03 DOI: 10.1137/23m1585131
Jie Jiang, Hailin Sun, Xiaojun Chen
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2917-2942, September 2024.
Abstract. This paper considers a multiproduct pricing problem under pure characteristics demand models when the probability distribution of the random parameter in the problem is uncertain. We formulate this problem as a distributionally robust optimization (DRO) problem based on a constructive approach to estimating pure characteristics demand models with pricing by Pang, Su, and Lee. In this model, the consumers’ purchase decision is to maximize their utility. We show that the DRO problem is well-defined, and the objective function is upper semicontinuous by using an equivalent hierarchical form. We also use the data-driven approach to analyze the DRO problem when the ambiguity set, i.e., a set of probability distributions that contains some exact information of the underlying probability distribution, is given by a general moment-based case. We give convergence results as the data size tends to infinity and analyze the quantitative statistical robustness in view of the possible contamination of driven data. Furthermore, we use the Lagrange duality to reformulate the DRO problem as a mathematical program with complementarity constraints, and give a numerical procedure for finding a global solution of the DRO problem under certain specific settings. Finally, we report numerical results that validate the effectiveness and scalability of our approach for the distributionally robust multiproduct pricing problem.
SIAM 优化期刊》,第 34 卷第 3 期,第 2917-2942 页,2024 年 9 月。 摘要本文考虑了在纯特征需求模型下,当问题中随机参数的概率分布不确定时的多产品定价问题。我们根据 Pang、Su 和 Lee 提出的纯特征需求模型定价估计的构造方法,将该问题表述为分布稳健优化(DRO)问题。在该模型中,消费者的购买决策是使其效用最大化。我们利用等效的分层形式证明了 DRO 问题定义明确,目标函数是上半连续的。我们还使用数据驱动法分析了当含糊集(即包含底层概率分布的某些精确信息的概率分布集)由基于矩的一般情况给出时的 DRO 问题。我们给出了数据规模趋于无穷大时的收敛结果,并分析了驱动数据可能受到污染时的定量统计稳健性。此外,我们还利用拉格朗日对偶性将 DRO 问题重新表述为具有互补约束的数学程序,并给出了在某些特定设置下找到 DRO 问题全局解的数值程序。最后,我们报告了数值结果,验证了我们的方法在分布稳健的多产品定价问题上的有效性和可扩展性。
{"title":"Data-Driven Distributionally Robust Multiproduct Pricing Problems under Pure Characteristics Demand Models","authors":"Jie Jiang, Hailin Sun, Xiaojun Chen","doi":"10.1137/23m1585131","DOIUrl":"https://doi.org/10.1137/23m1585131","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2917-2942, September 2024. <br/> Abstract. This paper considers a multiproduct pricing problem under pure characteristics demand models when the probability distribution of the random parameter in the problem is uncertain. We formulate this problem as a distributionally robust optimization (DRO) problem based on a constructive approach to estimating pure characteristics demand models with pricing by Pang, Su, and Lee. In this model, the consumers’ purchase decision is to maximize their utility. We show that the DRO problem is well-defined, and the objective function is upper semicontinuous by using an equivalent hierarchical form. We also use the data-driven approach to analyze the DRO problem when the ambiguity set, i.e., a set of probability distributions that contains some exact information of the underlying probability distribution, is given by a general moment-based case. We give convergence results as the data size tends to infinity and analyze the quantitative statistical robustness in view of the possible contamination of driven data. Furthermore, we use the Lagrange duality to reformulate the DRO problem as a mathematical program with complementarity constraints, and give a numerical procedure for finding a global solution of the DRO problem under certain specific settings. Finally, we report numerical results that validate the effectiveness and scalability of our approach for the distributionally robust multiproduct pricing problem.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142212459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MIP Relaxations in Factorable Programming 可因式编程中的 MIP 放松
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-08-26 DOI: 10.1137/22m1515537
Taotao He, Mohit Tawarmalani
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2856-2882, September 2024.
Abstract. In this paper, we develop new discrete relaxations for nonlinear expressions in factorable programming. We utilize specialized convexification results as well as composite relaxations to develop mixed-integer programming relaxations. Our relaxations rely on ideal formulations of convex hulls of outer-functions over a combinatorial structure that captures local inner-function structure. The resulting relaxations often require fewer variables and are tighter than currently prevalent ones. Finally, we provide computational evidence to demonstrate that our relaxations close approximately 60%–70% of the gap relative to McCormick relaxations and significantly improve the relaxations used in a state-of-the-art solver on various instances involving polynomial functions.
SIAM 优化期刊》,第 34 卷第 3 期,第 2856-2882 页,2024 年 9 月。 摘要本文为可因式编程中的非线性表达式开发了新的离散松弛。我们利用专门的凸化结果以及复合松弛来开发混合整数编程松弛。我们的松弛方法依赖于外函数在组合结构上的凸壳的理想表述,该组合结构捕捉了局部的内函数结构。由此产生的松弛往往需要更少的变量,而且比目前流行的松弛更严密。最后,我们提供了计算证据,证明我们的松弛方法缩小了与麦考密克松弛方法约 60%-70% 的差距,并在涉及多项式函数的各种实例上显著改善了最先进求解器中使用的松弛方法。
{"title":"MIP Relaxations in Factorable Programming","authors":"Taotao He, Mohit Tawarmalani","doi":"10.1137/22m1515537","DOIUrl":"https://doi.org/10.1137/22m1515537","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2856-2882, September 2024. <br/> Abstract. In this paper, we develop new discrete relaxations for nonlinear expressions in factorable programming. We utilize specialized convexification results as well as composite relaxations to develop mixed-integer programming relaxations. Our relaxations rely on ideal formulations of convex hulls of outer-functions over a combinatorial structure that captures local inner-function structure. The resulting relaxations often require fewer variables and are tighter than currently prevalent ones. Finally, we provide computational evidence to demonstrate that our relaxations close approximately 60%–70% of the gap relative to McCormick relaxations and significantly improve the relaxations used in a state-of-the-art solver on various instances involving polynomial functions.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142212461","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On a Differential Generalized Nash Equilibrium Problem with Mean Field Interaction 论具有平均场相互作用的差分广义纳什均衡问题
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-08-23 DOI: 10.1137/22m1489952
Michael Hintermüller, Thomas M. Surowiec, Mike Theiß
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2821-2855, September 2024.
Abstract. We consider a class of [math]-player linear quadratic differential generalized Nash equilibrium problems (GNEPs) with bound constraints on the individual control and state variables. In addition, we assume the individual players’ optimal control problems are coupled through their dynamics and objectives via a time-dependent mean field interaction term. This assumption allows us to model the realistic setting that strategic players in large games cannot observe the individual states of their competitors. We observe that the GNEPs require a constraint qualification, which necessitates sufficient robustness of the individuals, in order to prove the existence of an open-loop pure strategy Nash equilibrium and to derive optimality conditions. In order to gain qualitative insight into the [math]-player game, we assume that players are identical and pass to the limit in [math] to derive a type of first-order constrained mean field game (MFG). We prove that the mean field interaction terms converge to an absolutely continuous curve of probability measures on the set of possible state trajectories. Using variational convergence methods, we show that the optimal control problems converge to a representative agent problem. Under additional regularity assumptions, we provide an explicit form for the mean field term as the solution of a continuity equation and demonstrate the link back to the [math]-player GNEP.
SIAM 优化期刊》,第 34 卷第 3 期,第 2821-2855 页,2024 年 9 月。 摘要。我们考虑了一类[数学]玩家线性二次微分广义纳什均衡问题(GNEPs),其个体控制变量和状态变量都有约束。此外,我们还假定各个玩家的最优控制问题通过与时间相关的均值场交互项,与他们的动态和目标相耦合。这一假设使我们能够模拟大型博弈中战略参与者无法观察到竞争对手个体状态的现实情况。我们发现,GNEPs 需要一个约束条件,这就要求个体具有足够的鲁棒性,从而证明开环纯策略纳什均衡的存在,并推导出最优性条件。为了获得对[math]-玩家博弈的定性认识,我们假设玩家是相同的,并通过[math]中的极限推导出一种一阶约束均值场博弈(MFG)。我们证明,均值场相互作用项收敛于可能状态轨迹集上的概率度量绝对连续曲线。利用变分收敛方法,我们证明了最优控制问题收敛于一个代表性代理问题。在额外的规则性假设下,我们提供了平均场项作为连续性方程解的明确形式,并证明了与[数学]玩家 GNEP 的联系。
{"title":"On a Differential Generalized Nash Equilibrium Problem with Mean Field Interaction","authors":"Michael Hintermüller, Thomas M. Surowiec, Mike Theiß","doi":"10.1137/22m1489952","DOIUrl":"https://doi.org/10.1137/22m1489952","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2821-2855, September 2024. <br/> Abstract. We consider a class of [math]-player linear quadratic differential generalized Nash equilibrium problems (GNEPs) with bound constraints on the individual control and state variables. In addition, we assume the individual players’ optimal control problems are coupled through their dynamics and objectives via a time-dependent mean field interaction term. This assumption allows us to model the realistic setting that strategic players in large games cannot observe the individual states of their competitors. We observe that the GNEPs require a constraint qualification, which necessitates sufficient robustness of the individuals, in order to prove the existence of an open-loop pure strategy Nash equilibrium and to derive optimality conditions. In order to gain qualitative insight into the [math]-player game, we assume that players are identical and pass to the limit in [math] to derive a type of first-order constrained mean field game (MFG). We prove that the mean field interaction terms converge to an absolutely continuous curve of probability measures on the set of possible state trajectories. Using variational convergence methods, we show that the optimal control problems converge to a representative agent problem. Under additional regularity assumptions, we provide an explicit form for the mean field term as the solution of a continuity equation and demonstrate the link back to the [math]-player GNEP.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142212462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MGProx: A Nonsmooth Multigrid Proximal Gradient Method with Adaptive Restriction for Strongly Convex Optimization MGProx:用于强凸优化的带有自适应限制的非光滑多网格近端梯度法
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-08-13 DOI: 10.1137/23m1552140
Andersen Ang, Hans De Sterck, Stephen Vavasis
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2788-2820, September 2024.
Abstract. We study the combination of proximal gradient descent with multigrid for solving a class of possibly nonsmooth strongly convex optimization problems. We propose a multigrid proximal gradient method called MGProx, which accelerates the proximal gradient method by multigrid, based on using hierarchical information of the optimization problem. MGProx applies a newly introduced adaptive restriction operator to simplify the Minkowski sum of subdifferentials of the nondifferentiable objective function across different levels. We provide a theoretical characterization of MGProx. First we show that the MGProx update operator exhibits a fixed-point property. Next, we show that the coarse correction is a descent direction for the fine variable of the original fine level problem in the general nonsmooth case. Last, under some assumptions we provide the convergence rate for the algorithm. In the numerical tests on the elastic obstacle problem, which is an example of a nonsmooth convex optimization problem where the multigrid method can be applied, we show that MGProx has a faster convergence speed than competing methods.
SIAM 优化期刊》,第 34 卷第 3 期,第 2788-2820 页,2024 年 9 月。 摘要我们研究了近似梯度下降法与多网格法的结合,以求解一类可能是非光滑的强凸优化问题。我们提出了一种名为 MGProx 的多网格近似梯度法,它利用优化问题的层次信息,通过多网格加速近似梯度法。MGProx 应用了一种新引入的自适应限制算子,以简化不同层次的无差异目标函数的子差分的闵可夫斯基和。首先,我们证明了 MGProx 更新算子具有定点特性。接着,我们证明了在一般非光滑情况下,粗修正是原始精细问题中精细变量的下降方向。最后,在一些假设条件下,我们给出了算法的收敛速率。弹性障碍物问题是非光滑凸优化问题的一个例子,多网格方法可以应用于该问题,在对该问题的数值测试中,我们证明 MGProx 比其他竞争方法具有更快的收敛速度。
{"title":"MGProx: A Nonsmooth Multigrid Proximal Gradient Method with Adaptive Restriction for Strongly Convex Optimization","authors":"Andersen Ang, Hans De Sterck, Stephen Vavasis","doi":"10.1137/23m1552140","DOIUrl":"https://doi.org/10.1137/23m1552140","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2788-2820, September 2024. <br/> Abstract. We study the combination of proximal gradient descent with multigrid for solving a class of possibly nonsmooth strongly convex optimization problems. We propose a multigrid proximal gradient method called MGProx, which accelerates the proximal gradient method by multigrid, based on using hierarchical information of the optimization problem. MGProx applies a newly introduced adaptive restriction operator to simplify the Minkowski sum of subdifferentials of the nondifferentiable objective function across different levels. We provide a theoretical characterization of MGProx. First we show that the MGProx update operator exhibits a fixed-point property. Next, we show that the coarse correction is a descent direction for the fine variable of the original fine level problem in the general nonsmooth case. Last, under some assumptions we provide the convergence rate for the algorithm. In the numerical tests on the elastic obstacle problem, which is an example of a nonsmooth convex optimization problem where the multigrid method can be applied, we show that MGProx has a faster convergence speed than competing methods.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142212474","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Variational and Strong Variational Convexity in Infinite-Dimensional Variational Analysis 无穷维变分分析中的变分凸性和强变分凸性
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-08-12 DOI: 10.1137/23m1604667
P. D. Khanh, V. V. H. Khoa, B. S. Mordukhovich, V. T. Phat
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2756-2787, September 2024.
Abstract. This paper is devoted to a systematic study and characterizations of the fundamental notions of variational and strong variational convexity for lower semicontinuous functions. While these notions have been quite recently introduced by Rockafellar, the importance of them has already been recognized and documented in finite-dimensional variational analysis and optimization. Here we address general infinite-dimensional settings and derive comprehensive characterizations of both variational and strong variational convexity notions by developing novel techniques, which are essentially different from finite-dimensional counterparts. As a consequence of the obtained characterizations, we establish new quantitative and qualitative relationships between strong variational convexity and tilt stability of local minimizers in appropriate frameworks of Banach spaces.
SIAM 优化期刊》第 34 卷第 3 期第 2756-2787 页,2024 年 9 月。 摘要本文致力于系统研究低半连续函数的变凸性和强变凸性的基本概念及其特征。虽然这些概念是 Rockafellar 最近提出的,但它们的重要性在有限维变分分析和最优化中已经得到了认可和证明。在此,我们将讨论一般的无限维设置,并通过开发与有限维对应概念本质上不同的新技术,推导出变凸性和强变凸性概念的综合特征。根据所获得的特征,我们在巴拿赫空间的适当框架中建立了强变凸性与局部最小化的倾斜稳定性之间新的定量和定性关系。
{"title":"Variational and Strong Variational Convexity in Infinite-Dimensional Variational Analysis","authors":"P. D. Khanh, V. V. H. Khoa, B. S. Mordukhovich, V. T. Phat","doi":"10.1137/23m1604667","DOIUrl":"https://doi.org/10.1137/23m1604667","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2756-2787, September 2024. <br/> Abstract. This paper is devoted to a systematic study and characterizations of the fundamental notions of variational and strong variational convexity for lower semicontinuous functions. While these notions have been quite recently introduced by Rockafellar, the importance of them has already been recognized and documented in finite-dimensional variational analysis and optimization. Here we address general infinite-dimensional settings and derive comprehensive characterizations of both variational and strong variational convexity notions by developing novel techniques, which are essentially different from finite-dimensional counterparts. As a consequence of the obtained characterizations, we establish new quantitative and qualitative relationships between strong variational convexity and tilt stability of local minimizers in appropriate frameworks of Banach spaces.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142212475","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Convergence of Entropy-Regularized Natural Policy Gradient with Linear Function Approximation 采用线性函数逼近的熵细化自然策略梯度的收敛性
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-07-30 DOI: 10.1137/22m1540156
Semih Cayci, Niao He, R. Srikant
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2729-2755, September 2024.
Abstract. Natural policy gradient (NPG) methods, equipped with function approximation and entropy regularization, achieve impressive empirical success in reinforcement learning problems with large state-action spaces. However, their convergence properties and the impact of entropy regularization remain elusive in the function approximation regime. In this paper, we establish finite-time convergence analyses of entropy-regularized NPG with linear function approximation under softmax parameterization. In particular, we prove that entropy-regularized NPG with averaging satisfies the persistence of excitation condition, and achieves a fast convergence rate of [math] up to a function approximation error in regularized Markov decision processes. This convergence result does not require any a priori assumptions on the policies. Furthermore, under mild regularity conditions on the concentrability coefficient and basis vectors, we prove that entropy-regularized NPG exhibits linear convergence up to the compatible function approximation error. Finally, we provide sample complexity results for sample-based NPG with entropy regularization.
SIAM 优化期刊》,第 34 卷第 3 期,第 2729-2755 页,2024 年 9 月。 摘要自然策略梯度(NPG)方法配备了函数逼近和熵正则化,在具有大型状态-动作空间的强化学习问题上取得了令人印象深刻的经验成功。然而,在函数逼近机制中,它们的收敛特性和熵正则化的影响仍然难以捉摸。在本文中,我们建立了软最大参数化条件下线性函数逼近的熵正则化 NPG 的有限时间收敛分析。特别是,我们证明了带平均化的熵规整 NPG 满足激励持久性条件,并在规整马尔可夫决策过程中实现了函数近似误差以内 [math] 的快速收敛率。这一收敛结果不需要对策略做任何先验假设。此外,在同调系数和基向量的温和正则性条件下,我们证明了熵正则化 NPG 在兼容函数近似误差范围内表现出线性收敛性。最后,我们提供了基于样本的熵正则化 NPG 的样本复杂度结果。
{"title":"Convergence of Entropy-Regularized Natural Policy Gradient with Linear Function Approximation","authors":"Semih Cayci, Niao He, R. Srikant","doi":"10.1137/22m1540156","DOIUrl":"https://doi.org/10.1137/22m1540156","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2729-2755, September 2024. <br/> Abstract. Natural policy gradient (NPG) methods, equipped with function approximation and entropy regularization, achieve impressive empirical success in reinforcement learning problems with large state-action spaces. However, their convergence properties and the impact of entropy regularization remain elusive in the function approximation regime. In this paper, we establish finite-time convergence analyses of entropy-regularized NPG with linear function approximation under softmax parameterization. In particular, we prove that entropy-regularized NPG with averaging satisfies the persistence of excitation condition, and achieves a fast convergence rate of [math] up to a function approximation error in regularized Markov decision processes. This convergence result does not require any a priori assumptions on the policies. Furthermore, under mild regularity conditions on the concentrability coefficient and basis vectors, we prove that entropy-regularized NPG exhibits linear convergence up to the compatible function approximation error. Finally, we provide sample complexity results for sample-based NPG with entropy regularization.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863905","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stochastic Trust-Region Algorithm in Random Subspaces with Convergence and Expected Complexity Analyses 随机子空间中的随机信任区域算法及收敛性和预期复杂性分析
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-07-25 DOI: 10.1137/22m1524072
K. J. Dzahini, S. M. Wild
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2671-2699, September 2024.
Abstract. This work proposes a framework for large-scale stochastic derivative-free optimization (DFO) by introducing STARS, a trust-region method based on iterative minimization in random subspaces. This framework is both an algorithmic and theoretical extension of a random subspace derivative-free optimization (RSDFO) framework, and an algorithm for stochastic optimization with random models (STORM). Moreover, like RSDFO, STARS achieves scalability by minimizing interpolation models that approximate the objective in low-dimensional affine subspaces, thus significantly reducing per-iteration costs in terms of function evaluations and yielding strong performance on large-scale stochastic DFO problems. The user-determined dimension of these subspaces, when the latter are defined, for example, by the columns of so-called Johnson–Lindenstrauss transforms, turns out to be independent of the dimension of the problem. For convergence purposes, inspired by the analyses of RSDFO and STORM, both a particular quality of the subspace and the accuracies of random function estimates and models are required to hold with sufficiently high, but fixed, probabilities. Using martingale theory under the latter assumptions, an almost sure global convergence of STARS to a first-order stationary point is shown, and the expected number of iterations required to reach a desired first-order accuracy is proved to be similar to that of STORM and other stochastic DFO algorithms, up to constants.
SIAM 优化期刊》,第 34 卷第 3 期,第 2671-2699 页,2024 年 9 月。 摘要本文通过引入基于随机子空间迭代最小化的信任区域方法 STARS,提出了一种大规模随机无导数优化(DFO)框架。该框架是随机子空间无导数优化(RSDFO)框架和随机模型随机优化算法(STORM)在算法和理论上的扩展。此外,与 RSDFO 一样,STARS 通过最小化在低维仿射子空间中逼近目标的插值模型来实现可扩展性,从而显著降低了函数求值的每次迭代成本,并在大规模随机无导数优化问题上取得了优异的性能。用户确定的这些子空间的维度(例如由所谓的约翰逊-林登斯特劳斯变换的列定义的子空间)与问题的维度无关。受 RSDFO 和 STORM 分析的启发,为了达到收敛的目的,要求子空间的特定质量以及随机函数估计值和模型的精确度以足够高但固定的概率保持不变。在后一种假设条件下使用马丁格尔理论,证明了 STARS 几乎肯定会全局收敛到一阶静止点,并证明了达到所需一阶精度所需的预期迭代次数与 STORM 和其他随机 DFO 算法相似,直至常数。
{"title":"Stochastic Trust-Region Algorithm in Random Subspaces with Convergence and Expected Complexity Analyses","authors":"K. J. Dzahini, S. M. Wild","doi":"10.1137/22m1524072","DOIUrl":"https://doi.org/10.1137/22m1524072","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2671-2699, September 2024. <br/> Abstract. This work proposes a framework for large-scale stochastic derivative-free optimization (DFO) by introducing STARS, a trust-region method based on iterative minimization in random subspaces. This framework is both an algorithmic and theoretical extension of a random subspace derivative-free optimization (RSDFO) framework, and an algorithm for stochastic optimization with random models (STORM). Moreover, like RSDFO, STARS achieves scalability by minimizing interpolation models that approximate the objective in low-dimensional affine subspaces, thus significantly reducing per-iteration costs in terms of function evaluations and yielding strong performance on large-scale stochastic DFO problems. The user-determined dimension of these subspaces, when the latter are defined, for example, by the columns of so-called Johnson–Lindenstrauss transforms, turns out to be independent of the dimension of the problem. For convergence purposes, inspired by the analyses of RSDFO and STORM, both a particular quality of the subspace and the accuracies of random function estimates and models are required to hold with sufficiently high, but fixed, probabilities. Using martingale theory under the latter assumptions, an almost sure global convergence of STARS to a first-order stationary point is shown, and the expected number of iterations required to reach a desired first-order accuracy is proved to be similar to that of STORM and other stochastic DFO algorithms, up to constants.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141775205","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Convergence Analysis of a Norm Minimization-Based Convex Vector Optimization Algorithm 基于规范最小化的凸向量优化算法的收敛性分析
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-07-25 DOI: 10.1137/23m1574580
Çağin Ararat, Firdevs Ulus, Muhammad Umer
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2700-2728, September 2024.
Abstract. In this work, we propose an outer approximation algorithm for solving bounded convex vector optimization problems (CVOPs). The scalarization model solved iteratively within the algorithm is a modification of the norm-minimizing scalarization proposed in [Ç. Ararat, F. Ulus, and M. Umer, J. Optim. Theory Appl., 194 (2022), pp. 681–712]. For a predetermined tolerance [math], we prove that the algorithm terminates after finitely many iterations, and it returns a polyhedral outer approximation to the upper image of the CVOP such that the Hausdorff distance between the two is less than [math]. We show that for an arbitrary norm used in the scalarization models, the approximation error after [math] iterations decreases by the order of [math], where [math] is the dimension of the objective space. An improved convergence rate of [math] is proved for the special case of using the Euclidean norm.
SIAM 优化期刊》,第 34 卷第 3 期,第 2700-2728 页,2024 年 9 月。 摘要在这项工作中,我们提出了一种求解有界凸向量优化问题(CVOPs)的外近似算法。算法中迭代求解的标量化模型是对[Ç. Ararat, F. Ulus, Ç...]中提出的规范最小化标量化的修正。Ararat, F. Ulus, and M. Umer, J. Optim.理论应用》,194 (2022),第 681-712 页]。对于预定公差 [math],我们证明该算法在有限次迭代后终止,并返回 CVOP 上像的多面体外近似,且两者之间的豪斯多夫距离小于 [math]。我们证明,对于标量化模型中使用的任意规范,[math] 次迭代后的近似误差会以 [math] 的数量级减小,其中 [math] 是目标空间的维度。对于使用欧氏规范的特殊情况,我们证明了[math]的收敛率有所提高。
{"title":"Convergence Analysis of a Norm Minimization-Based Convex Vector Optimization Algorithm","authors":"Çağin Ararat, Firdevs Ulus, Muhammad Umer","doi":"10.1137/23m1574580","DOIUrl":"https://doi.org/10.1137/23m1574580","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2700-2728, September 2024. <br/> Abstract. In this work, we propose an outer approximation algorithm for solving bounded convex vector optimization problems (CVOPs). The scalarization model solved iteratively within the algorithm is a modification of the norm-minimizing scalarization proposed in [Ç. Ararat, F. Ulus, and M. Umer, J. Optim. Theory Appl., 194 (2022), pp. 681–712]. For a predetermined tolerance [math], we prove that the algorithm terminates after finitely many iterations, and it returns a polyhedral outer approximation to the upper image of the CVOP such that the Hausdorff distance between the two is less than [math]. We show that for an arbitrary norm used in the scalarization models, the approximation error after [math] iterations decreases by the order of [math], where [math] is the dimension of the objective space. An improved convergence rate of [math] is proved for the special case of using the Euclidean norm.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863976","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Small Errors in Random Zeroth-Order Optimization Are Imaginary 随机零阶优化中的小误差是虚构的
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-07-19 DOI: 10.1137/22m1510261
Wouter Jongeneel, Man-Chung Yue, Daniel Kuhn
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2638-2670, September 2024.
Abstract. Most zeroth-order optimization algorithms mimic a first-order algorithm but replace the gradient of the objective function with some gradient estimator that can be computed from a small number of function evaluations. This estimator is constructed randomly, and its expectation matches the gradient of a smooth approximation of the objective function whose quality improves as the underlying smoothing parameter [math] is reduced. Gradient estimators requiring a smaller number of function evaluations are preferable from a computational point of view. While estimators based on a single function evaluation can be obtained by use of the divergence theorem from vector calculus, their variance explodes as [math] tends to 0. Estimators based on multiple function evaluations, on the other hand, suffer from numerical cancellation when [math] tends to 0. To combat both effects simultaneously, we extend the objective function to the complex domain and construct a gradient estimator that evaluates the objective at a complex point whose coordinates have small imaginary parts of the order [math]. As this estimator requires only one function evaluation, it is immune to cancellation. In addition, its variance remains bounded as [math] tends to 0. We prove that zeroth-order algorithms that use our estimator offer the same theoretical convergence guarantees as the state-of-the-art methods. Numerical experiments suggest, however, that they often converge faster in practice.
SIAM 优化期刊》,第 34 卷第 3 期,第 2638-2670 页,2024 年 9 月。 摘要大多数零阶优化算法模仿一阶算法,但用一些梯度估计器代替目标函数的梯度。这种估计器是随机构建的,其期望值与目标函数平滑近似值的梯度相匹配,而目标函数平滑近似值的质量会随着基本平滑参数[数学]的降低而提高。从计算角度来看,需要较少函数评估次数的梯度估计器更为可取。虽然可以利用向量微积分中的发散定理获得基于单次函数求值的估计值,但当[math]趋近于 0 时,估计值的方差会爆炸性增长。由于这种估计器只需要一次函数评估,因此不会被抵消。此外,当[math]趋于 0 时,它的方差仍然是有界的。我们证明,使用我们的估计器的零阶算法能提供与最先进方法相同的理论收敛保证。然而,数值实验表明,它们在实践中的收敛速度往往更快。
{"title":"Small Errors in Random Zeroth-Order Optimization Are Imaginary","authors":"Wouter Jongeneel, Man-Chung Yue, Daniel Kuhn","doi":"10.1137/22m1510261","DOIUrl":"https://doi.org/10.1137/22m1510261","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2638-2670, September 2024. <br/> Abstract. Most zeroth-order optimization algorithms mimic a first-order algorithm but replace the gradient of the objective function with some gradient estimator that can be computed from a small number of function evaluations. This estimator is constructed randomly, and its expectation matches the gradient of a smooth approximation of the objective function whose quality improves as the underlying smoothing parameter [math] is reduced. Gradient estimators requiring a smaller number of function evaluations are preferable from a computational point of view. While estimators based on a single function evaluation can be obtained by use of the divergence theorem from vector calculus, their variance explodes as [math] tends to 0. Estimators based on multiple function evaluations, on the other hand, suffer from numerical cancellation when [math] tends to 0. To combat both effects simultaneously, we extend the objective function to the complex domain and construct a gradient estimator that evaluates the objective at a complex point whose coordinates have small imaginary parts of the order [math]. As this estimator requires only one function evaluation, it is immune to cancellation. In addition, its variance remains bounded as [math] tends to 0. We prove that zeroth-order algorithms that use our estimator offer the same theoretical convergence guarantees as the state-of-the-art methods. Numerical experiments suggest, however, that they often converge faster in practice.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141738174","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Square Root LASSO: Well-Posedness, Lipschitz Stability, and the Tuning Trade-Off 平方根 LASSO:拟合性、Lipschitz 稳定性和调整权衡
IF 3.1 1区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-07-18 DOI: 10.1137/23m1561968
Aaron Berk, Simone Brugiapaglia, Tim Hoheisel
SIAM Journal on Optimization, Volume 34, Issue 3, Page 2609-2637, September 2024.
Abstract. This paper studies well-posedness and parameter sensitivity of the square root LASSO (SR-LASSO), an optimization model for recovering sparse solutions to linear inverse problems in finite dimension. An advantage of the SR-LASSO (e.g., over the standard LASSO) is that the optimal tuning of the regularization parameter is robust with respect to measurement noise. This paper provides three point-based regularity conditions at a solution of the SR-LASSO: the weak, intermediate, and strong assumptions. It is shown that the weak assumption implies uniqueness of the solution in question. The intermediate assumption yields a directionally differentiable and locally Lipschitz solution map (with explicit Lipschitz bounds), whereas the strong assumption gives continuous differentiability of said map around the point in question. Our analysis leads to new theoretical insights on the comparison between SR-LASSO and LASSO from the viewpoint of tuning parameter sensitivity: noise-robust optimal parameter choice for SR-LASSO comes at the “price” of elevated tuning parameter sensitivity. Numerical results support and showcase the theoretical findings.
SIAM 优化期刊》,第 34 卷第 3 期,第 2609-2637 页,2024 年 9 月。 摘要本文研究了平方根 LASSO(SR-LASSO)的问题解决性和参数敏感性,SR-LASSO 是一种用于恢复有限维线性逆问题稀疏解的优化模型。与标准 LASSO 相比,SR-LASSO 的优势在于正则化参数的优化调整对测量噪声具有鲁棒性。本文提供了 SR-LASSO 解的三个基于点的正则性条件:弱假设、中假设和强假设。结果表明,弱假设意味着相关解的唯一性。中间假设产生了方向可微分和局部 Lipschitz 解映射(具有明确的 Lipschitz 边界),而强假设则给出了所述映射在相关点周围的连续可微分性。我们的分析从调谐参数灵敏度的角度,对 SR-LASSO 和 LASSO 的比较提出了新的理论见解:SR-LASSO 的噪声最优参数选择是以提高调谐参数灵敏度为 "代价 "的。数值结果支持并展示了理论发现。
{"title":"Square Root LASSO: Well-Posedness, Lipschitz Stability, and the Tuning Trade-Off","authors":"Aaron Berk, Simone Brugiapaglia, Tim Hoheisel","doi":"10.1137/23m1561968","DOIUrl":"https://doi.org/10.1137/23m1561968","url":null,"abstract":"SIAM Journal on Optimization, Volume 34, Issue 3, Page 2609-2637, September 2024. <br/> Abstract. This paper studies well-posedness and parameter sensitivity of the square root LASSO (SR-LASSO), an optimization model for recovering sparse solutions to linear inverse problems in finite dimension. An advantage of the SR-LASSO (e.g., over the standard LASSO) is that the optimal tuning of the regularization parameter is robust with respect to measurement noise. This paper provides three point-based regularity conditions at a solution of the SR-LASSO: the weak, intermediate, and strong assumptions. It is shown that the weak assumption implies uniqueness of the solution in question. The intermediate assumption yields a directionally differentiable and locally Lipschitz solution map (with explicit Lipschitz bounds), whereas the strong assumption gives continuous differentiability of said map around the point in question. Our analysis leads to new theoretical insights on the comparison between SR-LASSO and LASSO from the viewpoint of tuning parameter sensitivity: noise-robust optimal parameter choice for SR-LASSO comes at the “price” of elevated tuning parameter sensitivity. Numerical results support and showcase the theoretical findings.","PeriodicalId":49529,"journal":{"name":"SIAM Journal on Optimization","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141745955","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
SIAM Journal on Optimization
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1