首页 > 最新文献

INFORMS journal on optimization最新文献

英文 中文
Comparison-Based Algorithms for One-Dimensional Stochastic Convex Optimization 基于比较的一维随机凸优化算法
Pub Date : 2018-09-09 DOI: 10.1287/ijoo.2019.0022
Xi Chen, Qihang Lin, Zizhuo Wang
Stochastic optimization finds a wide range of applications in operations research and management science. However, existing stochastic optimization techniques usually require the information of random samples (e.g., demands in the newsvendor problem) or the objective values at the sampled points (e.g., the lost sales cost), which might not be available in practice. In this paper, we consider a new setup for stochastic optimization, in which the decision maker has access to only comparative information between a random sample and two chosen decision points in each iteration. We propose a comparison-based algorithm (CBA) to solve such problems in one dimension with convex objective functions. Particularly, the CBA properly chooses the two points in each iteration and constructs an unbiased gradient estimate for the original problem. We show that the CBA achieves the same convergence rate as the optimal stochastic gradient methods (with the samples observed). We also consider extensions of our approach to multi-dimensional quadratic problems as well as problems with non-convex objective functions. Numerical experiments show that the CBA performs well in test problems.
随机优化在运筹学和管理学中有着广泛的应用。然而,现有的随机优化技术通常需要随机样本的信息(例如,报贩问题中的需求)或采样点的客观值(例如,损失的销售成本),这在实践中可能无法获得。在本文中,我们考虑了一种新的随机优化设置,其中决策者在每次迭代中只能访问随机样本和两个选定决策点之间的比较信息。我们提出了一种基于比较的算法(CBA)来解决一维凸目标函数问题。特别是,CBA算法在每次迭代中正确选择两个点,并对原问题构造无偏梯度估计。我们证明了CBA与最优随机梯度方法具有相同的收敛速度(与观察到的样本)。我们也考虑将我们的方法扩展到多维二次问题以及非凸目标函数问题。数值实验表明,CBA在测试问题中表现良好。
{"title":"Comparison-Based Algorithms for One-Dimensional Stochastic Convex Optimization","authors":"Xi Chen, Qihang Lin, Zizhuo Wang","doi":"10.1287/ijoo.2019.0022","DOIUrl":"https://doi.org/10.1287/ijoo.2019.0022","url":null,"abstract":"Stochastic optimization finds a wide range of applications in operations research and management science. However, existing stochastic optimization techniques usually require the information of random samples (e.g., demands in the newsvendor problem) or the objective values at the sampled points (e.g., the lost sales cost), which might not be available in practice. In this paper, we consider a new setup for stochastic optimization, in which the decision maker has access to only comparative information between a random sample and two chosen decision points in each iteration. We propose a comparison-based algorithm (CBA) to solve such problems in one dimension with convex objective functions. Particularly, the CBA properly chooses the two points in each iteration and constructs an unbiased gradient estimate for the original problem. We show that the CBA achieves the same convergence rate as the optimal stochastic gradient methods (with the samples observed). We also consider extensions of our approach to multi-dimensional quadratic problems as well as problems with non-convex objective functions. Numerical experiments show that the CBA performs well in test problems.","PeriodicalId":73382,"journal":{"name":"INFORMS journal on optimization","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1287/ijoo.2019.0022","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45063077","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An Ensemble Learning Framework for Model Fitting and Evaluation in Inverse Linear Optimization 逆线性优化中模型拟合与评价的集成学习框架
Pub Date : 2018-04-12 DOI: 10.1287/IJOO.2019.0045
A. Babier, T. Chan, Taewoo Lee, Rafid Mahmood, Daria Terekhov
We develop a generalized inverse optimization framework for fitting the cost vector of a single linear optimization problem given multiple observed decisions. This setting is motivated by ensemble learning, where building consensus from base learners can yield better predictions. We unify several models in the inverse optimization literature under a single framework and derive assumption-free and exact solution methods for each one. We extend a goodness-of-fit metric previously introduced for the problem with a single observed decision to this new setting and demonstrate several important properties. Finally, we demonstrate our framework in a novel inverse optimization-driven procedure for automated radiation therapy treatment planning. Here, the inverse optimization model leverages an ensemble of dose predictions from different machine learning models to construct a consensus treatment plan that outperforms baseline methods. The consensus plan yields better trade-offs between the competing clinical criteria used for plan evaluation.
我们开发了一个广义逆优化框架,用于拟合给定多个观测决策的单个线性优化问题的代价向量。这种设置是由集成学习驱动的,在集成学习中,从基础学习器中建立共识可以产生更好的预测。将逆优化文献中的几个模型统一在一个框架下,推导出每个模型的无假设精确解方法。我们将之前为具有单个观察决策的问题引入的拟合优度度量扩展到这个新设置,并演示了几个重要属性。最后,我们在一个新的逆优化驱动程序中展示了我们的框架,用于自动化放射治疗治疗计划。在这里,逆优化模型利用来自不同机器学习模型的剂量预测集合来构建优于基线方法的共识治疗计划。共识计划在用于计划评估的竞争性临床标准之间产生更好的权衡。
{"title":"An Ensemble Learning Framework for Model Fitting and Evaluation in Inverse Linear Optimization","authors":"A. Babier, T. Chan, Taewoo Lee, Rafid Mahmood, Daria Terekhov","doi":"10.1287/IJOO.2019.0045","DOIUrl":"https://doi.org/10.1287/IJOO.2019.0045","url":null,"abstract":"We develop a generalized inverse optimization framework for fitting the cost vector of a single linear optimization problem given multiple observed decisions. This setting is motivated by ensemble learning, where building consensus from base learners can yield better predictions. We unify several models in the inverse optimization literature under a single framework and derive assumption-free and exact solution methods for each one. We extend a goodness-of-fit metric previously introduced for the problem with a single observed decision to this new setting and demonstrate several important properties. Finally, we demonstrate our framework in a novel inverse optimization-driven procedure for automated radiation therapy treatment planning. Here, the inverse optimization model leverages an ensemble of dose predictions from different machine learning models to construct a consensus treatment plan that outperforms baseline methods. The consensus plan yields better trade-offs between the competing clinical criteria used for plan evaluation.","PeriodicalId":73382,"journal":{"name":"INFORMS journal on optimization","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-04-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43779014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Improved Linear Programs for Discrete Barycenters 离散Barycenter的改进线性规划
Pub Date : 2018-03-30 DOI: 10.1287/ijoo.2019.0020
S. Borgwardt, Stephan Patterson
Discrete barycenters are the optimal solutions to mass transport problems for a set of discrete measures. They arise in applications of operations research and statistics. The best known algorithms are based on linear programming, but these programs scale exponentially in the number of measures, making them prohibitive for practical purposes. In this paper, we improve on these algorithms. First, by using the optimality conditions to restrict the search space, we provide a better linear program that reduces the number of variables dramatically. Second, we recall a proof method from the literature, which lends itself to a linear program that has not been considered for computations. We exhibit that this second formulation is a viable, and arguably the go-to approach, for data in general position. Third, we then combine the two programs into a single hybrid model that retains the best properties of both formulations for partially structured data. We then study the models through both a theoretical analysis and computational experiments. We consider both the hardness of constructing the models and their actual solution. In doing so, we exhibit that each of the improved linear programs becomes the best, go-to approach for data of different underlying structure.
离散重心是一组离散测度的质量传输问题的最优解。它们出现在运筹学和统计学的应用中。最著名的算法是基于线性规划的,但这些程序的度量数量呈指数级增长,这使得它们在实际应用中无法使用。在本文中,我们对这些算法进行了改进。首先,通过使用最优性条件来限制搜索空间,我们提供了一个更好的线性规划,可以显著减少变量的数量。其次,我们回顾了文献中的一种证明方法,它适用于一个未被考虑用于计算的线性程序。我们证明,对于一般情况下的数据,第二个公式是可行的,可以说是可行的方法。第三,我们将这两个程序组合成一个单一的混合模型,该模型保留了部分结构化数据的两个公式的最佳特性。然后,我们通过理论分析和计算实验对模型进行了研究。我们同时考虑了构建模型的难度及其实际解决方案。在这样做的过程中,我们展示了每一个改进的线性程序都成为不同底层结构数据的最佳方法。
{"title":"Improved Linear Programs for Discrete Barycenters","authors":"S. Borgwardt, Stephan Patterson","doi":"10.1287/ijoo.2019.0020","DOIUrl":"https://doi.org/10.1287/ijoo.2019.0020","url":null,"abstract":"Discrete barycenters are the optimal solutions to mass transport problems for a set of discrete measures. They arise in applications of operations research and statistics. The best known algorithms are based on linear programming, but these programs scale exponentially in the number of measures, making them prohibitive for practical purposes. \u0000In this paper, we improve on these algorithms. First, by using the optimality conditions to restrict the search space, we provide a better linear program that reduces the number of variables dramatically. Second, we recall a proof method from the literature, which lends itself to a linear program that has not been considered for computations. We exhibit that this second formulation is a viable, and arguably the go-to approach, for data in general position. Third, we then combine the two programs into a single hybrid model that retains the best properties of both formulations for partially structured data. \u0000We then study the models through both a theoretical analysis and computational experiments. We consider both the hardness of constructing the models and their actual solution. In doing so, we exhibit that each of the improved linear programs becomes the best, go-to approach for data of different underlying structure.","PeriodicalId":73382,"journal":{"name":"INFORMS journal on optimization","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1287/ijoo.2019.0020","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45991523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Inexact Nonconvex Newton-Type Methods 非精确非凸牛顿型方法
Pub Date : 2018-02-20 DOI: 10.1287/IJOO.2019.0043
Z. Yao, Peng Xu, Farbod Roosta-Khorasani, Michael W. Mahoney
The paper aims to extend the theory and application of nonconvex Newton-type methods, namely trust region and cubic regularization, to the settings in which, in addition to the solution of subproblems, the gradient and the Hessian of the objective function are approximated. Using certain conditions on such approximations, the paper establishes optimal worst-case iteration complexities as the exact counterparts. This paper is part of a broader research program on designing, analyzing, and implementing efficient second-order optimization methods for large-scale machine learning applications. The authors were based at UC Berkeley when the idea of the project was conceived. The first two authors were PhD students, the third author was a postdoc, all supervised by the fourth author.
本文旨在将非凸牛顿型方法(即信赖域和三次正则化)的理论和应用扩展到除了子问题的解之外,还近似目标函数的梯度和Hessian的环境中。利用这种近似的某些条件,本文建立了最优最坏情况迭代复杂度作为精确对应。本文是大规模机器学习应用中设计、分析和实现高效二阶优化方法的更广泛研究计划的一部分。该项目的构思时,作者就在加州大学伯克利分校。前两位作者是博士生,第三位作者是博士后,均由第四位作者指导。
{"title":"Inexact Nonconvex Newton-Type Methods","authors":"Z. Yao, Peng Xu, Farbod Roosta-Khorasani, Michael W. Mahoney","doi":"10.1287/IJOO.2019.0043","DOIUrl":"https://doi.org/10.1287/IJOO.2019.0043","url":null,"abstract":"The paper aims to extend the theory and application of nonconvex Newton-type methods, namely trust region and cubic regularization, to the settings in which, in addition to the solution of subproblems, the gradient and the Hessian of the objective function are approximated. Using certain conditions on such approximations, the paper establishes optimal worst-case iteration complexities as the exact counterparts. This paper is part of a broader research program on designing, analyzing, and implementing efficient second-order optimization methods for large-scale machine learning applications. The authors were based at UC Berkeley when the idea of the project was conceived. The first two authors were PhD students, the third author was a postdoc, all supervised by the fourth author.","PeriodicalId":73382,"journal":{"name":"INFORMS journal on optimization","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-02-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45998619","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Stochastic Trust Region Algorithm Based on Careful Step Normalization 基于谨慎步进归一化的随机信任域算法
Pub Date : 2017-12-29 DOI: 10.1287/IJOO.2018.0010
Frank E. Curtis, K. Scheinberg, R. Shi
An algorithm is proposed for solving stochastic and finite-sum minimization problems. Based on a trust region methodology, the algorithm employs normalized steps, at least as long as the norms of t...
提出了一种求解随机和有限和最小化问题的算法。基于信任域方法,该算法采用归一化步骤,至少只要t的范数…
{"title":"A Stochastic Trust Region Algorithm Based on Careful Step Normalization","authors":"Frank E. Curtis, K. Scheinberg, R. Shi","doi":"10.1287/IJOO.2018.0010","DOIUrl":"https://doi.org/10.1287/IJOO.2018.0010","url":null,"abstract":"An algorithm is proposed for solving stochastic and finite-sum minimization problems. Based on a trust region methodology, the algorithm employs normalized steps, at least as long as the norms of t...","PeriodicalId":73382,"journal":{"name":"INFORMS journal on optimization","volume":"1 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1287/IJOO.2018.0010","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42046930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
“Relative Continuity” for Non-Lipschitz Nonsmooth Convex Optimization Using Stochastic (or Deterministic) Mirror Descent 基于随机(或确定性)镜像下降的非lipschitz非光滑凸优化的“相对连续性”
Pub Date : 2017-10-12 DOI: 10.1287/IJOO.2018.0008
Haihao Lu
The usual approach to developing and analyzing first-order methods for non-smooth (stochastic or deterministic) convex optimization assumes that the objective function is uniformly Lipschitz continuous with parameter $M_f$. However, in many settings the non-differentiable convex function $f(cdot)$ is not uniformly Lipschitz continuous -- for example (i) the classical support vector machine (SVM) problem, (ii) the problem of minimizing the maximum of convex quadratic functions, and even (iii) the univariate setting with $f(x) := max{0, x} + x^2$. Herein we develop a notion of "relative continuity" that is determined relative to a user-specified "reference function" $h(cdot)$ (that should be computationally tractable for algorithms), and we show that many non-differentiable convex functions are relatively continuous with respect to a correspondingly fairly-simple reference function $h(cdot)$. We also similarly develop a notion of "relative stochastic continuity" for the stochastic setting. We analysis two standard algorithms -- the (deterministic) mirror descent algorithm and the stochastic mirror descent algorithm -- for solving optimization problems in these two new settings, and we develop for the first time computational guarantees for instances where the objective function is not uniformly Lipschitz continuous. This paper is a companion paper for non-differentiable convex optimization to the recent paper by Lu, Freund, and Nesterov, which developed similar sorts of results for differentiable convex optimization.
开发和分析非光滑(随机或确定性)凸优化的一阶方法的常用方法假设目标函数与参数$M_f$一致Lipschitz连续。然而,在许多设置中,不可微凸函数$f(cdot)$不是一致Lipschitz连续的——例如(i)经典支持向量机(SVM)问题,(ii)凸二次函数的最大值最小化问题,甚至(iii)$f(x):=max{0,x}+x^2$的单变量设置。在此,我们发展了一个“相对连续性”的概念,该概念是相对于用户指定的“参考函数”$h(cdot)$(对于算法来说应该是可计算的)确定的,并且我们证明了许多不可微凸函数相对于相应的相当简单的参考函数$h(/cdot)$是相对连续的。我们还类似地为随机设置发展了“相对随机连续性”的概念。我们分析了两种标准算法——(确定性)镜像下降算法和随机镜像下降算法——用于解决这两种新设置下的优化问题,并首次为目标函数不是一致Lipschitz连续的情况开发了计算保证。这篇论文是Lu、Freund和Nesterov最近的一篇关于不可微凸优化的论文的配套论文,Lu、Flund和Neterov最近的论文发展了类似的关于可微凸最优化的结果。
{"title":"“Relative Continuity” for Non-Lipschitz Nonsmooth Convex Optimization Using Stochastic (or Deterministic) Mirror Descent","authors":"Haihao Lu","doi":"10.1287/IJOO.2018.0008","DOIUrl":"https://doi.org/10.1287/IJOO.2018.0008","url":null,"abstract":"The usual approach to developing and analyzing first-order methods for non-smooth (stochastic or deterministic) convex optimization assumes that the objective function is uniformly Lipschitz continuous with parameter $M_f$. However, in many settings the non-differentiable convex function $f(cdot)$ is not uniformly Lipschitz continuous -- for example (i) the classical support vector machine (SVM) problem, (ii) the problem of minimizing the maximum of convex quadratic functions, and even (iii) the univariate setting with $f(x) := max{0, x} + x^2$. Herein we develop a notion of \"relative continuity\" that is determined relative to a user-specified \"reference function\" $h(cdot)$ (that should be computationally tractable for algorithms), and we show that many non-differentiable convex functions are relatively continuous with respect to a correspondingly fairly-simple reference function $h(cdot)$. We also similarly develop a notion of \"relative stochastic continuity\" for the stochastic setting. We analysis two standard algorithms -- the (deterministic) mirror descent algorithm and the stochastic mirror descent algorithm -- for solving optimization problems in these two new settings, and we develop for the first time computational guarantees for instances where the objective function is not uniformly Lipschitz continuous. This paper is a companion paper for non-differentiable convex optimization to the recent paper by Lu, Freund, and Nesterov, which developed similar sorts of results for differentiable convex optimization.","PeriodicalId":73382,"journal":{"name":"INFORMS journal on optimization","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1287/IJOO.2018.0008","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46349114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 56
Separable Convex Optimization with Nested Lower and Upper Constraints 具有嵌套上下约束的可分离凸优化
Pub Date : 2017-03-04 DOI: 10.1287/IJOO.2018.0004
Thibaut Vidal, Daniel Gribel, Patrick Jaillet
We study a convex resource allocation problem in which lower and upper bounds are imposed on partial sums of allocations. This model is linked to a large range of applications, including production...
研究了一类凸资源分配问题,其中资源分配的部分和具有上界和下界。该模型与广泛的应用程序相关联,包括生产…
{"title":"Separable Convex Optimization with Nested Lower and Upper Constraints","authors":"Thibaut Vidal, Daniel Gribel, Patrick Jaillet","doi":"10.1287/IJOO.2018.0004","DOIUrl":"https://doi.org/10.1287/IJOO.2018.0004","url":null,"abstract":"We study a convex resource allocation problem in which lower and upper bounds are imposed on partial sums of allocations. This model is linked to a large range of applications, including production...","PeriodicalId":73382,"journal":{"name":"INFORMS journal on optimization","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1287/IJOO.2018.0004","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48959141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Convergence Rate Analysis of a Stochastic Trust-Region Method via Supermartingales 基于上鞅的随机信赖域方法的收敛速度分析
Pub Date : 2016-09-23 DOI: 10.1287/IJOO.2019.0016
J. Blanchet, C. Cartis, M. Menickelly, K. Scheinberg
We propose a novel framework for analyzing convergence rates of stochastic optimization algorithms with adaptive step sizes. This framework is based on analyzing properties of an underlying generic stochastic process, in particular by deriving a bound on the expected stopping time of this process. We utilize this framework to analyze the bounds on expected global convergence rates of a stochastic variant of a traditional trust region method, introduced in cite{ChenMenickellyScheinberg2014}. While traditional trust region methods rely on exact computations of the gradient, Hessian and values of the objective function, this method assumes that these values are available up to some dynamically adjusted accuracy. Moreover, this accuracy is assumed to hold only with some sufficiently large, but fixed, probability, without any additional restrictions on the variance of the errors. This setting applies, for example, to standard stochastic optimization and machine learning formulations. Improving upon the analysis in cite{ChenMenickellyScheinberg2014}, we show that the stochastic process defined by the algorithm satisfies the assumptions of our proposed general framework, with the stopping time defined as reaching accuracy $|nabla f(x)|leq epsilon$. The resulting bound for this stopping time is $O(epsilon^{-2})$, under the assumption of sufficiently accurate stochastic gradient, and is the first global complexity bound for a stochastic trust-region method. Finally, we apply the same framework to derive second order complexity bound under some additional assumptions.
我们提出了一个新的框架来分析具有自适应步长的随机优化算法的收敛速度。这个框架是基于分析一个潜在的一般随机过程的性质,特别是通过推导该过程的期望停止时间的界限。我们利用这个框架来分析在cite{ChenMenickellyScheinberg2014}中介绍的传统信赖域方法的随机变体的期望全局收敛率的界。传统的信赖域方法依赖于梯度、Hessian和目标函数值的精确计算,而该方法假设这些值在一定的动态调整精度范围内是可用的。此外,假定这种精度只有在某些足够大但固定的概率下才成立,对误差的方差没有任何额外的限制。例如,这个设置适用于标准的随机优化和机器学习公式。在cite{ChenMenickellyScheinberg2014}分析的基础上,我们证明了算法定义的随机过程满足我们提出的一般框架的假设,其停止时间定义为达到精度$|nabla f(x)|leq epsilon$。在足够精确的随机梯度假设下,该停止时间的结果界为$O(epsilon^{-2})$,是随机信赖域方法的第一个全局复杂度界。最后,我们应用相同的框架,在一些附加的假设下推导二阶复杂度界。
{"title":"Convergence Rate Analysis of a Stochastic Trust-Region Method via Supermartingales","authors":"J. Blanchet, C. Cartis, M. Menickelly, K. Scheinberg","doi":"10.1287/IJOO.2019.0016","DOIUrl":"https://doi.org/10.1287/IJOO.2019.0016","url":null,"abstract":"We propose a novel framework for analyzing convergence rates of stochastic optimization algorithms with adaptive step sizes. This framework is based on analyzing properties of an underlying generic stochastic process, in particular by deriving a bound on the expected stopping time of this process. We utilize this framework to analyze the bounds on expected global convergence rates of a stochastic variant of a traditional trust region method, introduced in cite{ChenMenickellyScheinberg2014}. While traditional trust region methods rely on exact computations of the gradient, Hessian and values of the objective function, this method assumes that these values are available up to some dynamically adjusted accuracy. Moreover, this accuracy is assumed to hold only with some sufficiently large, but fixed, probability, without any additional restrictions on the variance of the errors. This setting applies, for example, to standard stochastic optimization and machine learning formulations. Improving upon the analysis in cite{ChenMenickellyScheinberg2014}, we show that the stochastic process defined by the algorithm satisfies the assumptions of our proposed general framework, with the stopping time defined as reaching accuracy $|nabla f(x)|leq epsilon$. The resulting bound for this stopping time is $O(epsilon^{-2})$, under the assumption of sufficiently accurate stochastic gradient, and is the first global complexity bound for a stochastic trust-region method. Finally, we apply the same framework to derive second order complexity bound under some additional assumptions.","PeriodicalId":73382,"journal":{"name":"INFORMS journal on optimization","volume":"1 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2016-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1287/IJOO.2019.0016","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"66363408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 91
The Power and Limits of Predictive Approaches to Observational Data-Driven Optimization: The Case of Pricing 观察数据驱动优化的预测方法的力量和局限性:定价案例
Pub Date : 2016-05-08 DOI: 10.1287/ijoo.2022.0077
D. Bertsimas, Nathan Kallus
We consider data-driven decision making in which data on historical decisions and outcomes are endogenous and lack the necessary features for causal identification (e.g., unconfoundedness or instruments), focusing on data-driven pricing. We study approaches that, for lack of better alternative, optimize the prediction of objective (revenue) given decision (price). Whereas data-driven decision making is transforming modern operations, most large-scale data are observational, with which confounding is inevitable and the strong assumptions necessary for causal identification are dubious. Nonetheless, the inevitable statistical biases may be irrelevant if impact on downstream optimization performance is limited. This paper seeks to formalize and empirically study this question. First, to study the power of decision making with confounded data, by leveraging a special optimization structure, we develop bounds on the suboptimality of pricing using the (noncausal) prediction of historical demand given price. Second, to study the limits of decision making with confounded data, we develop a new hypothesis test for optimality with respect to the true average causal effect on the objective and apply it to interest rate–setting data to assesses whether performance can be distinguished from optimal to statistical significance in practice. Our empirical study demonstrates that predictive approaches can generally be powerful in practice with some limitations.
我们考虑数据驱动的决策,其中历史决策和结果的数据是内生的,缺乏因果识别的必要特征(例如,非混淆性或工具),专注于数据驱动的定价。由于缺乏更好的选择,我们研究了在给定决策(价格)的情况下优化目标(收入)预测的方法。虽然数据驱动的决策正在改变现代业务,但大多数大规模数据都是观测数据,与这些数据混淆是不可避免的,而因果识别所必需的强有力的假设是可疑的。尽管如此,如果对下游优化性能的影响有限,不可避免的统计偏差可能无关紧要。本文试图对这一问题进行形式化和实证研究。首先,为了研究混合数据的决策能力,我们利用一个特殊的优化结构,利用给定价格的历史需求(非因果)预测,开发了定价的次优性界限。其次,为了研究混杂数据决策的局限性,我们开发了一个新的假设检验,即关于对目标的真实平均因果效应的最优性,并将其应用于利率设定数据,以评估在实践中绩效是否可以区分为最优和统计显著性。我们的实证研究表明,预测方法在实践中通常是强大的,但也有一些局限性。
{"title":"The Power and Limits of Predictive Approaches to Observational Data-Driven Optimization: The Case of Pricing","authors":"D. Bertsimas, Nathan Kallus","doi":"10.1287/ijoo.2022.0077","DOIUrl":"https://doi.org/10.1287/ijoo.2022.0077","url":null,"abstract":"We consider data-driven decision making in which data on historical decisions and outcomes are endogenous and lack the necessary features for causal identification (e.g., unconfoundedness or instruments), focusing on data-driven pricing. We study approaches that, for lack of better alternative, optimize the prediction of objective (revenue) given decision (price). Whereas data-driven decision making is transforming modern operations, most large-scale data are observational, with which confounding is inevitable and the strong assumptions necessary for causal identification are dubious. Nonetheless, the inevitable statistical biases may be irrelevant if impact on downstream optimization performance is limited. This paper seeks to formalize and empirically study this question. First, to study the power of decision making with confounded data, by leveraging a special optimization structure, we develop bounds on the suboptimality of pricing using the (noncausal) prediction of historical demand given price. Second, to study the limits of decision making with confounded data, we develop a new hypothesis test for optimality with respect to the true average causal effect on the objective and apply it to interest rate–setting data to assesses whether performance can be distinguished from optimal to statistical significance in practice. Our empirical study demonstrates that predictive approaches can generally be powerful in practice with some limitations.","PeriodicalId":73382,"journal":{"name":"INFORMS journal on optimization","volume":"1 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2016-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"66363416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
期刊
INFORMS journal on optimization
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1