Pub Date : 2022-01-01DOI: 10.1016/j.ejco.2022.100035
Fred Roosta , Yang Liu , Peng Xu , Michael W. Mahoney
We consider a variant of inexact Newton Method [20], [40], called Newton-MR, in which the least-squares sub-problems are solved approximately using Minimum Residual method [79]. By construction, Newton-MR can be readily applied for unconstrained optimization of a class of non-convex problems known as invex, which subsumes convexity as a sub-class. For invex optimization, instead of the classical Lipschitz continuity assumptions on gradient and Hessian, Newton-MR's global convergence can be guaranteed under a weaker notion of joint regularity of Hessian and gradient. We also obtain Newton-MR's problem-independent local convergence to the set of minima. We show that fast local/global convergence can be guaranteed under a novel inexactness condition, which, to our knowledge, is much weaker than the prior related works. Numerical results demonstrate the performance of Newton-MR as compared with several other Newton-type alternatives on a few machine learning problems.
{"title":"Newton-MR: Inexact Newton Method with minimum residual sub-problem solver","authors":"Fred Roosta , Yang Liu , Peng Xu , Michael W. Mahoney","doi":"10.1016/j.ejco.2022.100035","DOIUrl":"10.1016/j.ejco.2022.100035","url":null,"abstract":"<div><p>We consider a variant of inexact Newton Method <span>[20]</span>, <span>[40]</span>, called Newton-MR, in which the least-squares sub-problems are solved approximately using Minimum Residual method <span>[79]</span>. By construction, Newton-MR can be readily applied for unconstrained optimization of a class of non-convex problems known as invex, which subsumes convexity as a sub-class. For invex optimization, instead of the classical Lipschitz continuity assumptions on gradient and Hessian, Newton-MR's global convergence can be guaranteed under a weaker notion of joint regularity of Hessian and gradient. We also obtain Newton-MR's problem-independent local convergence to the set of minima. We show that fast local/global convergence can be guaranteed under a novel inexactness condition, which, to our knowledge, is much weaker than the prior related works. Numerical results demonstrate the performance of Newton-MR as compared with several other Newton-type alternatives on a few machine learning problems.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"10 ","pages":"Article 100035"},"PeriodicalIF":2.4,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2192440622000119/pdfft?md5=d469cd05ef15c6b063a51fd431c7a8dd&pid=1-s2.0-S2192440622000119-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123761580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-01-01DOI: 10.1016/j.ejco.2022.100031
Thorsten Koch , Timo Berthold , Jaap Pedersen , Charlie Vanaret
This study investigates the progress made in lp and milp solver performance during the last two decades by comparing the solver software from the beginning of the millennium with the codes available today. On average, we found out that for solving lp/milp, computer hardware got about 20 times faster, and the algorithms improved by a factor of about nine for lp and around 50 for milp, which gives a total speed-up of about 180 and 1,000 times, respectively. However, these numbers have a very high variance and they considerably underestimate the progress made on the algorithmic side: many problem instances can nowadays be solved within seconds, which the old codes are not able to solve within any reasonable time.
{"title":"Progress in mathematical programming solvers from 2001 to 2020","authors":"Thorsten Koch , Timo Berthold , Jaap Pedersen , Charlie Vanaret","doi":"10.1016/j.ejco.2022.100031","DOIUrl":"10.1016/j.ejco.2022.100031","url":null,"abstract":"<div><p>This study investigates the progress made in <span>lp</span> and <span>milp</span> solver performance during the last two decades by comparing the solver software from the beginning of the millennium with the codes available today. On average, we found out that for solving <span>lp</span>/<span>milp</span>, computer hardware got about 20 times faster, and the algorithms improved by a factor of about nine for <span>lp</span> and around 50 for <span>milp</span>, which gives a total speed-up of about 180 and 1,000 times, respectively. However, these numbers have a very high variance and they considerably underestimate the progress made on the algorithmic side: many problem instances can nowadays be solved within seconds, which the old codes are not able to solve within any reasonable time.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"10 ","pages":"Article 100031"},"PeriodicalIF":2.4,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2192440622000077/pdfft?md5=79377e1d524040849993372a12f99ead&pid=1-s2.0-S2192440622000077-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131506187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-01-01DOI: 10.1016/j.ejco.2022.100049
Roberto Montemanni , Xiaochen Chou , Derek H. Smith
B-coloring is a problem in graph theory. It can model some real applications, as well as being used to enhance solution methods for the classical graph coloring problem. In turn, improved solutions for the classical coloring problem would impact a larger pool of practical applications in several different fields such as scheduling, timetabling and telecommunications. Given a graph , the b-coloring problem aims to maximize the number of colors used while assigning a color to every vertex in V, preventing adjacent vertices from receiving the same color, with every color represented by a special vertex, called a b-vertex. A vertex can be a b-vertex only if the set of colors assigned to its adjacent vertices includes all the colors, apart from the one assigned to the vertex itself.
This work employs methods based on Linear Programming to derive new upper and lower bounds for the problem. In particular, starting from a Mixed Integer Linear Programming model recently presented, upper bounds are obtained through partial linear relaxations of this model, while lower bounds are derived by considering different variations of the original model, modified to target a specific number of colors provided as input. The experimental campaign documented in the paper led to several improvements to the state-of-the-art results.
{"title":"Upper and lower bounds based on linear programming for the b-coloring problem","authors":"Roberto Montemanni , Xiaochen Chou , Derek H. Smith","doi":"10.1016/j.ejco.2022.100049","DOIUrl":"10.1016/j.ejco.2022.100049","url":null,"abstract":"<div><p>B-coloring is a problem in graph theory. It can model some real applications, as well as being used to enhance solution methods for the classical graph coloring problem. In turn, improved solutions for the classical coloring problem would impact a larger pool of practical applications in several different fields such as scheduling, timetabling and telecommunications. Given a graph <span><math><mi>G</mi><mo>=</mo><mo>(</mo><mi>V</mi><mo>,</mo><mi>E</mi><mo>)</mo></math></span>, the <em>b-coloring problem</em> aims to maximize the number of colors used while assigning a color to every vertex in <em>V</em>, preventing adjacent vertices from receiving the same color, with every color represented by a special vertex, called a b-vertex. A vertex can be a <em>b-vertex</em> only if the set of colors assigned to its adjacent vertices includes all the colors, apart from the one assigned to the vertex itself.</p><p>This work employs methods based on Linear Programming to derive new upper and lower bounds for the problem. In particular, starting from a Mixed Integer Linear Programming model recently presented, upper bounds are obtained through partial linear relaxations of this model, while lower bounds are derived by considering different variations of the original model, modified to target a specific number of colors provided as input. The experimental campaign documented in the paper led to several improvements to the state-of-the-art results.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"10 ","pages":"Article 100049"},"PeriodicalIF":2.4,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2192440622000259/pdfft?md5=4554fc69f0108b024eff3393ae695fc6&pid=1-s2.0-S2192440622000259-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128592840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-01-01DOI: 10.1016/j.ejco.2022.100044
Rémi Chan–Renous-Legoubin , Clément W. Royer
Nonlinear conjugate gradients are among the most popular techniques for solving continuous optimization problems. Although these schemes have long been studied from a global convergence standpoint, their worst-case complexity properties have yet to be fully understood, especially in the nonconvex setting. In particular, it is unclear whether nonlinear conjugate gradient methods possess better guarantees than first-order methods such as gradient descent. Meanwhile, recent experiments have shown impressive performance of standard nonlinear conjugate gradient techniques on certain nonconvex problems, even when compared with methods endowed with the best known complexity guarantees.
In this paper, we propose a nonlinear conjugate gradient scheme based on a simple line-search paradigm and a modified restart condition. These two ingredients allow for monitoring the properties of the search directions, which is instrumental in obtaining complexity guarantees. Our complexity results illustrate the possible discrepancy between nonlinear conjugate gradient methods and classical gradient descent. A numerical investigation on nonconvex robust regression problems as well as a standard benchmark illustrate that the restarting condition can track the behavior of a standard implementation.
{"title":"A nonlinear conjugate gradient method with complexity guarantees and its application to nonconvex regression","authors":"Rémi Chan–Renous-Legoubin , Clément W. Royer","doi":"10.1016/j.ejco.2022.100044","DOIUrl":"10.1016/j.ejco.2022.100044","url":null,"abstract":"<div><p>Nonlinear conjugate gradients are among the most popular techniques for solving continuous optimization problems. Although these schemes have long been studied from a global convergence standpoint, their worst-case complexity properties have yet to be fully understood, especially in the nonconvex setting. In particular, it is unclear whether nonlinear conjugate gradient methods possess better guarantees than first-order methods such as gradient descent. Meanwhile, recent experiments have shown impressive performance of standard nonlinear conjugate gradient techniques on certain nonconvex problems, even when compared with methods endowed with the best known complexity guarantees.</p><p>In this paper, we propose a nonlinear conjugate gradient scheme based on a simple line-search paradigm and a modified restart condition. These two ingredients allow for monitoring the properties of the search directions, which is instrumental in obtaining complexity guarantees. Our complexity results illustrate the possible discrepancy between nonlinear conjugate gradient methods and classical gradient descent. A numerical investigation on nonconvex robust regression problems as well as a standard benchmark illustrate that the restarting condition can track the behavior of a standard implementation.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"10 ","pages":"Article 100044"},"PeriodicalIF":2.4,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S219244062200020X/pdfft?md5=32a8c7d35ac8b53e431514d2573efa79&pid=1-s2.0-S219244062200020X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114696474","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-01-01DOI: 10.1016/j.ejco.2022.100048
Ekaterina Borodich , Vladislav Tominin , Yaroslav Tominin , Dmitry Kovalev , Alexander Gasnikov , Pavel Dvurechensky
We consider composite minimax optimization problems where the goal is to find a saddle-point of a large sum of non-bilinear objective functions augmented by simple composite regularizers for the primal and dual variables. For such problems, under the average-smoothness assumption, we propose accelerated stochastic variance-reduced algorithms with optimal up to logarithmic factors complexity bounds. In particular, we consider strongly-convex-strongly-concave, convex-strongly-concave, and convex-concave objectives. To the best of our knowledge, these are the first nearly-optimal algorithms for this setting.
{"title":"Accelerated variance-reduced methods for saddle-point problems","authors":"Ekaterina Borodich , Vladislav Tominin , Yaroslav Tominin , Dmitry Kovalev , Alexander Gasnikov , Pavel Dvurechensky","doi":"10.1016/j.ejco.2022.100048","DOIUrl":"10.1016/j.ejco.2022.100048","url":null,"abstract":"<div><p>We consider composite minimax optimization problems where the goal is to find a saddle-point of a large sum of non-bilinear objective functions augmented by simple composite regularizers for the primal and dual variables. For such problems, under the average-smoothness assumption, we propose accelerated stochastic variance-reduced algorithms with optimal up to logarithmic factors complexity bounds. In particular, we consider strongly-convex-strongly-concave, convex-strongly-concave, and convex-concave objectives. To the best of our knowledge, these are the first nearly-optimal algorithms for this setting.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"10 ","pages":"Article 100048"},"PeriodicalIF":2.4,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2192440622000247/pdfft?md5=41248ad222d5ad361783568adf860824&pid=1-s2.0-S2192440622000247-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116295129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-01-01DOI: 10.1016/j.ejco.2022.100033
Tibor Illés , Tamás Terlaky
This brief note presents a personal recollection of the early history of EUROpt, the Continuous Optimization Working Group of EURO. This historical note details the events that happened before the formation of EUROpt Working Group and the first five years of its existence. During the early years EUROpt Working Group established a conference series, organized thematic EURO Mini conferences, launched the EUROpt Fellow program, developed an effective rotating management structure, and grown to a large, matured, very active and high impact EURO Working Group.
{"title":"EUROpt, the Continuous Optimization Working Group of EURO: From idea to maturity","authors":"Tibor Illés , Tamás Terlaky","doi":"10.1016/j.ejco.2022.100033","DOIUrl":"10.1016/j.ejco.2022.100033","url":null,"abstract":"<div><p>This brief note presents a personal recollection of the early history of EUR<em>O</em>pt, the Continuous Optimization Working Group of EURO. This historical note details the events that happened before the formation of EUR<em>O</em>pt Working Group and the first five years of its existence. During the early years EUR<em>O</em>pt Working Group established a conference series, organized thematic EURO Mini conferences, launched the EUR<em>O</em>pt Fellow program, developed an effective rotating management structure, and grown to a large, matured, very active and high impact EURO Working Group.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"10 ","pages":"Article 100033"},"PeriodicalIF":2.4,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2192440622000090/pdfft?md5=a62c5ab91e77a43689d735471635b334&pid=1-s2.0-S2192440622000090-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115932034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-01-01DOI: 10.1016/j.ejco.2022.100034
Héctor G.-de-Alba , Samuel Nucamendi-Guillén , Oliver Avalos-Rosales
In this paper, the unrelated parallel machine scheduling problem with the objective of minimizing the total tardiness is addressed. For such a problem, a mixed-integer linear programming (MILP) formulation, that considers assignment and positional variables, is presented. In addition, an iterated local search (ILS) algorithm that produces high-quality solutions in reasonable times is proposed for large size instances. The ILS robustness was determined by comparing its performance with the results provided by the MILP. The instances used in this paper were constructed under a new approach which results in tighter due dates than the previous generation method for this problem. The proposed MILP formulation was able to solve instances of up to 150 jobs and 20 machines. Regarding the ILS, it yielded high-quality solutions in a reasonable time, solving instances of a size up to 400 jobs and 20 machines. Experimental results confirm that both approaches are efficient and promising.
{"title":"A mixed integer formulation and an efficient metaheuristic for the unrelated parallel machine scheduling problem: Total tardiness minimization","authors":"Héctor G.-de-Alba , Samuel Nucamendi-Guillén , Oliver Avalos-Rosales","doi":"10.1016/j.ejco.2022.100034","DOIUrl":"https://doi.org/10.1016/j.ejco.2022.100034","url":null,"abstract":"<div><p>In this paper, the unrelated parallel machine scheduling problem with the objective of minimizing the total tardiness is addressed. For such a problem, a mixed-integer linear programming (MILP) formulation, that considers assignment and positional variables, is presented. In addition, an iterated local search (ILS) algorithm that produces high-quality solutions in reasonable times is proposed for large size instances. The ILS robustness was determined by comparing its performance with the results provided by the MILP. The instances used in this paper were constructed under a new approach which results in tighter due dates than the previous generation method for this problem. The proposed MILP formulation was able to solve instances of up to 150 jobs and 20 machines. Regarding the ILS, it yielded high-quality solutions in a reasonable time, solving instances of a size up to 400 jobs and 20 machines. Experimental results confirm that both approaches are efficient and promising.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"10 ","pages":"Article 100034"},"PeriodicalIF":2.4,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2192440622000107/pdfft?md5=fe6b0c8e039b76ee7c40763ee43095a1&pid=1-s2.0-S2192440622000107-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"92090668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-01-01DOI: 10.1016/j.ejco.2022.100045
Pavel Dvurechensky , Dmitry Kamzolov , Aleksandr Lukashevich , Soomin Lee , Erik Ordentlich , César A. Uribe , Alexander Gasnikov
Statistical preconditioning enables fast methods for distributed large-scale empirical risk minimization problems. In this approach, multiple worker nodes compute gradients in parallel, which are then used by the central node to update the parameter by solving an auxiliary (preconditioned) smaller-scale optimization problem. The recently proposed Statistically Preconditioned Accelerated Gradient (SPAG) method [1] has complexity bounds superior to other such algorithms but requires an exact solution for computationally intensive auxiliary optimization problems at every iteration. In this paper, we propose an Inexact SPAG (InSPAG) and explicitly characterize the accuracy by which the corresponding auxiliary subproblem needs to be solved to guarantee the same convergence rate as the exact method. We build our results by first developing an inexact adaptive accelerated Bregman proximal gradient method for general optimization problems under relative smoothness and strong convexity assumptions, which may be of independent interest. Moreover, we explore the properties of the auxiliary problem in the InSPAG algorithm assuming Lipschitz third-order derivatives and strong convexity. For such problem class, we develop a linearly convergent Hyperfast second-order method and estimate the total complexity of the InSPAG method with hyperfast auxiliary problem solver. Finally, we illustrate the proposed method's practical efficiency by performing large-scale numerical experiments on logistic regression models. To the best of our knowledge, these are the first empirical results on implementing high-order methods on large-scale problems, as we work with data where the dimension is of the order of 3 million, and the number of samples is 700 million.
{"title":"Hyperfast second-order local solvers for efficient statistically preconditioned distributed optimization","authors":"Pavel Dvurechensky , Dmitry Kamzolov , Aleksandr Lukashevich , Soomin Lee , Erik Ordentlich , César A. Uribe , Alexander Gasnikov","doi":"10.1016/j.ejco.2022.100045","DOIUrl":"10.1016/j.ejco.2022.100045","url":null,"abstract":"<div><p>Statistical preconditioning enables fast methods for distributed large-scale empirical risk minimization problems. In this approach, multiple worker nodes compute gradients in parallel, which are then used by the central node to update the parameter by solving an auxiliary (preconditioned) smaller-scale optimization problem. The recently proposed Statistically Preconditioned Accelerated Gradient (SPAG) method <span>[1]</span> has complexity bounds superior to other such algorithms but requires an exact solution for computationally intensive auxiliary optimization problems at every iteration. In this paper, we propose an Inexact SPAG (InSPAG) and explicitly characterize the accuracy by which the corresponding auxiliary subproblem needs to be solved to guarantee the same convergence rate as the exact method. We build our results by first developing an inexact adaptive accelerated Bregman proximal gradient method for general optimization problems under relative smoothness and strong convexity assumptions, which may be of independent interest. Moreover, we explore the properties of the auxiliary problem in the InSPAG algorithm assuming Lipschitz third-order derivatives and strong convexity. For such problem class, we develop a linearly convergent Hyperfast second-order method and estimate the total complexity of the InSPAG method with hyperfast auxiliary problem solver. Finally, we illustrate the proposed method's practical efficiency by performing large-scale numerical experiments on logistic regression models. To the best of our knowledge, these are the first empirical results on implementing high-order methods on large-scale problems, as we work with data where the dimension is of the order of 3 million, and the number of samples is 700 million.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"10 ","pages":"Article 100045"},"PeriodicalIF":2.4,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2192440622000211/pdfft?md5=295cb611041330f3ffad8993cf73fef2&pid=1-s2.0-S2192440622000211-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121213587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-01-01DOI: 10.1016/j.ejco.2022.100043
S. Bellavia , G. Gurioli , B. Morini , Ph.L. Toint
A trust-region algorithm is presented for finding approximate minimizers of smooth unconstrained functions whose values and derivatives are subject to random noise. It is shown that, under suitable probabilistic assumptions, the new method finds (in expectation) an ϵ-approximate minimizer of arbitrary order in at most inexact evaluations of the function and its derivatives, providing the first such result for general optimality orders. The impact of intrinsic noise limiting the validity of the assumptions is also discussed and it is shown that difficulties are unlikely to occur in the first-order version of the algorithm for sufficiently large gradients. Conversely, should these assumptions fail for specific realizations, then “degraded” optimality guarantees are shown to hold when failure occurs. These conclusions are then discussed and illustrated in the context of subsampling methods for finite-sum optimization.
{"title":"Trust-region algorithms: Probabilistic complexity and intrinsic noise with applications to subsampling techniques","authors":"S. Bellavia , G. Gurioli , B. Morini , Ph.L. Toint","doi":"10.1016/j.ejco.2022.100043","DOIUrl":"10.1016/j.ejco.2022.100043","url":null,"abstract":"<div><p>A trust-region algorithm is presented for finding approximate minimizers of smooth unconstrained functions whose values and derivatives are subject to random noise. It is shown that, under suitable probabilistic assumptions, the new method finds (in expectation) an <em>ϵ</em>-approximate minimizer of arbitrary order <span><math><mi>q</mi><mo>≥</mo><mn>1</mn></math></span> in at most <span><math><mi>O</mi><mo>(</mo><msup><mrow><mi>ϵ</mi></mrow><mrow><mo>−</mo><mo>(</mo><mi>q</mi><mo>+</mo><mn>1</mn><mo>)</mo></mrow></msup><mo>)</mo></math></span> inexact evaluations of the function and its derivatives, providing the first such result for general optimality orders. The impact of intrinsic noise limiting the validity of the assumptions is also discussed and it is shown that difficulties are unlikely to occur in the first-order version of the algorithm for sufficiently large gradients. Conversely, should these assumptions fail for specific realizations, then “degraded” optimality guarantees are shown to hold when failure occurs. These conclusions are then discussed and illustrated in the context of subsampling methods for finite-sum optimization.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"10 ","pages":"Article 100043"},"PeriodicalIF":2.4,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2192440622000193/pdfft?md5=746d8300ed25b919398d91159dcb575f&pid=1-s2.0-S2192440622000193-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124064710","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-01-01DOI: 10.1016/j.ejco.2022.100037
Bo Peng
Recently and simultaneously, two MILP-based approaches to copositivity testing were proposed. This note tries a performance comparison, using a group of test sets containing a large number of designed instances. According to the numerical results, we find that one copositivity detection approach performs better when the function value of the defined function h of a matrix is large while the other one performs better when the dimension of problems is increasing moderately. A problem set that is hard for both approaches is also presented, which may be used as a test bed for future competing approaches. An improved variant of one of the approaches is also proposed to handle those hard instances more efficiently.
{"title":"Performance comparison of two recently proposed copositivity tests","authors":"Bo Peng","doi":"10.1016/j.ejco.2022.100037","DOIUrl":"10.1016/j.ejco.2022.100037","url":null,"abstract":"<div><p>Recently and simultaneously, two MILP-based approaches to copositivity testing were proposed. This note tries a performance comparison, using a group of test sets containing a large number of designed instances. According to the numerical results, we find that one copositivity detection approach performs better when the function value of the defined function <em>h</em> of a matrix is large while the other one performs better when the dimension of problems is increasing moderately. A problem set that is hard for both approaches is also presented, which may be used as a test bed for future competing approaches. An improved variant of one of the approaches is also proposed to handle those hard instances more efficiently.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"10 ","pages":"Article 100037"},"PeriodicalIF":2.4,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2192440622000132/pdfft?md5=abbd19fbc87e563c0963318349831747&pid=1-s2.0-S2192440622000132-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115890151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}