Pub Date : 2024-02-28DOI: 10.1007/s10898-024-01373-5
Marco Baioletti, Valentino Santucci, Marco Tomassini
During the last decades many metaheuristics for global numerical optimization have been proposed. Among them, Basin Hopping is very simple and straightforward to implement, although rarely used outside its original Physical Chemistry community. In this work, our aim is to compare Basin Hopping, and two population variants of it, with readily available implementations of the well known metaheuristics Differential Evolution, Particle Swarm Optimization, and Covariance Matrix Adaptation Evolution Strategy. We perform numerical experiments using the IOH profiler environment with the BBOB test function set and two difficult real-world problems. The experiments were carried out in two different but complementary ways: by measuring the performance under a fixed budget of function evaluations and by considering a fixed target value. The general conclusion is that Basin Hopping and its newly introduced population variant are almost as good as Covariance Matrix Adaptation on the synthetic benchmark functions and better than it on the two hard cluster energy minimization problems. Thus, the proposed analyses show that Basin Hopping can be considered a good candidate for global numerical optimization problems along with the more established metaheuristics, especially if one wants to obtain quick and reliable results on an unknown problem.
{"title":"A performance analysis of Basin hopping compared to established metaheuristics for global optimization","authors":"Marco Baioletti, Valentino Santucci, Marco Tomassini","doi":"10.1007/s10898-024-01373-5","DOIUrl":"https://doi.org/10.1007/s10898-024-01373-5","url":null,"abstract":"<p>During the last decades many metaheuristics for global numerical optimization have been proposed. Among them, Basin Hopping is very simple and straightforward to implement, although rarely used outside its original Physical Chemistry community. In this work, our aim is to compare Basin Hopping, and two population variants of it, with readily available implementations of the well known metaheuristics Differential Evolution, Particle Swarm Optimization, and Covariance Matrix Adaptation Evolution Strategy. We perform numerical experiments using the <i>IOH profiler</i> environment with the BBOB test function set and two difficult real-world problems. The experiments were carried out in two different but complementary ways: by measuring the performance under a fixed budget of function evaluations and by considering a fixed target value. The general conclusion is that Basin Hopping and its newly introduced population variant are almost as good as Covariance Matrix Adaptation on the synthetic benchmark functions and better than it on the two hard cluster energy minimization problems. Thus, the proposed analyses show that Basin Hopping can be considered a good candidate for global numerical optimization problems along with the more established metaheuristics, especially if one wants to obtain quick and reliable results on an unknown problem.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":null,"pages":null},"PeriodicalIF":1.8,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140005393","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-22DOI: 10.1007/s10898-024-01371-7
Cheng Lu, Wenguo Yang
We study the non-submodular maximization problem, in which the objective function is characterized by parameters, subject to a cardinality or (p)-system constraint. By adapting the Threshold-Greedy algorithm for the submodular maximization, we present two deterministic algorithms for approximately solving the non-submodular maximization problem. Our analysis shows that the algorithms we propose requires much less function evaluations than existing algorithms, while providing comparable approximation guarantees. Moreover, numerical experiment results are presented to validate the theoretical analysis. Our results not only fill a gap in the (non-)submodular maximization, but also generalize and improve several existing results on closely related optimization problems.
{"title":"Fast deterministic algorithms for non-submodular maximization with strong performance guarantees","authors":"Cheng Lu, Wenguo Yang","doi":"10.1007/s10898-024-01371-7","DOIUrl":"https://doi.org/10.1007/s10898-024-01371-7","url":null,"abstract":"<p>We study the non-submodular maximization problem, in which the objective function is characterized by parameters, subject to a cardinality or <span>(p)</span>-system constraint. By adapting the <span>Threshold-Greedy</span> algorithm for the submodular maximization, we present two deterministic algorithms for approximately solving the non-submodular maximization problem. Our analysis shows that the algorithms we propose requires much less function evaluations than existing algorithms, while providing comparable approximation guarantees. Moreover, numerical experiment results are presented to validate the theoretical analysis. Our results not only fill a gap in the (non-)submodular maximization, but also generalize and improve several existing results on closely related optimization problems.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":null,"pages":null},"PeriodicalIF":1.8,"publicationDate":"2024-02-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139924361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-22DOI: 10.1007/s10898-023-01360-2
Abstract
This paper explores a class of nonlinear Adjustable Robust Optimization (ARO) problems, containing here-and-now and wait-and-see variables, with uncertainty in the objective function and constraints. By applying Fenchel’s duality on the wait-and-see variables, we obtain an equivalent dual reformulation, which is a nonlinear static robust optimization problem. Using the dual formulation, we provide conditions under which the ARO problem is convex on the here-and-now decision. Furthermore, since the dual formulation contains a non-concave maximization on the uncertain parameter, we use perspective relaxation and an alternating method to handle the non-concavity. By employing the perspective relaxation, we obtain an upper bound, which we show is the same as the static relaxation of the considered problem. Moreover, invoking the alternating method, we design a new dual-based cutting plane algorithm that is able to find a reasonable lower bound for the optimal objective value of the considered nonlinear ARO model. In addition to sketching and establishing the theoretical features of the algorithms, including convergence analysis, by numerical experiments we reveal the abilities of our cutting plane algorithm in producing locally robust solutions with an acceptable optimality gap.
摘要 本文探讨了一类非线性可调稳健优化(ARO)问题,该问题包含此时此地和等待观察变量,目标函数和约束条件具有不确定性。通过对 "等待-观察 "变量应用 Fenchel 对偶,我们得到了一个等价的对偶重述,即一个非线性静态鲁棒优化问题。利用对偶表述,我们提供了 ARO 问题在此时此地的决策上具有凸性的条件。此外,由于对偶表述包含对不确定参数的非凹性最大化,我们使用透视松弛和交替法来处理非凹性。通过使用透视松弛法,我们得到了一个上界,并证明它与所考虑问题的静态松弛法相同。此外,利用交替法,我们设计了一种新的基于对偶的切割面算法,能够为所考虑的非线性 ARO 模型的最优目标值找到一个合理的下界。除了勾勒和建立算法的理论特征(包括收敛性分析)外,我们还通过数值实验揭示了我们的切割面算法在产生具有可接受最优性差距的局部稳健解方面的能力。
{"title":"A new dual-based cutting plane algorithm for nonlinear adjustable robust optimization","authors":"","doi":"10.1007/s10898-023-01360-2","DOIUrl":"https://doi.org/10.1007/s10898-023-01360-2","url":null,"abstract":"<h3>Abstract</h3> <p>This paper explores a class of nonlinear Adjustable Robust Optimization (ARO) problems, containing here-and-now and wait-and-see variables, with uncertainty in the objective function and constraints. By applying Fenchel’s duality on the wait-and-see variables, we obtain an equivalent dual reformulation, which is a nonlinear static robust optimization problem. Using the dual formulation, we provide conditions under which the ARO problem is convex on the here-and-now decision. Furthermore, since the dual formulation contains a non-concave maximization on the uncertain parameter, we use perspective relaxation and an alternating method to handle the non-concavity. By employing the perspective relaxation, we obtain an upper bound, which we show is the same as the static relaxation of the considered problem. Moreover, invoking the alternating method, we design a new dual-based cutting plane algorithm that is able to find a reasonable lower bound for the optimal objective value of the considered nonlinear ARO model. In addition to sketching and establishing the theoretical features of the algorithms, including convergence analysis, by numerical experiments we reveal the abilities of our cutting plane algorithm in producing locally robust solutions with an acceptable optimality gap.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":null,"pages":null},"PeriodicalIF":1.8,"publicationDate":"2024-02-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139924360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-15DOI: 10.1007/s10898-023-01358-w
Hongwei Jiao, Binbin Li, Wenqiang Yang
In this paper, we investigate a generalized multiplicative problem (GMP) that is known to be NP-hard even with one linear product term. We first introduce some criterion-space variables to obtain an equivalent problem of the GMP. A criterion-space branch-reduction-bound algorithm is then designed, which integrates some basic operations such as the two-level linear relaxation technique, rectangle branching rule and criterion-space region reduction technologies. The global convergence of the presented algorithm is proved by means of the subsequent solutions of a series of linear relaxation problems, and its maximum number of iterations is estimated on the basis of exhaustiveness of branching rule. Finally, numerical results demonstrate the presented algorithm can efficiently find the global optimum solutions for some test instances with the robustness.
{"title":"A criterion-space branch-reduction-bound algorithm for solving generalized multiplicative problems","authors":"Hongwei Jiao, Binbin Li, Wenqiang Yang","doi":"10.1007/s10898-023-01358-w","DOIUrl":"https://doi.org/10.1007/s10898-023-01358-w","url":null,"abstract":"<p>In this paper, we investigate a generalized multiplicative problem (GMP) that is known to be NP-hard even with one linear product term. We first introduce some criterion-space variables to obtain an equivalent problem of the GMP. A criterion-space branch-reduction-bound algorithm is then designed, which integrates some basic operations such as the two-level linear relaxation technique, rectangle branching rule and criterion-space region reduction technologies. The global convergence of the presented algorithm is proved by means of the subsequent solutions of a series of linear relaxation problems, and its maximum number of iterations is estimated on the basis of exhaustiveness of branching rule. Finally, numerical results demonstrate the presented algorithm can efficiently find the global optimum solutions for some test instances with the robustness.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":null,"pages":null},"PeriodicalIF":1.8,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139759405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-15DOI: 10.1007/s10898-024-01364-6
Sabah Bushaj, İ. Esra Büyüktahtakın
In this paper, we address the difficulty of solving large-scale multi-dimensional knapsack instances (MKP), presenting a novel deep reinforcement learning (DRL) framework. In this DRL framework, we train different agents compatible with a discrete action space for sequential decision-making while still satisfying any resource constraint of the MKP. This novel framework incorporates the decision variable values in the 2D DRL where the agent is responsible for assigning a value of 1 or 0 to each of the variables. To the best of our knowledge, this is the first DRL model of its kind in which a 2D environment is formulated, and an element of the DRL solution matrix represents an item of the MKP. Our framework is configured to solve MKP instances of different dimensions and distributions. We propose a K-means approach to obtain an initial feasible solution that is used to train the DRL agent. We train four different agents in our framework and present the results comparing each of them with the CPLEX commercial solver. The results show that our agents can learn and generalize over instances with different sizes and distributions. Our DRL framework shows that it can solve medium-sized instances at least 45 times faster in CPU solution time and at least 10 times faster for large instances, with a maximum solution gap of 0.28% compared to the performance of CPLEX. Furthermore, at least 95% of the items are predicted in line with the CPLEX solution. Computations with DRL also provide a better optimality gap with respect to state-of-the-art approaches.
{"title":"A K-means Supported Reinforcement Learning Framework to Multi-dimensional Knapsack","authors":"Sabah Bushaj, İ. Esra Büyüktahtakın","doi":"10.1007/s10898-024-01364-6","DOIUrl":"https://doi.org/10.1007/s10898-024-01364-6","url":null,"abstract":"<p>In this paper, we address the difficulty of solving large-scale multi-dimensional knapsack instances (MKP), presenting a novel deep reinforcement learning (DRL) framework. In this DRL framework, we train different agents compatible with a discrete action space for sequential decision-making while still satisfying any resource constraint of the MKP. This novel framework incorporates the decision variable values in the 2D DRL where the agent is responsible for assigning a value of 1 or 0 to each of the variables. To the best of our knowledge, this is the first DRL model of its kind in which a 2D environment is formulated, and an element of the DRL solution matrix represents an item of the MKP. Our framework is configured to solve MKP instances of different dimensions and distributions. We propose a K-means approach to obtain an initial feasible solution that is used to train the DRL agent. We train four different agents in our framework and present the results comparing each of them with the CPLEX commercial solver. The results show that our agents can learn and generalize over instances with different sizes and distributions. Our DRL framework shows that it can solve medium-sized instances at least 45 times faster in CPU solution time and at least 10 times faster for large instances, with a maximum solution gap of 0.28% compared to the performance of CPLEX. Furthermore, at least 95% of the items are predicted in line with the CPLEX solution. Computations with DRL also provide a better optimality gap with respect to state-of-the-art approaches.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":null,"pages":null},"PeriodicalIF":1.8,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139759470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-15DOI: 10.1007/s10898-024-01365-5
Zehao Xiao, Liwei Zhang
An online majorized semi-proximal alternating direction method of multiplier (Online-mspADMM) is proposed for a broad class of online linearly constrained composite optimization problems. A majorized technique is adopted to produce subproblems which can be easily solved. Under mild assumptions, we establish (mathcal {O}(sqrt{N})) objective regret and (mathcal {O}(sqrt{N})) constraint violation regret at round N. We apply the Online-mspADMM to solve different types of online regularized logistic regression problems. The numerical results on synthetic data sets verify the theoretical result about regrets.
针对各类在线线性约束复合优化问题,提出了一种在线大化半近似交替方向乘法(Online-mspADMM)。该方法采用大化技术来生成易于求解的子问题。在温和的假设条件下,我们在第 N 轮建立了 (mathcal {O}(sqrt{N})) 目标遗憾和 (mathcal {O}(sqrt{N})) 约束违反遗憾。在合成数据集上的数值结果验证了关于遗憾的理论结果。
{"title":"Regret analysis of an online majorized semi-proximal ADMM for online composite optimization","authors":"Zehao Xiao, Liwei Zhang","doi":"10.1007/s10898-024-01365-5","DOIUrl":"https://doi.org/10.1007/s10898-024-01365-5","url":null,"abstract":"<p>An online majorized semi-proximal alternating direction method of multiplier (Online-mspADMM) is proposed for a broad class of online linearly constrained composite optimization problems. A majorized technique is adopted to produce subproblems which can be easily solved. Under mild assumptions, we establish <span>(mathcal {O}(sqrt{N}))</span> objective regret and <span>(mathcal {O}(sqrt{N}))</span> constraint violation regret at round <i>N</i>. We apply the Online-mspADMM to solve different types of online regularized logistic regression problems. The numerical results on synthetic data sets verify the theoretical result about regrets.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":null,"pages":null},"PeriodicalIF":1.8,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139759413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-15DOI: 10.1007/s10898-024-01369-1
Kathrin Klamroth, Michael Stiglmayr, Claudia Totzeck
We propose a multi-swarm approach to approximate the Pareto front of general multi-objective optimization problems that is based on the consensus-based optimization method (CBO). The algorithm is motivated step by step beginning with a simple extension of CBO based on fixed scalarization weights. To overcome the issue of choosing the weights we propose an adaptive weight strategy in the second modeling step. The modeling process is concluded with the incorporation of a penalty strategy that avoids clusters along the Pareto front and a diffusion term that prevents collapsing swarms. Altogether the proposed K-swarm CBO algorithm is tailored for a diverse approximation of the Pareto front and, simultaneously, the efficient set of general non-convex multi-objective problems. The feasibility of the approach is justified by analytic results, including convergence proofs, and a performance comparison to the well-known non-dominated sorting genetic algorithms NSGA2 and NSGA3 as well as the recently proposed one-swarm approach for multi-objective problems involving consensus-based optimization.
{"title":"Consensus-based optimization for multi-objective problems: a multi-swarm approach","authors":"Kathrin Klamroth, Michael Stiglmayr, Claudia Totzeck","doi":"10.1007/s10898-024-01369-1","DOIUrl":"https://doi.org/10.1007/s10898-024-01369-1","url":null,"abstract":"<p>We propose a multi-swarm approach to approximate the Pareto front of general multi-objective optimization problems that is based on the consensus-based optimization method (CBO). The algorithm is motivated step by step beginning with a simple extension of CBO based on fixed scalarization weights. To overcome the issue of choosing the weights we propose an adaptive weight strategy in the second modeling step. The modeling process is concluded with the incorporation of a penalty strategy that avoids clusters along the Pareto front and a diffusion term that prevents collapsing swarms. Altogether the proposed <i>K</i>-swarm CBO algorithm is tailored for a diverse approximation of the Pareto front and, simultaneously, the efficient set of general non-convex multi-objective problems. The feasibility of the approach is justified by analytic results, including convergence proofs, and a performance comparison to the well-known non-dominated sorting genetic algorithms NSGA2 and NSGA3 as well as the recently proposed one-swarm approach for multi-objective problems involving consensus-based optimization.\u0000</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":null,"pages":null},"PeriodicalIF":1.8,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139759336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-13DOI: 10.1007/s10898-024-01372-6
Abstract
The problem of finding a globally optimal k-partition of a set (mathcal {A}) is a very intricate optimization problem for which in general, except in the case of one-dimensional data, i.e., for data with one feature ((mathcal {A}subset mathbb {R})), there is no method to solve. Only in the one-dimensional case, there are efficient methods based on the fact that the search for a globally optimal k-partition is equivalent to solving a global optimization problem for a symmetric Lipschitz-continuous function using the global optimization algorithm DIRECT. In the present paper, we propose a method for finding a globally optimal k-partition in the general case ((mathcal {A}subset mathbb {R}^n), (nge 1)), generalizing an idea for solving the Lipschitz global optimization for symmetric functions. To do this, we propose a method that combines a global optimization algorithm with linear constraints and the k-means algorithm. The first of these two algorithms is used only to find a good initial approximation for the k-means algorithm. The method was tested on a number of artificial datasets and on several examples from the UCI Machine Learning Repository, and an application in spectral clustering for linearly non-separable datasets is also demonstrated. Our proposed method proved to be very efficient.
{"title":"A method for searching for a globally optimal k-partition of higher-dimensional datasets","authors":"","doi":"10.1007/s10898-024-01372-6","DOIUrl":"https://doi.org/10.1007/s10898-024-01372-6","url":null,"abstract":"<h3>Abstract</h3> <p>The problem of finding a globally optimal <em>k</em>-partition of a set <span> <span>(mathcal {A})</span> </span> is a very intricate optimization problem for which in general, except in the case of one-dimensional data, i.e., for data with one feature (<span> <span>(mathcal {A}subset mathbb {R})</span> </span>), there is no method to solve. Only in the one-dimensional case, there are efficient methods based on the fact that the search for a globally optimal <em>k</em>-partition is equivalent to solving a global optimization problem for a symmetric Lipschitz-continuous function using the global optimization algorithm <span>DIRECT</span>. In the present paper, we propose a method for finding a globally optimal <em>k</em>-partition in the general case (<span> <span>(mathcal {A}subset mathbb {R}^n)</span> </span>, <span> <span>(nge 1)</span> </span>), generalizing an idea for solving the Lipschitz global optimization for symmetric functions. To do this, we propose a method that combines a global optimization algorithm with linear constraints and the <em>k</em>-means algorithm. The first of these two algorithms is used only to find a good initial approximation for the <em>k</em>-means algorithm. The method was tested on a number of artificial datasets and on several examples from the UCI Machine Learning Repository, and an application in spectral clustering for linearly non-separable datasets is also demonstrated. Our proposed method proved to be very efficient. </p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":null,"pages":null},"PeriodicalIF":1.8,"publicationDate":"2024-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139759418","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Interval branch-and-bound solvers provide reliable algorithms for handling non-convex optimization problems by ensuring the feasibility and the optimality of the computed solutions, i.e. independently from the floating-point rounding errors. Moreover, these solvers deal with a wide variety of mathematical operators. However, these solvers are not dedicated to quadratic optimization and do not exploit nonlinear convex relaxations in their framework. We present an interval branch-and-bound method that can efficiently solve quadratic optimization problems. At each node explored by the algorithm, our solver uses a quadratic convex relaxation which is as strong as a semi-definite programming relaxation, and a variable selection strategy dedicated to quadratic problems. The interval features can then propagate efficiently this information for contracting all variable domains. We also propose to make our algorithm rigorous by certifying firstly the convexity of the objective function of our relaxation, and secondly the validity of the lower bound calculated at each node. In the non-rigorous case, our experiments show significant speedups on general integer quadratic instances, and when reliability is required, our first results show that we are able to handle medium-sized instances in a reasonable running time.
{"title":"Global solution of quadratic problems using interval methods and convex relaxations","authors":"Sourour Elloumi, Amélie Lambert, Bertrand Neveu, Gilles Trombettoni","doi":"10.1007/s10898-024-01370-8","DOIUrl":"https://doi.org/10.1007/s10898-024-01370-8","url":null,"abstract":"<p>Interval branch-and-bound solvers provide reliable algorithms for handling non-convex optimization problems by ensuring the feasibility and the optimality of the computed solutions, i.e. independently from the floating-point rounding errors. Moreover, these solvers deal with a wide variety of mathematical operators. However, these solvers are not dedicated to quadratic optimization and do not exploit nonlinear convex relaxations in their framework. We present an interval branch-and-bound method that can efficiently solve quadratic optimization problems. At each node explored by the algorithm, our solver uses a quadratic convex relaxation which is as strong as a semi-definite programming relaxation, and a variable selection strategy dedicated to quadratic problems. The interval features can then propagate efficiently this information for contracting all variable domains. We also propose to make our algorithm rigorous by certifying firstly the convexity of the objective function of our relaxation, and secondly the validity of the lower bound calculated at each node. In the non-rigorous case, our experiments show significant speedups on general integer quadratic instances, and when reliability is required, our first results show that we are able to handle medium-sized instances in a reasonable running time.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":null,"pages":null},"PeriodicalIF":1.8,"publicationDate":"2024-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139759499","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper is devoted to the study of a class of multiobjective semi-infinite programming problems on Hadamard manifolds (in short, (MOSIP-HM)). We derive some alternative theorems analogous to Tucker’s theorem, Tucker’s first and second existence theorem, and Motzkin’s theorem of alternative in the framework of Hadamard manifolds. We employ Motzkin’s theorem of alternative to establish necessary and sufficient conditions that characterize KKT pseudoconvex functions using strong KKT vector critical points and efficient solutions of (MOSIP-HM). Moreover, we formulate the Mond-Weir and Wolfe-type dual problems related to (MOSIP-HM) and derive the weak and converse duality theorems relating (MOSIP-HM) and the dual problems. Several non-trivial numerical examples are provided to illustrate the significance of the derived results. The results deduced in the paper extend and generalize several notable works existing in the literature.
{"title":"Efficiency conditions and duality for multiobjective semi-infinite programming problems on Hadamard manifolds","authors":"Balendu Bhooshan Upadhyay, Arnav Ghosh, Savin Treanţă","doi":"10.1007/s10898-024-01367-3","DOIUrl":"https://doi.org/10.1007/s10898-024-01367-3","url":null,"abstract":"<p>This paper is devoted to the study of a class of multiobjective semi-infinite programming problems on Hadamard manifolds (in short, (MOSIP-HM)). We derive some alternative theorems analogous to Tucker’s theorem, Tucker’s first and second existence theorem, and Motzkin’s theorem of alternative in the framework of Hadamard manifolds. We employ Motzkin’s theorem of alternative to establish necessary and sufficient conditions that characterize KKT pseudoconvex functions using strong KKT vector critical points and efficient solutions of (MOSIP-HM). Moreover, we formulate the Mond-Weir and Wolfe-type dual problems related to (MOSIP-HM) and derive the weak and converse duality theorems relating (MOSIP-HM) and the dual problems. Several non-trivial numerical examples are provided to illustrate the significance of the derived results. The results deduced in the paper extend and generalize several notable works existing in the literature.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":null,"pages":null},"PeriodicalIF":1.8,"publicationDate":"2024-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139656331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}