Pub Date : 2024-03-19DOI: 10.1007/s10898-024-01382-4
Le Thi Khanh Hien, Dimitri Papadimitriou
In this paper, we propose an inertial alternating direction method of multipliers for solving a class of non-convex multi-block optimization problems with nonlinear coupling constraints. Distinctive features of our proposed method, when compared with other alternating direction methods of multipliers for solving non-convex problems with nonlinear coupling constraints, include: (i) we apply the inertial technique to the update of primal variables and (ii) we apply a non-standard update rule for the multiplier by scaling the multiplier by a factor before moving along the ascent direction where a relaxation parameter is allowed. Subsequential convergence and global convergence are presented for the proposed algorithm.
{"title":"An inertial ADMM for a class of nonconvex composite optimization with nonlinear coupling constraints","authors":"Le Thi Khanh Hien, Dimitri Papadimitriou","doi":"10.1007/s10898-024-01382-4","DOIUrl":"https://doi.org/10.1007/s10898-024-01382-4","url":null,"abstract":"<p>In this paper, we propose an inertial alternating direction method of multipliers for solving a class of non-convex multi-block optimization problems with <i>nonlinear coupling constraints</i>. Distinctive features of our proposed method, when compared with other alternating direction methods of multipliers for solving non-convex problems with nonlinear coupling constraints, include: (i) we apply the inertial technique to the update of primal variables and (ii) we apply a non-standard update rule for the multiplier by scaling the multiplier by a factor before moving along the ascent direction where a relaxation parameter is allowed. Subsequential convergence and global convergence are presented for the proposed algorithm.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"34 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140168165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-18DOI: 10.1007/s10898-024-01378-0
Zhou Sheng, Gonglin Yuan
Trust-region methods have received massive attention in a variety of continuous optimization. They aim to obtain a trial step by minimizing a quadratic model in a region of a certain trust-region radius around the current iterate. This paper proposes an adaptive Riemannian trust-region algorithm for optimization on manifolds, in which the trust-region radius depends linearly on the norm of the Riemannian gradient at each iteration. Under mild assumptions, we establish the liminf-type convergence, lim-type convergence, and global convergence results of the proposed algorithm. In addition, the proposed algorithm is shown to reach the conclusion that the norm of the Riemannian gradient is smaller than (epsilon ) within ({mathcal {O}}(frac{1}{epsilon ^2})) iterations. Some numerical examples of tensor approximations are carried out to reveal the performances of the proposed algorithm compared to the classical Riemannian trust-region algorithm.
{"title":"Convergence and worst-case complexity of adaptive Riemannian trust-region methods for optimization on manifolds","authors":"Zhou Sheng, Gonglin Yuan","doi":"10.1007/s10898-024-01378-0","DOIUrl":"https://doi.org/10.1007/s10898-024-01378-0","url":null,"abstract":"<p>Trust-region methods have received massive attention in a variety of continuous optimization. They aim to obtain a trial step by minimizing a quadratic model in a region of a certain trust-region radius around the current iterate. This paper proposes an adaptive Riemannian trust-region algorithm for optimization on manifolds, in which the trust-region radius depends linearly on the norm of the Riemannian gradient at each iteration. Under mild assumptions, we establish the liminf-type convergence, lim-type convergence, and global convergence results of the proposed algorithm. In addition, the proposed algorithm is shown to reach the conclusion that the norm of the Riemannian gradient is smaller than <span>(epsilon )</span> within <span>({mathcal {O}}(frac{1}{epsilon ^2}))</span> iterations. Some numerical examples of tensor approximations are carried out to reveal the performances of the proposed algorithm compared to the classical Riemannian trust-region algorithm.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"27 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140154986","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-14DOI: 10.1007/s10898-024-01377-1
Yonghong Yao, Abubakar Adamu, Yekini Shehu, Jen-Chih Yao
This paper proposes two simple and elegant proximal-type algorithms to solve equilibrium problems with pseudo-monotone bifunctions in the setting of Hilbert spaces. The proposed algorithms use one proximal point evaluation of the bifunction at each iteration. Consequently, prove that the sequences of iterates generated by the first algorithm converge weakly to a solution of the equilibrium problem (assuming existence) and obtain a linear convergence rate under standard assumptions. We also design a viscosity version of the first algorithm and obtain its corresponding strong convergence result. Some popular existing algorithms in the literature are recovered. We finally give some numerical tests and compare our algorithms with some related ones to show the performance and efficiency of our proposed algorithms.
{"title":"Simple proximal-type algorithms for equilibrium problems","authors":"Yonghong Yao, Abubakar Adamu, Yekini Shehu, Jen-Chih Yao","doi":"10.1007/s10898-024-01377-1","DOIUrl":"https://doi.org/10.1007/s10898-024-01377-1","url":null,"abstract":"<p>This paper proposes two simple and elegant proximal-type algorithms to solve equilibrium problems with pseudo-monotone bifunctions in the setting of Hilbert spaces. The proposed algorithms use one proximal point evaluation of the bifunction at each iteration. Consequently, prove that the sequences of iterates generated by the first algorithm converge weakly to a solution of the equilibrium problem (assuming existence) and obtain a linear convergence rate under standard assumptions. We also design a viscosity version of the first algorithm and obtain its corresponding strong convergence result. Some popular existing algorithms in the literature are recovered. We finally give some numerical tests and compare our algorithms with some related ones to show the performance and efficiency of our proposed algorithms.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"27 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140154990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-05DOI: 10.1007/s10898-024-01366-4
Hongwei Liu, Ting Wang, Zexian Liu
In this paper, we consider the problem that minimizing the sum of a nonsmooth function with a smooth one in the nonconvex setting, which arising in many contemporary applications such as machine learning, statistics, and signal/image processing. To solve this problem, we propose a new nonmonotone accelerated proximal gradient method with variable stepsize strategy. Note that incorporating inertial term into proximal gradient method is a simple and efficient acceleration technique, while the descent property of the proximal gradient algorithm will lost. In our algorithm, the iterates generated by inertial proximal gradient scheme are accepted when the objective function values decrease or increase appropriately; otherwise, the iteration point is generated by proximal gradient scheme, which makes the function values on a subset of iterates are decreasing. We also introduce a variable stepsize strategy, which does not need a line search or does not need to know the Lipschitz constant and makes the algorithm easy to implement. We show that the sequence of iterates generated by the algorithm converges to a critical point of the objective function. Further, under the assumption that the objective function satisfies the Kurdyka–Łojasiewicz inequality, we prove the convergence rates of the objective function values and the iterates. Moreover, numerical results on both convex and nonconvex problems are reported to demonstrate the effectiveness and superiority of the proposed method and stepsize strategy.
{"title":"A nonmonotone accelerated proximal gradient method with variable stepsize strategy for nonsmooth and nonconvex minimization problems","authors":"Hongwei Liu, Ting Wang, Zexian Liu","doi":"10.1007/s10898-024-01366-4","DOIUrl":"https://doi.org/10.1007/s10898-024-01366-4","url":null,"abstract":"<p>In this paper, we consider the problem that minimizing the sum of a nonsmooth function with a smooth one in the nonconvex setting, which arising in many contemporary applications such as machine learning, statistics, and signal/image processing. To solve this problem, we propose a new nonmonotone accelerated proximal gradient method with variable stepsize strategy. Note that incorporating inertial term into proximal gradient method is a simple and efficient acceleration technique, while the descent property of the proximal gradient algorithm will lost. In our algorithm, the iterates generated by inertial proximal gradient scheme are accepted when the objective function values decrease or increase appropriately; otherwise, the iteration point is generated by proximal gradient scheme, which makes the function values on a subset of iterates are decreasing. We also introduce a variable stepsize strategy, which does not need a line search or does not need to know the Lipschitz constant and makes the algorithm easy to implement. We show that the sequence of iterates generated by the algorithm converges to a critical point of the objective function. Further, under the assumption that the objective function satisfies the Kurdyka–Łojasiewicz inequality, we prove the convergence rates of the objective function values and the iterates. Moreover, numerical results on both convex and nonconvex problems are reported to demonstrate the effectiveness and superiority of the proposed method and stepsize strategy.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"192 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-03-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140032824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1007/s10898-024-01374-4
Abstract
Nonnegative tensor factorizations (NTF) have applications in statistics, computer vision, exploratory multi-way data analysis, and blind source separation. This paper studies randomized multiplicative updating algorithms for symmetric NTF via random projections and random samplings. For random projections, we consider two methods to generate the random matrix and analyze the computational complexity, while for random samplings the uniform sampling strategy and its variants are examined. The mixing of these two strategies is then considered. Some theoretical results are presented based on the bounds of the singular values of sub-Gaussian matrices and the fact that randomly sampling rows from an orthogonal matrix results in a well-conditioned matrix. These algorithms are easy to implement, and their efficiency is verified via test tensors from both synthetic and real datasets, such as for clustering facial images.
{"title":"Sketch-based multiplicative updating algorithms for symmetric nonnegative tensor factorizations with applications to face image clustering","authors":"","doi":"10.1007/s10898-024-01374-4","DOIUrl":"https://doi.org/10.1007/s10898-024-01374-4","url":null,"abstract":"<h3>Abstract</h3> <p>Nonnegative tensor factorizations (NTF) have applications in statistics, computer vision, exploratory multi-way data analysis, and blind source separation. This paper studies randomized multiplicative updating algorithms for symmetric NTF via random projections and random samplings. For random projections, we consider two methods to generate the random matrix and analyze the computational complexity, while for random samplings the uniform sampling strategy and its variants are examined. The mixing of these two strategies is then considered. Some theoretical results are presented based on the bounds of the singular values of sub-Gaussian matrices and the fact that randomly sampling rows from an orthogonal matrix results in a well-conditioned matrix. These algorithms are easy to implement, and their efficiency is verified via test tensors from both synthetic and real datasets, such as for clustering facial images.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"12 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140005115","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1007/s10898-023-01351-3
Gabriela Kováčová, Firdevs Ulus
It is possible to solve unbounded convex vector optimization problems (CVOPs) in two phases: (1) computing or approximating the recession cone of the upper image and (2) solving the equivalent bounded CVOP where the ordering cone is extended based on the first phase. In this paper, we consider unbounded CVOPs and propose an alternative solution methodology to compute or approximate the recession cone of the upper image. In particular, we relate the dual of the recession cone with the Lagrange dual of weighted sum scalarization problems whenever the dual problem can be written explicitly. Computing this set requires solving a convex (or polyhedral) projection problem. We show that this methodology can be applied to semidefinite, quadratic, and linear vector optimization problems and provide some numerical examples.
{"title":"Computing the recession cone of a convex upper image via convex projection","authors":"Gabriela Kováčová, Firdevs Ulus","doi":"10.1007/s10898-023-01351-3","DOIUrl":"https://doi.org/10.1007/s10898-023-01351-3","url":null,"abstract":"<p>It is possible to solve unbounded convex vector optimization problems (CVOPs) in two phases: (1) computing or approximating the recession cone of the upper image and (2) solving the equivalent bounded CVOP where the ordering cone is extended based on the first phase. In this paper, we consider unbounded CVOPs and propose an alternative solution methodology to compute or approximate the recession cone of the upper image. In particular, we relate the dual of the recession cone with the Lagrange dual of weighted sum scalarization problems whenever the dual problem can be written explicitly. Computing this set requires solving a convex (or polyhedral) projection problem. We show that this methodology can be applied to semidefinite, quadratic, and linear vector optimization problems and provide some numerical examples.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"115 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140004997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-28DOI: 10.1007/s10898-024-01373-5
Marco Baioletti, Valentino Santucci, Marco Tomassini
During the last decades many metaheuristics for global numerical optimization have been proposed. Among them, Basin Hopping is very simple and straightforward to implement, although rarely used outside its original Physical Chemistry community. In this work, our aim is to compare Basin Hopping, and two population variants of it, with readily available implementations of the well known metaheuristics Differential Evolution, Particle Swarm Optimization, and Covariance Matrix Adaptation Evolution Strategy. We perform numerical experiments using the IOH profiler environment with the BBOB test function set and two difficult real-world problems. The experiments were carried out in two different but complementary ways: by measuring the performance under a fixed budget of function evaluations and by considering a fixed target value. The general conclusion is that Basin Hopping and its newly introduced population variant are almost as good as Covariance Matrix Adaptation on the synthetic benchmark functions and better than it on the two hard cluster energy minimization problems. Thus, the proposed analyses show that Basin Hopping can be considered a good candidate for global numerical optimization problems along with the more established metaheuristics, especially if one wants to obtain quick and reliable results on an unknown problem.
{"title":"A performance analysis of Basin hopping compared to established metaheuristics for global optimization","authors":"Marco Baioletti, Valentino Santucci, Marco Tomassini","doi":"10.1007/s10898-024-01373-5","DOIUrl":"https://doi.org/10.1007/s10898-024-01373-5","url":null,"abstract":"<p>During the last decades many metaheuristics for global numerical optimization have been proposed. Among them, Basin Hopping is very simple and straightforward to implement, although rarely used outside its original Physical Chemistry community. In this work, our aim is to compare Basin Hopping, and two population variants of it, with readily available implementations of the well known metaheuristics Differential Evolution, Particle Swarm Optimization, and Covariance Matrix Adaptation Evolution Strategy. We perform numerical experiments using the <i>IOH profiler</i> environment with the BBOB test function set and two difficult real-world problems. The experiments were carried out in two different but complementary ways: by measuring the performance under a fixed budget of function evaluations and by considering a fixed target value. The general conclusion is that Basin Hopping and its newly introduced population variant are almost as good as Covariance Matrix Adaptation on the synthetic benchmark functions and better than it on the two hard cluster energy minimization problems. Thus, the proposed analyses show that Basin Hopping can be considered a good candidate for global numerical optimization problems along with the more established metaheuristics, especially if one wants to obtain quick and reliable results on an unknown problem.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"914 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140005393","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-22DOI: 10.1007/s10898-024-01371-7
Cheng Lu, Wenguo Yang
We study the non-submodular maximization problem, in which the objective function is characterized by parameters, subject to a cardinality or (p)-system constraint. By adapting the Threshold-Greedy algorithm for the submodular maximization, we present two deterministic algorithms for approximately solving the non-submodular maximization problem. Our analysis shows that the algorithms we propose requires much less function evaluations than existing algorithms, while providing comparable approximation guarantees. Moreover, numerical experiment results are presented to validate the theoretical analysis. Our results not only fill a gap in the (non-)submodular maximization, but also generalize and improve several existing results on closely related optimization problems.
{"title":"Fast deterministic algorithms for non-submodular maximization with strong performance guarantees","authors":"Cheng Lu, Wenguo Yang","doi":"10.1007/s10898-024-01371-7","DOIUrl":"https://doi.org/10.1007/s10898-024-01371-7","url":null,"abstract":"<p>We study the non-submodular maximization problem, in which the objective function is characterized by parameters, subject to a cardinality or <span>(p)</span>-system constraint. By adapting the <span>Threshold-Greedy</span> algorithm for the submodular maximization, we present two deterministic algorithms for approximately solving the non-submodular maximization problem. Our analysis shows that the algorithms we propose requires much less function evaluations than existing algorithms, while providing comparable approximation guarantees. Moreover, numerical experiment results are presented to validate the theoretical analysis. Our results not only fill a gap in the (non-)submodular maximization, but also generalize and improve several existing results on closely related optimization problems.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"1 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-02-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139924361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-22DOI: 10.1007/s10898-023-01360-2
Abstract
This paper explores a class of nonlinear Adjustable Robust Optimization (ARO) problems, containing here-and-now and wait-and-see variables, with uncertainty in the objective function and constraints. By applying Fenchel’s duality on the wait-and-see variables, we obtain an equivalent dual reformulation, which is a nonlinear static robust optimization problem. Using the dual formulation, we provide conditions under which the ARO problem is convex on the here-and-now decision. Furthermore, since the dual formulation contains a non-concave maximization on the uncertain parameter, we use perspective relaxation and an alternating method to handle the non-concavity. By employing the perspective relaxation, we obtain an upper bound, which we show is the same as the static relaxation of the considered problem. Moreover, invoking the alternating method, we design a new dual-based cutting plane algorithm that is able to find a reasonable lower bound for the optimal objective value of the considered nonlinear ARO model. In addition to sketching and establishing the theoretical features of the algorithms, including convergence analysis, by numerical experiments we reveal the abilities of our cutting plane algorithm in producing locally robust solutions with an acceptable optimality gap.
摘要 本文探讨了一类非线性可调稳健优化(ARO)问题,该问题包含此时此地和等待观察变量,目标函数和约束条件具有不确定性。通过对 "等待-观察 "变量应用 Fenchel 对偶,我们得到了一个等价的对偶重述,即一个非线性静态鲁棒优化问题。利用对偶表述,我们提供了 ARO 问题在此时此地的决策上具有凸性的条件。此外,由于对偶表述包含对不确定参数的非凹性最大化,我们使用透视松弛和交替法来处理非凹性。通过使用透视松弛法,我们得到了一个上界,并证明它与所考虑问题的静态松弛法相同。此外,利用交替法,我们设计了一种新的基于对偶的切割面算法,能够为所考虑的非线性 ARO 模型的最优目标值找到一个合理的下界。除了勾勒和建立算法的理论特征(包括收敛性分析)外,我们还通过数值实验揭示了我们的切割面算法在产生具有可接受最优性差距的局部稳健解方面的能力。
{"title":"A new dual-based cutting plane algorithm for nonlinear adjustable robust optimization","authors":"","doi":"10.1007/s10898-023-01360-2","DOIUrl":"https://doi.org/10.1007/s10898-023-01360-2","url":null,"abstract":"<h3>Abstract</h3> <p>This paper explores a class of nonlinear Adjustable Robust Optimization (ARO) problems, containing here-and-now and wait-and-see variables, with uncertainty in the objective function and constraints. By applying Fenchel’s duality on the wait-and-see variables, we obtain an equivalent dual reformulation, which is a nonlinear static robust optimization problem. Using the dual formulation, we provide conditions under which the ARO problem is convex on the here-and-now decision. Furthermore, since the dual formulation contains a non-concave maximization on the uncertain parameter, we use perspective relaxation and an alternating method to handle the non-concavity. By employing the perspective relaxation, we obtain an upper bound, which we show is the same as the static relaxation of the considered problem. Moreover, invoking the alternating method, we design a new dual-based cutting plane algorithm that is able to find a reasonable lower bound for the optimal objective value of the considered nonlinear ARO model. In addition to sketching and establishing the theoretical features of the algorithms, including convergence analysis, by numerical experiments we reveal the abilities of our cutting plane algorithm in producing locally robust solutions with an acceptable optimality gap.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"36 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-02-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139924360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-15DOI: 10.1007/s10898-023-01358-w
Hongwei Jiao, Binbin Li, Wenqiang Yang
In this paper, we investigate a generalized multiplicative problem (GMP) that is known to be NP-hard even with one linear product term. We first introduce some criterion-space variables to obtain an equivalent problem of the GMP. A criterion-space branch-reduction-bound algorithm is then designed, which integrates some basic operations such as the two-level linear relaxation technique, rectangle branching rule and criterion-space region reduction technologies. The global convergence of the presented algorithm is proved by means of the subsequent solutions of a series of linear relaxation problems, and its maximum number of iterations is estimated on the basis of exhaustiveness of branching rule. Finally, numerical results demonstrate the presented algorithm can efficiently find the global optimum solutions for some test instances with the robustness.
{"title":"A criterion-space branch-reduction-bound algorithm for solving generalized multiplicative problems","authors":"Hongwei Jiao, Binbin Li, Wenqiang Yang","doi":"10.1007/s10898-023-01358-w","DOIUrl":"https://doi.org/10.1007/s10898-023-01358-w","url":null,"abstract":"<p>In this paper, we investigate a generalized multiplicative problem (GMP) that is known to be NP-hard even with one linear product term. We first introduce some criterion-space variables to obtain an equivalent problem of the GMP. A criterion-space branch-reduction-bound algorithm is then designed, which integrates some basic operations such as the two-level linear relaxation technique, rectangle branching rule and criterion-space region reduction technologies. The global convergence of the presented algorithm is proved by means of the subsequent solutions of a series of linear relaxation problems, and its maximum number of iterations is estimated on the basis of exhaustiveness of branching rule. Finally, numerical results demonstrate the presented algorithm can efficiently find the global optimum solutions for some test instances with the robustness.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"56 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139759405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}