Pub Date : 2024-03-14DOI: 10.1007/s10898-024-01377-1
Yonghong Yao, Abubakar Adamu, Yekini Shehu, Jen-Chih Yao
This paper proposes two simple and elegant proximal-type algorithms to solve equilibrium problems with pseudo-monotone bifunctions in the setting of Hilbert spaces. The proposed algorithms use one proximal point evaluation of the bifunction at each iteration. Consequently, prove that the sequences of iterates generated by the first algorithm converge weakly to a solution of the equilibrium problem (assuming existence) and obtain a linear convergence rate under standard assumptions. We also design a viscosity version of the first algorithm and obtain its corresponding strong convergence result. Some popular existing algorithms in the literature are recovered. We finally give some numerical tests and compare our algorithms with some related ones to show the performance and efficiency of our proposed algorithms.
{"title":"Simple proximal-type algorithms for equilibrium problems","authors":"Yonghong Yao, Abubakar Adamu, Yekini Shehu, Jen-Chih Yao","doi":"10.1007/s10898-024-01377-1","DOIUrl":"https://doi.org/10.1007/s10898-024-01377-1","url":null,"abstract":"<p>This paper proposes two simple and elegant proximal-type algorithms to solve equilibrium problems with pseudo-monotone bifunctions in the setting of Hilbert spaces. The proposed algorithms use one proximal point evaluation of the bifunction at each iteration. Consequently, prove that the sequences of iterates generated by the first algorithm converge weakly to a solution of the equilibrium problem (assuming existence) and obtain a linear convergence rate under standard assumptions. We also design a viscosity version of the first algorithm and obtain its corresponding strong convergence result. Some popular existing algorithms in the literature are recovered. We finally give some numerical tests and compare our algorithms with some related ones to show the performance and efficiency of our proposed algorithms.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"27 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140154990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-05DOI: 10.1007/s10898-024-01366-4
Hongwei Liu, Ting Wang, Zexian Liu
In this paper, we consider the problem that minimizing the sum of a nonsmooth function with a smooth one in the nonconvex setting, which arising in many contemporary applications such as machine learning, statistics, and signal/image processing. To solve this problem, we propose a new nonmonotone accelerated proximal gradient method with variable stepsize strategy. Note that incorporating inertial term into proximal gradient method is a simple and efficient acceleration technique, while the descent property of the proximal gradient algorithm will lost. In our algorithm, the iterates generated by inertial proximal gradient scheme are accepted when the objective function values decrease or increase appropriately; otherwise, the iteration point is generated by proximal gradient scheme, which makes the function values on a subset of iterates are decreasing. We also introduce a variable stepsize strategy, which does not need a line search or does not need to know the Lipschitz constant and makes the algorithm easy to implement. We show that the sequence of iterates generated by the algorithm converges to a critical point of the objective function. Further, under the assumption that the objective function satisfies the Kurdyka–Łojasiewicz inequality, we prove the convergence rates of the objective function values and the iterates. Moreover, numerical results on both convex and nonconvex problems are reported to demonstrate the effectiveness and superiority of the proposed method and stepsize strategy.
{"title":"A nonmonotone accelerated proximal gradient method with variable stepsize strategy for nonsmooth and nonconvex minimization problems","authors":"Hongwei Liu, Ting Wang, Zexian Liu","doi":"10.1007/s10898-024-01366-4","DOIUrl":"https://doi.org/10.1007/s10898-024-01366-4","url":null,"abstract":"<p>In this paper, we consider the problem that minimizing the sum of a nonsmooth function with a smooth one in the nonconvex setting, which arising in many contemporary applications such as machine learning, statistics, and signal/image processing. To solve this problem, we propose a new nonmonotone accelerated proximal gradient method with variable stepsize strategy. Note that incorporating inertial term into proximal gradient method is a simple and efficient acceleration technique, while the descent property of the proximal gradient algorithm will lost. In our algorithm, the iterates generated by inertial proximal gradient scheme are accepted when the objective function values decrease or increase appropriately; otherwise, the iteration point is generated by proximal gradient scheme, which makes the function values on a subset of iterates are decreasing. We also introduce a variable stepsize strategy, which does not need a line search or does not need to know the Lipschitz constant and makes the algorithm easy to implement. We show that the sequence of iterates generated by the algorithm converges to a critical point of the objective function. Further, under the assumption that the objective function satisfies the Kurdyka–Łojasiewicz inequality, we prove the convergence rates of the objective function values and the iterates. Moreover, numerical results on both convex and nonconvex problems are reported to demonstrate the effectiveness and superiority of the proposed method and stepsize strategy.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"192 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-03-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140032824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1007/s10898-024-01374-4
Abstract
Nonnegative tensor factorizations (NTF) have applications in statistics, computer vision, exploratory multi-way data analysis, and blind source separation. This paper studies randomized multiplicative updating algorithms for symmetric NTF via random projections and random samplings. For random projections, we consider two methods to generate the random matrix and analyze the computational complexity, while for random samplings the uniform sampling strategy and its variants are examined. The mixing of these two strategies is then considered. Some theoretical results are presented based on the bounds of the singular values of sub-Gaussian matrices and the fact that randomly sampling rows from an orthogonal matrix results in a well-conditioned matrix. These algorithms are easy to implement, and their efficiency is verified via test tensors from both synthetic and real datasets, such as for clustering facial images.
{"title":"Sketch-based multiplicative updating algorithms for symmetric nonnegative tensor factorizations with applications to face image clustering","authors":"","doi":"10.1007/s10898-024-01374-4","DOIUrl":"https://doi.org/10.1007/s10898-024-01374-4","url":null,"abstract":"<h3>Abstract</h3> <p>Nonnegative tensor factorizations (NTF) have applications in statistics, computer vision, exploratory multi-way data analysis, and blind source separation. This paper studies randomized multiplicative updating algorithms for symmetric NTF via random projections and random samplings. For random projections, we consider two methods to generate the random matrix and analyze the computational complexity, while for random samplings the uniform sampling strategy and its variants are examined. The mixing of these two strategies is then considered. Some theoretical results are presented based on the bounds of the singular values of sub-Gaussian matrices and the fact that randomly sampling rows from an orthogonal matrix results in a well-conditioned matrix. These algorithms are easy to implement, and their efficiency is verified via test tensors from both synthetic and real datasets, such as for clustering facial images.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"12 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140005115","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1007/s10898-023-01351-3
Gabriela Kováčová, Firdevs Ulus
It is possible to solve unbounded convex vector optimization problems (CVOPs) in two phases: (1) computing or approximating the recession cone of the upper image and (2) solving the equivalent bounded CVOP where the ordering cone is extended based on the first phase. In this paper, we consider unbounded CVOPs and propose an alternative solution methodology to compute or approximate the recession cone of the upper image. In particular, we relate the dual of the recession cone with the Lagrange dual of weighted sum scalarization problems whenever the dual problem can be written explicitly. Computing this set requires solving a convex (or polyhedral) projection problem. We show that this methodology can be applied to semidefinite, quadratic, and linear vector optimization problems and provide some numerical examples.
{"title":"Computing the recession cone of a convex upper image via convex projection","authors":"Gabriela Kováčová, Firdevs Ulus","doi":"10.1007/s10898-023-01351-3","DOIUrl":"https://doi.org/10.1007/s10898-023-01351-3","url":null,"abstract":"<p>It is possible to solve unbounded convex vector optimization problems (CVOPs) in two phases: (1) computing or approximating the recession cone of the upper image and (2) solving the equivalent bounded CVOP where the ordering cone is extended based on the first phase. In this paper, we consider unbounded CVOPs and propose an alternative solution methodology to compute or approximate the recession cone of the upper image. In particular, we relate the dual of the recession cone with the Lagrange dual of weighted sum scalarization problems whenever the dual problem can be written explicitly. Computing this set requires solving a convex (or polyhedral) projection problem. We show that this methodology can be applied to semidefinite, quadratic, and linear vector optimization problems and provide some numerical examples.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"115 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140004997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-28DOI: 10.1007/s10898-024-01373-5
Marco Baioletti, Valentino Santucci, Marco Tomassini
During the last decades many metaheuristics for global numerical optimization have been proposed. Among them, Basin Hopping is very simple and straightforward to implement, although rarely used outside its original Physical Chemistry community. In this work, our aim is to compare Basin Hopping, and two population variants of it, with readily available implementations of the well known metaheuristics Differential Evolution, Particle Swarm Optimization, and Covariance Matrix Adaptation Evolution Strategy. We perform numerical experiments using the IOH profiler environment with the BBOB test function set and two difficult real-world problems. The experiments were carried out in two different but complementary ways: by measuring the performance under a fixed budget of function evaluations and by considering a fixed target value. The general conclusion is that Basin Hopping and its newly introduced population variant are almost as good as Covariance Matrix Adaptation on the synthetic benchmark functions and better than it on the two hard cluster energy minimization problems. Thus, the proposed analyses show that Basin Hopping can be considered a good candidate for global numerical optimization problems along with the more established metaheuristics, especially if one wants to obtain quick and reliable results on an unknown problem.
{"title":"A performance analysis of Basin hopping compared to established metaheuristics for global optimization","authors":"Marco Baioletti, Valentino Santucci, Marco Tomassini","doi":"10.1007/s10898-024-01373-5","DOIUrl":"https://doi.org/10.1007/s10898-024-01373-5","url":null,"abstract":"<p>During the last decades many metaheuristics for global numerical optimization have been proposed. Among them, Basin Hopping is very simple and straightforward to implement, although rarely used outside its original Physical Chemistry community. In this work, our aim is to compare Basin Hopping, and two population variants of it, with readily available implementations of the well known metaheuristics Differential Evolution, Particle Swarm Optimization, and Covariance Matrix Adaptation Evolution Strategy. We perform numerical experiments using the <i>IOH profiler</i> environment with the BBOB test function set and two difficult real-world problems. The experiments were carried out in two different but complementary ways: by measuring the performance under a fixed budget of function evaluations and by considering a fixed target value. The general conclusion is that Basin Hopping and its newly introduced population variant are almost as good as Covariance Matrix Adaptation on the synthetic benchmark functions and better than it on the two hard cluster energy minimization problems. Thus, the proposed analyses show that Basin Hopping can be considered a good candidate for global numerical optimization problems along with the more established metaheuristics, especially if one wants to obtain quick and reliable results on an unknown problem.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"914 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140005393","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-22DOI: 10.1007/s10898-024-01371-7
Cheng Lu, Wenguo Yang
We study the non-submodular maximization problem, in which the objective function is characterized by parameters, subject to a cardinality or (p)-system constraint. By adapting the Threshold-Greedy algorithm for the submodular maximization, we present two deterministic algorithms for approximately solving the non-submodular maximization problem. Our analysis shows that the algorithms we propose requires much less function evaluations than existing algorithms, while providing comparable approximation guarantees. Moreover, numerical experiment results are presented to validate the theoretical analysis. Our results not only fill a gap in the (non-)submodular maximization, but also generalize and improve several existing results on closely related optimization problems.
{"title":"Fast deterministic algorithms for non-submodular maximization with strong performance guarantees","authors":"Cheng Lu, Wenguo Yang","doi":"10.1007/s10898-024-01371-7","DOIUrl":"https://doi.org/10.1007/s10898-024-01371-7","url":null,"abstract":"<p>We study the non-submodular maximization problem, in which the objective function is characterized by parameters, subject to a cardinality or <span>(p)</span>-system constraint. By adapting the <span>Threshold-Greedy</span> algorithm for the submodular maximization, we present two deterministic algorithms for approximately solving the non-submodular maximization problem. Our analysis shows that the algorithms we propose requires much less function evaluations than existing algorithms, while providing comparable approximation guarantees. Moreover, numerical experiment results are presented to validate the theoretical analysis. Our results not only fill a gap in the (non-)submodular maximization, but also generalize and improve several existing results on closely related optimization problems.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"1 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-02-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139924361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-22DOI: 10.1007/s10898-023-01360-2
Abstract
This paper explores a class of nonlinear Adjustable Robust Optimization (ARO) problems, containing here-and-now and wait-and-see variables, with uncertainty in the objective function and constraints. By applying Fenchel’s duality on the wait-and-see variables, we obtain an equivalent dual reformulation, which is a nonlinear static robust optimization problem. Using the dual formulation, we provide conditions under which the ARO problem is convex on the here-and-now decision. Furthermore, since the dual formulation contains a non-concave maximization on the uncertain parameter, we use perspective relaxation and an alternating method to handle the non-concavity. By employing the perspective relaxation, we obtain an upper bound, which we show is the same as the static relaxation of the considered problem. Moreover, invoking the alternating method, we design a new dual-based cutting plane algorithm that is able to find a reasonable lower bound for the optimal objective value of the considered nonlinear ARO model. In addition to sketching and establishing the theoretical features of the algorithms, including convergence analysis, by numerical experiments we reveal the abilities of our cutting plane algorithm in producing locally robust solutions with an acceptable optimality gap.
摘要 本文探讨了一类非线性可调稳健优化(ARO)问题,该问题包含此时此地和等待观察变量,目标函数和约束条件具有不确定性。通过对 "等待-观察 "变量应用 Fenchel 对偶,我们得到了一个等价的对偶重述,即一个非线性静态鲁棒优化问题。利用对偶表述,我们提供了 ARO 问题在此时此地的决策上具有凸性的条件。此外,由于对偶表述包含对不确定参数的非凹性最大化,我们使用透视松弛和交替法来处理非凹性。通过使用透视松弛法,我们得到了一个上界,并证明它与所考虑问题的静态松弛法相同。此外,利用交替法,我们设计了一种新的基于对偶的切割面算法,能够为所考虑的非线性 ARO 模型的最优目标值找到一个合理的下界。除了勾勒和建立算法的理论特征(包括收敛性分析)外,我们还通过数值实验揭示了我们的切割面算法在产生具有可接受最优性差距的局部稳健解方面的能力。
{"title":"A new dual-based cutting plane algorithm for nonlinear adjustable robust optimization","authors":"","doi":"10.1007/s10898-023-01360-2","DOIUrl":"https://doi.org/10.1007/s10898-023-01360-2","url":null,"abstract":"<h3>Abstract</h3> <p>This paper explores a class of nonlinear Adjustable Robust Optimization (ARO) problems, containing here-and-now and wait-and-see variables, with uncertainty in the objective function and constraints. By applying Fenchel’s duality on the wait-and-see variables, we obtain an equivalent dual reformulation, which is a nonlinear static robust optimization problem. Using the dual formulation, we provide conditions under which the ARO problem is convex on the here-and-now decision. Furthermore, since the dual formulation contains a non-concave maximization on the uncertain parameter, we use perspective relaxation and an alternating method to handle the non-concavity. By employing the perspective relaxation, we obtain an upper bound, which we show is the same as the static relaxation of the considered problem. Moreover, invoking the alternating method, we design a new dual-based cutting plane algorithm that is able to find a reasonable lower bound for the optimal objective value of the considered nonlinear ARO model. In addition to sketching and establishing the theoretical features of the algorithms, including convergence analysis, by numerical experiments we reveal the abilities of our cutting plane algorithm in producing locally robust solutions with an acceptable optimality gap.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"36 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-02-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139924360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-15DOI: 10.1007/s10898-023-01358-w
Hongwei Jiao, Binbin Li, Wenqiang Yang
In this paper, we investigate a generalized multiplicative problem (GMP) that is known to be NP-hard even with one linear product term. We first introduce some criterion-space variables to obtain an equivalent problem of the GMP. A criterion-space branch-reduction-bound algorithm is then designed, which integrates some basic operations such as the two-level linear relaxation technique, rectangle branching rule and criterion-space region reduction technologies. The global convergence of the presented algorithm is proved by means of the subsequent solutions of a series of linear relaxation problems, and its maximum number of iterations is estimated on the basis of exhaustiveness of branching rule. Finally, numerical results demonstrate the presented algorithm can efficiently find the global optimum solutions for some test instances with the robustness.
{"title":"A criterion-space branch-reduction-bound algorithm for solving generalized multiplicative problems","authors":"Hongwei Jiao, Binbin Li, Wenqiang Yang","doi":"10.1007/s10898-023-01358-w","DOIUrl":"https://doi.org/10.1007/s10898-023-01358-w","url":null,"abstract":"<p>In this paper, we investigate a generalized multiplicative problem (GMP) that is known to be NP-hard even with one linear product term. We first introduce some criterion-space variables to obtain an equivalent problem of the GMP. A criterion-space branch-reduction-bound algorithm is then designed, which integrates some basic operations such as the two-level linear relaxation technique, rectangle branching rule and criterion-space region reduction technologies. The global convergence of the presented algorithm is proved by means of the subsequent solutions of a series of linear relaxation problems, and its maximum number of iterations is estimated on the basis of exhaustiveness of branching rule. Finally, numerical results demonstrate the presented algorithm can efficiently find the global optimum solutions for some test instances with the robustness.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"56 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139759405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-15DOI: 10.1007/s10898-024-01364-6
Sabah Bushaj, İ. Esra Büyüktahtakın
In this paper, we address the difficulty of solving large-scale multi-dimensional knapsack instances (MKP), presenting a novel deep reinforcement learning (DRL) framework. In this DRL framework, we train different agents compatible with a discrete action space for sequential decision-making while still satisfying any resource constraint of the MKP. This novel framework incorporates the decision variable values in the 2D DRL where the agent is responsible for assigning a value of 1 or 0 to each of the variables. To the best of our knowledge, this is the first DRL model of its kind in which a 2D environment is formulated, and an element of the DRL solution matrix represents an item of the MKP. Our framework is configured to solve MKP instances of different dimensions and distributions. We propose a K-means approach to obtain an initial feasible solution that is used to train the DRL agent. We train four different agents in our framework and present the results comparing each of them with the CPLEX commercial solver. The results show that our agents can learn and generalize over instances with different sizes and distributions. Our DRL framework shows that it can solve medium-sized instances at least 45 times faster in CPU solution time and at least 10 times faster for large instances, with a maximum solution gap of 0.28% compared to the performance of CPLEX. Furthermore, at least 95% of the items are predicted in line with the CPLEX solution. Computations with DRL also provide a better optimality gap with respect to state-of-the-art approaches.
{"title":"A K-means Supported Reinforcement Learning Framework to Multi-dimensional Knapsack","authors":"Sabah Bushaj, İ. Esra Büyüktahtakın","doi":"10.1007/s10898-024-01364-6","DOIUrl":"https://doi.org/10.1007/s10898-024-01364-6","url":null,"abstract":"<p>In this paper, we address the difficulty of solving large-scale multi-dimensional knapsack instances (MKP), presenting a novel deep reinforcement learning (DRL) framework. In this DRL framework, we train different agents compatible with a discrete action space for sequential decision-making while still satisfying any resource constraint of the MKP. This novel framework incorporates the decision variable values in the 2D DRL where the agent is responsible for assigning a value of 1 or 0 to each of the variables. To the best of our knowledge, this is the first DRL model of its kind in which a 2D environment is formulated, and an element of the DRL solution matrix represents an item of the MKP. Our framework is configured to solve MKP instances of different dimensions and distributions. We propose a K-means approach to obtain an initial feasible solution that is used to train the DRL agent. We train four different agents in our framework and present the results comparing each of them with the CPLEX commercial solver. The results show that our agents can learn and generalize over instances with different sizes and distributions. Our DRL framework shows that it can solve medium-sized instances at least 45 times faster in CPU solution time and at least 10 times faster for large instances, with a maximum solution gap of 0.28% compared to the performance of CPLEX. Furthermore, at least 95% of the items are predicted in line with the CPLEX solution. Computations with DRL also provide a better optimality gap with respect to state-of-the-art approaches.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"121 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139759470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-15DOI: 10.1007/s10898-024-01365-5
Zehao Xiao, Liwei Zhang
An online majorized semi-proximal alternating direction method of multiplier (Online-mspADMM) is proposed for a broad class of online linearly constrained composite optimization problems. A majorized technique is adopted to produce subproblems which can be easily solved. Under mild assumptions, we establish (mathcal {O}(sqrt{N})) objective regret and (mathcal {O}(sqrt{N})) constraint violation regret at round N. We apply the Online-mspADMM to solve different types of online regularized logistic regression problems. The numerical results on synthetic data sets verify the theoretical result about regrets.
针对各类在线线性约束复合优化问题,提出了一种在线大化半近似交替方向乘法(Online-mspADMM)。该方法采用大化技术来生成易于求解的子问题。在温和的假设条件下,我们在第 N 轮建立了 (mathcal {O}(sqrt{N})) 目标遗憾和 (mathcal {O}(sqrt{N})) 约束违反遗憾。在合成数据集上的数值结果验证了关于遗憾的理论结果。
{"title":"Regret analysis of an online majorized semi-proximal ADMM for online composite optimization","authors":"Zehao Xiao, Liwei Zhang","doi":"10.1007/s10898-024-01365-5","DOIUrl":"https://doi.org/10.1007/s10898-024-01365-5","url":null,"abstract":"<p>An online majorized semi-proximal alternating direction method of multiplier (Online-mspADMM) is proposed for a broad class of online linearly constrained composite optimization problems. A majorized technique is adopted to produce subproblems which can be easily solved. Under mild assumptions, we establish <span>(mathcal {O}(sqrt{N}))</span> objective regret and <span>(mathcal {O}(sqrt{N}))</span> constraint violation regret at round <i>N</i>. We apply the Online-mspADMM to solve different types of online regularized logistic regression problems. The numerical results on synthetic data sets verify the theoretical result about regrets.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"68 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139759413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}