Pub Date : 2024-01-22DOI: 10.1007/s10898-023-01362-0
Abstract
In this article, we propose an interval constraint programming method for globally solving catalog-based categorical optimization problems. It supports catalogs of arbitrary size and properties of arbitrary dimension, and does not require any modeling effort from the user. A novel catalog-based contractor (or filtering operator) guarantees consistency between the categorical properties and the existing catalog items. This results in an intuitive and generic approach that is exact, rigorous (robust to roundoff errors) and can be easily implemented in an off-the-shelf interval-based continuous solver that interleaves branching and constraint propagation. We demonstrate the validity of the approach on a numerical problem in which a categorical variable is described by a two-dimensional property space. A Julia prototype is available as open-source software under the MIT license at https://github.com/cvanaret/CateGOrical.jl.
摘要 本文提出了一种区间约束编程方法,用于全局求解基于目录的分类优化问题。该方法支持任意大小的目录和任意维度的属性,并且不需要用户做任何建模工作。一种新颖的基于目录的承包商(或过滤算子)保证了分类属性与现有目录项之间的一致性。这就产生了一种直观而通用的方法,它精确、严谨(对舍入误差具有鲁棒性),并且可以在现成的基于区间的连续求解器中轻松实现,该求解器将分支和约束传播交织在一起。我们在一个二维属性空间描述分类变量的数值问题上演示了该方法的有效性。Julia 原型作为开源软件,在 MIT 许可下发布在 https://github.com/cvanaret/CateGOrical.jl 上。
{"title":"Interval constraint programming for globally solving catalog-based categorical optimization","authors":"","doi":"10.1007/s10898-023-01362-0","DOIUrl":"https://doi.org/10.1007/s10898-023-01362-0","url":null,"abstract":"<h3>Abstract</h3> <p>In this article, we propose an interval constraint programming method for globally solving catalog-based categorical optimization problems. It supports catalogs of arbitrary size and properties of arbitrary dimension, and does not require any modeling effort from the user. A novel catalog-based contractor (or filtering operator) guarantees consistency between the categorical properties and the existing catalog items. This results in an intuitive and generic approach that is exact, rigorous (robust to roundoff errors) and can be easily implemented in an off-the-shelf interval-based continuous solver that interleaves branching and constraint propagation. We demonstrate the validity of the approach on a numerical problem in which a categorical variable is described by a two-dimensional property space. A Julia prototype is available as open-source software under the MIT license at https://github.com/cvanaret/CateGOrical.jl.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"44 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139518719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-22DOI: 10.1007/s10898-023-01356-y
Luze Xu, Jon Lee
MINLO (mixed-integer nonlinear optimization) formulations of the disjunction between the origin and a polytope via a binary indicator variable have broad applicability in nonlinear combinatorial optimization, for modeling a fixed cost c associated with carrying out a set of d activities and a convex variable cost function f associated with the levels of the activities. The perspective relaxation is often used to solve such models to optimality in a branch-and-bound context, especially in the context in which f is univariate (e.g., in Markowitz-style portfolio optimization). But such a relaxation typically requires conic solvers and are typically not compatible with general-purpose NLP software which can accommodate additional classes of constraints. This motivates the study of weaker relaxations to investigate when simpler relaxations may be adequate. Comparing the volume (i.e., Lebesgue measure) of the relaxations as means of comparing them, we lift some of the results related to univariate functions f to the multivariate case. Along the way, we survey, connect and extend relevant results on integration over a simplex, some of which we concretely employ, and others of which can be used for further exploration on our main subject.
MINLO(混合整数非线性优化)公式是通过二元指示变量对原点和多面体之间的析取,在非线性组合优化中具有广泛的适用性,可用于模拟与开展一组 d 项活动相关的固定成本 c 和与活动水平相关的凸变量成本函数 f。透视松弛法常用于在分支和边界情境中求解此类模型的最优性,尤其是在 f 是单变量的情境中(例如,在马科维茨式的组合优化中)。但这种松弛通常需要圆锥求解器,而且通常与能容纳更多类别约束的通用 NLP 软件不兼容。这就促使我们对较弱的松弛进行研究,以探究更简单的松弛何时可以满足要求。通过比较松弛的体积(即 Lebesgue 度量),我们将一些与单变量函数 f 相关的结果推广到多变量情况中。在此过程中,我们考察、连接并扩展了关于单纯形上积分的相关结果,其中一些我们已具体运用,另一些则可用于进一步探索我们的主要课题。
{"title":"Gaining or losing perspective for convex multivariate functions on a simplex","authors":"Luze Xu, Jon Lee","doi":"10.1007/s10898-023-01356-y","DOIUrl":"https://doi.org/10.1007/s10898-023-01356-y","url":null,"abstract":"<p>MINLO (mixed-integer nonlinear optimization) formulations of the disjunction between the origin and a polytope via a binary indicator variable have broad applicability in nonlinear combinatorial optimization, for modeling a fixed cost <i>c</i> associated with carrying out a set of <i>d</i> activities and a convex variable cost function <i>f</i> associated with the levels of the activities. The perspective relaxation is often used to solve such models to optimality in a branch-and-bound context, especially in the context in which <i>f</i> is univariate (e.g., in Markowitz-style portfolio optimization). But such a relaxation typically requires conic solvers and are typically not compatible with general-purpose NLP software which can accommodate additional classes of constraints. This motivates the study of weaker relaxations to investigate when simpler relaxations may be adequate. Comparing the volume (i.e., Lebesgue measure) of the relaxations as means of comparing them, we lift some of the results related to univariate functions <i>f</i> to the multivariate case. Along the way, we survey, connect and extend relevant results on integration over a simplex, some of which we concretely employ, and others of which can be used for further exploration on our main subject.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"121 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139514938","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-22DOI: 10.1007/s10898-023-01335-3
Cao Thanh Tinh, Thai Doan Chuong
This paper studies a class of multiobjective convex polynomial problems, where both the constraint and objective functions involve uncertain parameters that reside in ellipsoidal uncertainty sets. Employing the robust deterministic approach, we provide necessary conditions and sufficient conditions, which are exhibited in relation to second order cone conditions, for robust (weak) Pareto solutions of the uncertain multiobjective optimization problem. A dual multiobjective problem is proposed to examine robust converse, robust weak and robust strong duality relations between the primal and dual problems. Moreover, we establish robust solution relationships between the uncertain multiobjective optimization program and a (scalar) second order cone programming relaxation problem of a corresponding weighted-sum optimization problem. This in particular shows that we can find a robust (weak) Pareto solution of the uncertain multiobjective optimization problem by solving a second order cone programming relaxation.
{"title":"Robust second order cone conditions and duality for multiobjective problems under uncertainty data","authors":"Cao Thanh Tinh, Thai Doan Chuong","doi":"10.1007/s10898-023-01335-3","DOIUrl":"https://doi.org/10.1007/s10898-023-01335-3","url":null,"abstract":"<p>This paper studies a class of multiobjective convex polynomial problems, where both the constraint and objective functions involve uncertain parameters that reside in ellipsoidal uncertainty sets. Employing the robust deterministic approach, we provide necessary conditions and sufficient conditions, which are exhibited in relation to second order cone conditions, for robust (weak) Pareto solutions of the uncertain multiobjective optimization problem. A dual multiobjective problem is proposed to examine robust converse, robust weak and robust strong duality relations between the primal and dual problems. Moreover, we establish robust solution relationships between the uncertain multiobjective optimization program and a (scalar) second order cone programming relaxation problem of a corresponding weighted-sum optimization problem. This in particular shows that we can find a robust (weak) Pareto solution of the uncertain multiobjective optimization problem by solving a second order cone programming relaxation.\u0000</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"1 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139518645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-12DOI: 10.1007/s10898-023-01354-0
Sanchit Singh, Subhash C. Sarin, Ming Cheng
In this paper, we consider an application of lot-streaming for processing a lot of multiple items in a hybrid flow shop (HFS) for the objective of minimizing makespan. The HFS that we consider consists of two stages with a single machine available for processing in Stage 1 and m identical parallel machines in Stage 2. We call this problem a 1 + m TSHFS-LSP (two-stage hybrid flow shop, lot streaming problem), and show it to be NP-hard in general, except for the case when the sublot sizes are treated to be continuous. The novelty of our work is in obtaining closed-form expressions for optimal continuous sublot sizes that can be solved in polynomial time, for a given number of sublots. A fast linear search algorithm is also developed for determining the optimal number of sublots for the case of continuous sublot sizes. For the case when the sublot sizes are discrete, we propose a branch-and-bound-based heuristic to determine both the number of sublots and sublot sizes and demonstrate its efficacy by comparing its performance against that of a direct solution of a mixed-integer formulation of the problem by CPLEX®.
在本文中,我们考虑在混合流程车间(HFS)中应用批量流来处理包含多个项目的批量,以实现最小生产间隔的目标。我们考虑的 HFS 由两个阶段组成,第一阶段有一台可用于加工的机器,第二阶段有 m 台相同的并行机器。我们把这个问题称为 1 + m TSHFS-LSP(两阶段混合流程车间,批量流问题),并证明它在一般情况下是 NP-困难的,但子批量大小被视为连续的情况除外。我们工作的新颖之处在于获得了最优连续子批量大小的闭式表达式,对于给定数量的子批量,可以在多项式时间内求解。我们还开发了一种快速线性搜索算法,用于确定连续子槽尺寸情况下的最佳子槽数量。对于子槽大小离散的情况,我们提出了一种基于分支和边界的启发式方法来确定子槽数量和子槽大小,并通过比较其性能与 CPLEX® 对问题的混合整数表述的直接求解性能来证明其有效性。
{"title":"Single-lot, lot-streaming problem for a 1 + m hybrid flow shop","authors":"Sanchit Singh, Subhash C. Sarin, Ming Cheng","doi":"10.1007/s10898-023-01354-0","DOIUrl":"https://doi.org/10.1007/s10898-023-01354-0","url":null,"abstract":"<p>In this paper, we consider an application of lot-streaming for processing a lot of multiple items in a hybrid flow shop (HFS) for the objective of minimizing makespan. The HFS that we consider consists of two stages with a single machine available for processing in Stage 1 and <i>m</i> identical parallel machines in Stage 2. We call this problem a 1 + <i>m</i> TSHFS-LSP (two-stage hybrid flow shop, lot streaming problem), and show it to be NP-hard in general, except for the case when the sublot sizes are treated to be continuous. The novelty of our work is in obtaining closed-form expressions for optimal continuous sublot sizes that can be solved in polynomial time, for a given number of sublots. A fast linear search algorithm is also developed for determining the optimal number of sublots for the case of continuous sublot sizes. For the case when the sublot sizes are discrete, we propose a branch-and-bound-based heuristic to determine both the number of sublots and sublot sizes and demonstrate its efficacy by comparing its performance against that of a direct solution of a mixed-integer formulation of the problem by CPLEX<sup>®</sup>.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"2 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139458859","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-06DOI: 10.1007/s10898-023-01353-1
Xiangkai Sun, Jiayi Huang, Kok Lay Teo
In this paper, we deal with a new class of SOS-convex (sum of squares convex) polynomial optimization problems with spectrahedral uncertainty data in both the objective and constraints. By using robust optimization and a weighted-sum scalarization methodology, we first present the relationship between robust solutions of this uncertain SOS-convex polynomial optimization problem and that of its corresponding scalar optimization problem. Then, by using a normal cone constraint qualification condition, we establish necessary and sufficient optimality conditions for robust weakly efficient solutions of this uncertain SOS-convex polynomial optimization problem based on scaled diagonally dominant sums of squares conditions and linear matrix inequalities. Moreover, we introduce a semidefinite programming relaxation problem of its weighted-sum scalar optimization problem, and show that robust weakly efficient solutions of the uncertain SOS-convex polynomial optimization problem can be found by solving the corresponding semidefinite programming relaxation problem.
本文讨论了一类新的 SOS-凸(平方凸和)多项式优化问题,该问题的目标和约束条件中均包含光谱不确定性数据。通过使用鲁棒优化和加权求和标量化方法,我们首先提出了这种不确定 SOS 凸多项式优化问题的鲁棒解与其相应标量优化问题的鲁棒解之间的关系。然后,通过使用法锥约束限定条件,我们基于比例对角显性平方和条件和线性矩阵不等式,建立了该不确定 SOS 凸多项式优化问题的鲁棒弱有效解的必要和充分最优条件。此外,我们还引入了其加权和标量优化问题的半有限编程松弛问题,并证明通过求解相应的半有限编程松弛问题,可以找到不确定 SOS 凸多项式优化问题的稳健弱高效解。
{"title":"On semidefinite programming relaxations for a class of robust SOS-convex polynomial optimization problems","authors":"Xiangkai Sun, Jiayi Huang, Kok Lay Teo","doi":"10.1007/s10898-023-01353-1","DOIUrl":"https://doi.org/10.1007/s10898-023-01353-1","url":null,"abstract":"<p>In this paper, we deal with a new class of SOS-convex (sum of squares convex) polynomial optimization problems with spectrahedral uncertainty data in both the objective and constraints. By using robust optimization and a weighted-sum scalarization methodology, we first present the relationship between robust solutions of this uncertain SOS-convex polynomial optimization problem and that of its corresponding scalar optimization problem. Then, by using a normal cone constraint qualification condition, we establish necessary and sufficient optimality conditions for robust weakly efficient solutions of this uncertain SOS-convex polynomial optimization problem based on scaled diagonally dominant sums of squares conditions and linear matrix inequalities. Moreover, we introduce a semidefinite programming relaxation problem of its weighted-sum scalar optimization problem, and show that robust weakly efficient solutions of the uncertain SOS-convex polynomial optimization problem can be found by solving the corresponding semidefinite programming relaxation problem.\u0000</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"206 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139373847","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-05DOI: 10.1007/s10898-023-01349-x
Jing-jing Wang, Li-ping Tang, Xin-min Yang
In this paper, we propose a spectral projected subgradient method with a 1-memory momentum term for solving constrained convex multiobjective optimization problem. This method combines the subgradient-type algorithm for multiobjective optimization problems with the idea of the spectral projected algorithm to accelerate the convergence process. Additionally, a 1-memory momentum term is added to the subgradient direction in the early iterations. The 1-memory momentum term incorporates, in the present iteration, some of the influence of the past iterations, and this can help to improve the search direction. Under suitable assumptions, we show that the sequence generated by the method converges to a weakly Pareto efficient solution and derive the sublinear convergence rates for the proposed method. Finally, computational experiments are given to demonstrate the effectiveness of the proposed method.
{"title":"Spectral projected subgradient method with a 1-memory momentum term for constrained multiobjective optimization problem","authors":"Jing-jing Wang, Li-ping Tang, Xin-min Yang","doi":"10.1007/s10898-023-01349-x","DOIUrl":"https://doi.org/10.1007/s10898-023-01349-x","url":null,"abstract":"<p>In this paper, we propose a spectral projected subgradient method with a 1-memory momentum term for solving constrained convex multiobjective optimization problem. This method combines the subgradient-type algorithm for multiobjective optimization problems with the idea of the spectral projected algorithm to accelerate the convergence process. Additionally, a 1-memory momentum term is added to the subgradient direction in the early iterations. The 1-memory momentum term incorporates, in the present iteration, some of the influence of the past iterations, and this can help to improve the search direction. Under suitable assumptions, we show that the sequence generated by the method converges to a weakly Pareto efficient solution and derive the sublinear convergence rates for the proposed method. Finally, computational experiments are given to demonstrate the effectiveness of the proposed method.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"31 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139373808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-03DOI: 10.1007/s10898-023-01350-4
Abstract
A new exact projective penalty method is proposed for the equivalent reduction of constrained optimization problems to nonsmooth unconstrained ones. In the method, the original objective function is extended to infeasible points by summing its value at the projection of an infeasible point on the feasible set with the distance to the projection. Beside Euclidean projections, also a pointed projection in the direction of some fixed internal feasible point can be used. The equivalence means that local and global minimums of the problems coincide. Nonconvex sets with multivalued Euclidean projections are admitted, and the objective function may be lower semicontinuous. The particular case of convex problems is included. The obtained unconstrained or box constrained problem is solved by a version of the branch and bound method combined with local optimization. In principle, any local optimizer can be used within the branch and bound scheme but in numerical experiments sequential quadratic programming method was successfully used. So the proposed exact penalty method does not assume the existence of the objective function outside the allowable area and does not require the selection of the penalty coefficient.
{"title":"The exact projective penalty method for constrained optimization","authors":"","doi":"10.1007/s10898-023-01350-4","DOIUrl":"https://doi.org/10.1007/s10898-023-01350-4","url":null,"abstract":"<h3>Abstract</h3> <p>A new exact projective penalty method is proposed for the equivalent reduction of constrained optimization problems to nonsmooth unconstrained ones. In the method, the original objective function is extended to infeasible points by summing its value at the projection of an infeasible point on the feasible set with the distance to the projection. Beside Euclidean projections, also a pointed projection in the direction of some fixed internal feasible point can be used. The equivalence means that local and global minimums of the problems coincide. Nonconvex sets with multivalued Euclidean projections are admitted, and the objective function may be lower semicontinuous. The particular case of convex problems is included. The obtained unconstrained or box constrained problem is solved by a version of the branch and bound method combined with local optimization. In principle, any local optimizer can be used within the branch and bound scheme but in numerical experiments sequential quadratic programming method was successfully used. So the proposed exact penalty method does not assume the existence of the objective function outside the allowable area and does not require the selection of the penalty coefficient.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"30 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139093752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-02DOI: 10.1007/s10898-023-01336-2
Abstract
While constrained, multiobjective optimization is generally very difficult, there is a special case in which such problems can be solved with a simple, elegant branch-and-bound algorithm. This special case is when the objective and constraint functions are Lipschitz continuous with known Lipschitz constants. Given these Lipschitz constants, one can compute lower bounds on the functions over subregions of the search space. This allows one to iteratively partition the search space into rectangles, deleting those rectangles which—based on the lower bounds—contain points that are all provably infeasible or provably dominated by previously sampled point(s). As the algorithm proceeds, the rectangles that have not been deleted provide a tight covering of the Pareto set in the input space. Unfortunately, for black-box optimization this elegant algorithm cannot be applied, as we would not know the Lipschitz constants. In this paper, we show how one can heuristically extend this branch-and-bound algorithm to the case when the problem functions are black-box using an approach similar to that used in the well-known DIRECT global optimization algorithm. We call the resulting method “simDIRECT.” Initial experience with simDIRECT on test problems suggests that it performs similar to, or better than, multiobjective evolutionary algorithms when solving problems with small numbers of variables (up to 12) and a limited number of runs (up to 600).
{"title":"Constrained multiobjective optimization of expensive black-box functions using a heuristic branch-and-bound approach","authors":"","doi":"10.1007/s10898-023-01336-2","DOIUrl":"https://doi.org/10.1007/s10898-023-01336-2","url":null,"abstract":"<h3>Abstract</h3> <p>While constrained, multiobjective optimization is generally very difficult, there is a special case in which such problems can be solved with a simple, elegant branch-and-bound algorithm. This special case is when the objective and constraint functions are Lipschitz continuous with known Lipschitz constants. Given these Lipschitz constants, one can compute lower bounds on the functions over subregions of the search space. This allows one to iteratively partition the search space into rectangles, deleting those rectangles which—based on the lower bounds—contain points that are all provably infeasible or provably dominated by previously sampled point(s). As the algorithm proceeds, the rectangles that have not been deleted provide a tight covering of the Pareto set in the input space. Unfortunately, for black-box optimization this elegant algorithm cannot be applied, as we would not know the Lipschitz constants. In this paper, we show how one can heuristically extend this branch-and-bound algorithm to the case when the problem functions are black-box using an approach similar to that used in the well-known DIRECT global optimization algorithm. We call the resulting method “simDIRECT.” Initial experience with simDIRECT on test problems suggests that it performs similar to, or better than, multiobjective evolutionary algorithms when solving problems with small numbers of variables (up to 12) and a limited number of runs (up to 600). </p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"1 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139080250","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-02DOI: 10.1007/s10898-023-01352-2
Daniela Lera, Maria Chiara Nasso, Mikhail Posypkin, Yaroslav D. Sergeyev
In this paper, the problem of approximating and visualizing the solution set of systems of nonlinear inequalities is considered. It is supposed that left-hand parts of the inequalities can be multiextremal and non-differentiable. Thus, traditional local methods using gradients cannot be applied in these circumstances. Problems of this kind arise in many scientific applications, in particular, in finding working spaces of robots where it is necessary to determine not one but all the solutions of the system of nonlinear inequalities. Global optimization algorithms can be taken as an inspiration for developing methods for solving this problem. In this article, two new methods using two different approximations of Peano–Hilbert space-filling curves actively used in global optimization are proposed. Convergence conditions of the new methods are established. Numerical experiments executed on problems regarding finding the working spaces of several robots show a promising performance of the new algorithms.
{"title":"Determining solution set of nonlinear inequalities using space-filling curves for finding working spaces of planar robots","authors":"Daniela Lera, Maria Chiara Nasso, Mikhail Posypkin, Yaroslav D. Sergeyev","doi":"10.1007/s10898-023-01352-2","DOIUrl":"https://doi.org/10.1007/s10898-023-01352-2","url":null,"abstract":"<p>In this paper, the problem of approximating and visualizing the solution set of systems of nonlinear inequalities is considered. It is supposed that left-hand parts of the inequalities can be multiextremal and non-differentiable. Thus, traditional local methods using gradients cannot be applied in these circumstances. Problems of this kind arise in many scientific applications, in particular, in finding working spaces of robots where it is necessary to determine not one but <i>all</i> the solutions of the system of nonlinear inequalities. Global optimization algorithms can be taken as an inspiration for developing methods for solving this problem. In this article, two new methods using two different approximations of Peano–Hilbert space-filling curves actively used in global optimization are proposed. Convergence conditions of the new methods are established. Numerical experiments executed on problems regarding finding the working spaces of several robots show a promising performance of the new algorithms.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"29 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139077175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-02DOI: 10.1007/s10898-023-01344-2
Abstract
We discuss two key problems related to learning and optimization of neural networks: the computation of the adversarial attack for adversarial robustness and approximate optimization of complex functions. We show that both problems can be cast as instances of DC-programming. We give an explicit decomposition of the corresponding functions as differences of convex functions (DC) and report the results of experiments demonstrating the effectiveness of the DCA algorithm applied to these problems.
摘要 我们讨论了与神经网络学习和优化相关的两个关键问题:计算对抗鲁棒性的对抗攻击和复杂函数的近似优化。我们证明,这两个问题都可以作为 DC 编程的实例。我们给出了相应函数作为凸函数差分 (DC) 的明确分解,并报告了实验结果,证明了 DCA 算法应用于这些问题的有效性。
{"title":"DC-programming for neural network optimizations","authors":"","doi":"10.1007/s10898-023-01344-2","DOIUrl":"https://doi.org/10.1007/s10898-023-01344-2","url":null,"abstract":"<h3>Abstract</h3> <p>We discuss two key problems related to learning and optimization of neural networks: the computation of the adversarial attack for adversarial robustness and approximate optimization of complex functions. We show that both problems can be cast as instances of DC-programming. We give an explicit decomposition of the corresponding functions as differences of convex functions (DC) and report the results of experiments demonstrating the effectiveness of the DCA algorithm applied to these problems. </p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"25 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2024-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139080275","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}