Pub Date : 2024-08-06DOI: 10.1007/s10957-024-02498-z
Yuanhang Liu, Weijia Wu, Donghui Yang
This paper focuses on investigating the optimal actuator location for achieving minimum norm controls in the context of approximate controllability for degenerate parabolic equations. We propose a formulation of the optimization problem that encompasses both the actuator location and its associated minimum norm control. Specifically, we transform the problem into a two-person zero-sum game problem, resulting in the development of four equivalent formulations. Finally, we establish the crucial result that the solution to the relaxed optimization problem serves as an optimal actuator location for the classical problem.
{"title":"Optimal Actuator Location of the Norm Optimal Controls for Degenerate Parabolic Equations","authors":"Yuanhang Liu, Weijia Wu, Donghui Yang","doi":"10.1007/s10957-024-02498-z","DOIUrl":"https://doi.org/10.1007/s10957-024-02498-z","url":null,"abstract":"<p>This paper focuses on investigating the optimal actuator location for achieving minimum norm controls in the context of approximate controllability for degenerate parabolic equations. We propose a formulation of the optimization problem that encompasses both the actuator location and its associated minimum norm control. Specifically, we transform the problem into a two-person zero-sum game problem, resulting in the development of four equivalent formulations. Finally, we establish the crucial result that the solution to the relaxed optimization problem serves as an optimal actuator location for the classical problem.</p>","PeriodicalId":50100,"journal":{"name":"Journal of Optimization Theory and Applications","volume":"160 1","pages":""},"PeriodicalIF":1.9,"publicationDate":"2024-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141935096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-05DOI: 10.1007/s10957-024-02504-4
Daylen K. Thimm
Consider three closed linear subspaces (C_1, C_2,) and (C_3) of a Hilbert space H and the orthogonal projections (P_1, P_2) and (P_3) onto them. Halperin showed that a point in (C_1cap C_2 cap C_3) can be found by iteratively projecting any point (x_0 in H) onto all the sets in a periodic fashion. The limit point is then the projection of (x_0) onto (C_1cap C_2 cap C_3). Nevertheless, a non-periodic projection order may lead to a non-convergent projection series, as shown by Kopecká, Müller, and Paszkiewicz. This raises the question how many projection orders in ({1,2,3}^{mathbb {N}}) are “well behaved” in the sense that they lead to a convergent projection series. Melo, da Cruz Neto, and de Brito provided a necessary and sufficient condition under which the projection series converges and showed that the “well behaved” projection orders form a large subset in the sense of having full product measure. We show that also from a topological viewpoint the set of “well behaved” projection orders is a large subset: it contains a dense (G_delta ) subset with respect to the product topology. Furthermore, we analyze why the proof of the measure theoretic case cannot be directly adapted to the topological setting.
{"title":"Most Iterations of Projections Converge","authors":"Daylen K. Thimm","doi":"10.1007/s10957-024-02504-4","DOIUrl":"https://doi.org/10.1007/s10957-024-02504-4","url":null,"abstract":"<p>Consider three closed linear subspaces <span>(C_1, C_2,)</span> and <span>(C_3)</span> of a Hilbert space <i>H</i> and the orthogonal projections <span>(P_1, P_2)</span> and <span>(P_3)</span> onto them. Halperin showed that a point in <span>(C_1cap C_2 cap C_3)</span> can be found by iteratively projecting any point <span>(x_0 in H)</span> onto all the sets in a periodic fashion. The limit point is then the projection of <span>(x_0)</span> onto <span>(C_1cap C_2 cap C_3)</span>. Nevertheless, a non-periodic projection order may lead to a non-convergent projection series, as shown by Kopecká, Müller, and Paszkiewicz. This raises the question how many projection orders in <span>({1,2,3}^{mathbb {N}})</span> are “well behaved” in the sense that they lead to a convergent projection series. Melo, da Cruz Neto, and de Brito provided a necessary and sufficient condition under which the projection series converges and showed that the “well behaved” projection orders form a large subset in the sense of having full product measure. We show that also from a topological viewpoint the set of “well behaved” projection orders is a large subset: it contains a dense <span>(G_delta )</span> subset with respect to the product topology. Furthermore, we analyze why the proof of the measure theoretic case cannot be directly adapted to the topological setting.</p>","PeriodicalId":50100,"journal":{"name":"Journal of Optimization Theory and Applications","volume":"9 1","pages":""},"PeriodicalIF":1.9,"publicationDate":"2024-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141935102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-04DOI: 10.1007/s10957-024-02491-6
Eric Luxenberg, Dhruv Malik, Yuanzhi Li, Aarti Singh, Stephen Boyd
We consider robust empirical risk minimization (ERM), where model parameters are chosen to minimize the worst-case empirical loss when each data point varies over a given convex uncertainty set. In some simple cases, such problems can be expressed in an analytical form. In general the problem can be made tractable via dualization, which turns a min-max problem into a min-min problem. Dualization requires expertise and is tedious and error-prone. We demonstrate how CVXPY can be used to automate this dualization procedure in a user-friendly manner. Our framework allows practitioners to specify and solve robust ERM problems with a general class of convex losses, capturing many standard regression and classification problems. Users can easily specify any complex uncertainty set that is representable via disciplined convex programming (DCP) constraints.
{"title":"Specifying and Solving Robust Empirical Risk Minimization Problems Using CVXPY","authors":"Eric Luxenberg, Dhruv Malik, Yuanzhi Li, Aarti Singh, Stephen Boyd","doi":"10.1007/s10957-024-02491-6","DOIUrl":"https://doi.org/10.1007/s10957-024-02491-6","url":null,"abstract":"<p>We consider robust empirical risk minimization (ERM), where model parameters are chosen to minimize the worst-case empirical loss when each data point varies over a given convex uncertainty set. In some simple cases, such problems can be expressed in an analytical form. In general the problem can be made tractable via dualization, which turns a min-max problem into a min-min problem. Dualization requires expertise and is tedious and error-prone. We demonstrate how CVXPY can be used to automate this dualization procedure in a user-friendly manner. Our framework allows practitioners to specify and solve robust ERM problems with a general class of convex losses, capturing many standard regression and classification problems. Users can easily specify any complex uncertainty set that is representable via disciplined convex programming (DCP) constraints.</p>","PeriodicalId":50100,"journal":{"name":"Journal of Optimization Theory and Applications","volume":"9 1","pages":""},"PeriodicalIF":1.9,"publicationDate":"2024-08-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141935014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-02DOI: 10.1007/s10957-024-02501-7
Yu Cao, Yuanheng Wang, Habib ur Rehman, Yekini Shehu, Jen-Chih Yao
In this paper, we propose a new splitting algorithm to find the zero of a monotone inclusion problem that features the sum of three maximal monotone operators and a Lipschitz continuous monotone operator in Hilbert spaces. We prove that the sequence of iterates generated by our proposed splitting algorithm converges weakly to the zero of the considered inclusion problem under mild conditions on the iterative parameters. Several splitting algorithms in the literature are recovered as special cases of our proposed algorithm. Another interesting feature of our algorithm is that one forward evaluation of the Lipschitz continuous monotone operator is utilized at each iteration. Numerical results are given to support the theoretical analysis.
{"title":"Convergence Analysis of a New Forward-Reflected-Backward Algorithm for Four Operators Without Cocoercivity","authors":"Yu Cao, Yuanheng Wang, Habib ur Rehman, Yekini Shehu, Jen-Chih Yao","doi":"10.1007/s10957-024-02501-7","DOIUrl":"https://doi.org/10.1007/s10957-024-02501-7","url":null,"abstract":"<p>In this paper, we propose a new splitting algorithm to find the zero of a monotone inclusion problem that features the sum of three maximal monotone operators and a Lipschitz continuous monotone operator in Hilbert spaces. We prove that the sequence of iterates generated by our proposed splitting algorithm converges weakly to the zero of the considered inclusion problem under mild conditions on the iterative parameters. Several splitting algorithms in the literature are recovered as special cases of our proposed algorithm. Another interesting feature of our algorithm is that one forward evaluation of the Lipschitz continuous monotone operator is utilized at each iteration. Numerical results are given to support the theoretical analysis.</p>","PeriodicalId":50100,"journal":{"name":"Journal of Optimization Theory and Applications","volume":"26 1","pages":""},"PeriodicalIF":1.9,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141882507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-02DOI: 10.1007/s10957-024-02502-6
Cheik Traoré, Vassilis Apidopoulos, Saverio Salzo, Silvia Villa
In the context of finite sums minimization, variance reduction techniques are widely used to improve the performance of state-of-the-art stochastic gradient methods. Their practical impact is clear, as well as their theoretical properties. Stochastic proximal point algorithms have been studied as an alternative to stochastic gradient algorithms since they are more stable with respect to the choice of the step size. However, their variance-reduced versions are not as well studied as the gradient ones. In this work, we propose the first unified study of variance reduction techniques for stochastic proximal point algorithms. We introduce a generic stochastic proximal-based algorithm that can be specified to give the proximal version of SVRG, SAGA, and some of their variants. For this algorithm, in the smooth setting, we provide several convergence rates for the iterates and the objective function values, which are faster than those of the vanilla stochastic proximal point algorithm. More specifically, for convex functions, we prove a sublinear convergence rate of O(1/k). In addition, under the Polyak-łojasiewicz condition, we obtain linear convergence rates. Finally, our numerical experiments demonstrate the advantages of the proximal variance reduction methods over their gradient counterparts in terms of the stability with respect to the choice of the step size in most cases, especially for difficult problems.
{"title":"Variance Reduction Techniques for Stochastic Proximal Point Algorithms","authors":"Cheik Traoré, Vassilis Apidopoulos, Saverio Salzo, Silvia Villa","doi":"10.1007/s10957-024-02502-6","DOIUrl":"https://doi.org/10.1007/s10957-024-02502-6","url":null,"abstract":"<p>In the context of finite sums minimization, variance reduction techniques are widely used to improve the performance of state-of-the-art stochastic gradient methods. Their practical impact is clear, as well as their theoretical properties. Stochastic proximal point algorithms have been studied as an alternative to stochastic gradient algorithms since they are more stable with respect to the choice of the step size. However, their variance-reduced versions are not as well studied as the gradient ones. In this work, we propose the first unified study of variance reduction techniques for stochastic proximal point algorithms. We introduce a generic stochastic proximal-based algorithm that can be specified to give the proximal version of SVRG, SAGA, and some of their variants. For this algorithm, in the smooth setting, we provide several convergence rates for the iterates and the objective function values, which are faster than those of the vanilla stochastic proximal point algorithm. More specifically, for convex functions, we prove a sublinear convergence rate of <i>O</i>(1/<i>k</i>). In addition, under the Polyak-łojasiewicz condition, we obtain linear convergence rates. Finally, our numerical experiments demonstrate the advantages of the proximal variance reduction methods over their gradient counterparts in terms of the stability with respect to the choice of the step size in most cases, especially for difficult problems.</p>","PeriodicalId":50100,"journal":{"name":"Journal of Optimization Theory and Applications","volume":"5 1","pages":""},"PeriodicalIF":1.9,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141882639","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-02DOI: 10.1007/s10957-024-02506-2
Atsushi Hori, Daisuke Tsuyuguchi, Ellen H. Fukuda
The multi-leader–multi-follower game (MLMFG) involves two or more leaders and followers and serves as a generalization of the Stackelberg game and the single-leader–multi-follower game. Although MLMFG covers wide range of real-world applications, its research is still sparse. Notably, fundamental solution methods for this class of problems remain insufficiently established. A prevailing approach is to recast the MLMFG as an equilibrium problem with equilibrium constraints (EPEC) and solve it using a solver. Meanwhile, interpreting the solution to the EPEC in the context of MLMFG may be complex due to shared decision variables among all leaders, followers’ strategies that each leader can unilaterally change, but the variables are essentially controlled by followers. To address this issue, we introduce a response function of followers’ noncooperative game that is a function with leaders’ strategies as a variable. Employing this approach allows the MLMFG to be solved as a single-level differentiable variational inequality using a smoothing scheme for the followers’ response function. We also demonstrate that the sequence of solutions to the smoothed variational inequality converges to a stationary equilibrium of the MLMFG. Finally, we illustrate the behavior of the smoothing method by numerical experiments.
{"title":"A Method for Multi-Leader–Multi-Follower Games by Smoothing the Followers’ Response Function","authors":"Atsushi Hori, Daisuke Tsuyuguchi, Ellen H. Fukuda","doi":"10.1007/s10957-024-02506-2","DOIUrl":"https://doi.org/10.1007/s10957-024-02506-2","url":null,"abstract":"<p>The multi-leader–multi-follower game (MLMFG) involves two or more leaders and followers and serves as a generalization of the Stackelberg game and the single-leader–multi-follower game. Although MLMFG covers wide range of real-world applications, its research is still sparse. Notably, fundamental solution methods for this class of problems remain insufficiently established. A prevailing approach is to recast the MLMFG as an equilibrium problem with equilibrium constraints (EPEC) and solve it using a solver. Meanwhile, interpreting the solution to the EPEC in the context of MLMFG may be complex due to shared decision variables among all leaders, followers’ strategies that each leader can unilaterally change, but the variables are essentially controlled by followers. To address this issue, we introduce a response function of followers’ noncooperative game that is a function with leaders’ strategies as a variable. Employing this approach allows the MLMFG to be solved as a single-level differentiable variational inequality using a smoothing scheme for the followers’ response function. We also demonstrate that the sequence of solutions to the smoothed variational inequality converges to a stationary equilibrium of the MLMFG. Finally, we illustrate the behavior of the smoothing method by numerical experiments.</p>","PeriodicalId":50100,"journal":{"name":"Journal of Optimization Theory and Applications","volume":"22 1","pages":""},"PeriodicalIF":1.9,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141882506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01DOI: 10.1007/s10957-024-02499-y
Alexey S. Matveev, Dmitrii V. Sugak
This article is concerned with optimal control problems for plants described by systems of high order nonlinear PDE’s (whose linear approximation is elliptic in the sense of Douglis-Nirenberg), with a special attention being given to their particular case: the standard stationary system of non-linear Navier–Stokes equations. The objective is to establish an analog of the Pontryagin’s maximum principle. The major challenge stems from the presence of infinitely many point-wise constraints on the system’s state, which are imposed at any point from a given continuum set of independent variables. Necessary conditions for optimality in the form of an “abstract” maximum principle are first obtained for a general optimal control problem couched in the language of functional analysis. This result is targeted at a wide class of problems, with an idea to absorb, in its proof, a great deal of technical work needed for derivation of optimality conditions so that only an interpretation of the discussed result would be basically needed to handle a particular problem. The applicability of this approach is demonstrated via obtaining the afore-mentioned analog of the Pontryagin’s maximum principle for a state-constrained system of high-order elliptic equations and the Navier–Stokes equations.
{"title":"Pontryagin’s Maximum Principle for a State-Constrained System of Douglis-Nirenberg Type","authors":"Alexey S. Matveev, Dmitrii V. Sugak","doi":"10.1007/s10957-024-02499-y","DOIUrl":"https://doi.org/10.1007/s10957-024-02499-y","url":null,"abstract":"<p>This article is concerned with optimal control problems for plants described by systems of high order nonlinear PDE’s (whose linear approximation is elliptic in the sense of Douglis-Nirenberg), with a special attention being given to their particular case: the standard stationary system of non-linear Navier–Stokes equations. The objective is to establish an analog of the Pontryagin’s maximum principle. The major challenge stems from the presence of infinitely many point-wise constraints on the system’s state, which are imposed at any point from a given continuum set of independent variables. Necessary conditions for optimality in the form of an “abstract” maximum principle are first obtained for a general optimal control problem couched in the language of functional analysis. This result is targeted at a wide class of problems, with an idea to absorb, in its proof, a great deal of technical work needed for derivation of optimality conditions so that only an interpretation of the discussed result would be basically needed to handle a particular problem. The applicability of this approach is demonstrated via obtaining the afore-mentioned analog of the Pontryagin’s maximum principle for a state-constrained system of high-order elliptic equations and the Navier–Stokes equations.\u0000</p>","PeriodicalId":50100,"journal":{"name":"Journal of Optimization Theory and Applications","volume":"4 1","pages":""},"PeriodicalIF":1.9,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141882505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-30DOI: 10.1007/s10957-024-02497-0
Jie Jiang
In this paper, we use the distributionally robust approach to study stochastic variational inequalities under the ambiguity of the true probability distribution, which is referred to as distributionally robust variational inequalities (DRVIs). First of all, we adopt a relaxed value function approach to relax the DRVI and obtain its relaxation counterpart. This is mainly motivated by the robust requirement in the modeling process as well as the possible calculation error in the numerical process. After that, we investigate qualitative convergence properties as the relaxation parameter tends to zero. Considering the perturbation of ambiguity sets, we further study the quantitative stability of the relaxation DRVI. Finally, when the ambiguity set is given by the general moment information, the discrete approximation of the relaxation DRVI is discussed.
{"title":"Distributionally Robust Variational Inequalities: Relaxation, Quantification and Discretization","authors":"Jie Jiang","doi":"10.1007/s10957-024-02497-0","DOIUrl":"https://doi.org/10.1007/s10957-024-02497-0","url":null,"abstract":"<p>In this paper, we use the distributionally robust approach to study stochastic variational inequalities under the ambiguity of the true probability distribution, which is referred to as distributionally robust variational inequalities (DRVIs). First of all, we adopt a relaxed value function approach to relax the DRVI and obtain its relaxation counterpart. This is mainly motivated by the robust requirement in the modeling process as well as the possible calculation error in the numerical process. After that, we investigate qualitative convergence properties as the relaxation parameter tends to zero. Considering the perturbation of ambiguity sets, we further study the quantitative stability of the relaxation DRVI. Finally, when the ambiguity set is given by the general moment information, the discrete approximation of the relaxation DRVI is discussed.</p>","PeriodicalId":50100,"journal":{"name":"Journal of Optimization Theory and Applications","volume":"361 1","pages":""},"PeriodicalIF":1.9,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141868262","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-30DOI: 10.1007/s10957-024-02490-7
Meir Pachter, Isaac E. Weintraub
In this paper we advocate for Isaacs’ method for the solution of differential games to be applied to the solution of optimal control problems. To make the argument, the vehicle employed is Pontryagin’s canonical optimal control example, which entails a double integrator plant. However, rather than controlling the state to the origin, we require the end state to reach a terminal set that contains the origin in its interior. Indeed, in practice, it is required to control to a prescribed tolerance rather than reach a desired end state; constraining the end state to a terminal manifold of co-dimension n − 1 renders the optimal control problem easier to solve. The global solution of the optimal control problem is obtained and the synthesized optimal control law is in state feedback form. In this respect, two target sets are considered: a smooth circular target and a square target with corners. Closed-loop state-feedback control laws are synthesized that drive the double integrator plant from an arbitrary initial state to the target set in minimum time. This is accomplished using Isaacs’ method for the solution of differential games, which entails dynamic programming (DP), working backward from the usable part of the target set, as opposed to obtaining the optimal trajectories using the necessary conditions for optimality provided by Pontryagin’s Maximum Principle (PMP). In this paper, the case is made for Isaacs’ method for the solution of differential games to be applied to the solution of optimal control problems by way of the juxtaposition of the PMP and DP methods.
在本文中,我们主张将艾萨克斯的微分博弈求解方法应用于最优控制问题的求解。为了进行论证,我们采用了庞特里亚金的典型最优控制范例,该范例涉及一个双积分器工厂。不过,我们要求的不是将状态控制到原点,而是最终状态到达一个内部包含原点的终端集。事实上,在实际应用中,我们需要将状态控制在规定的公差范围内,而不是达到一个理想的终点状态;将终点状态限制在一个共维为 n - 1 的终端流形上,会使最优控制问题更容易求解。最优控制问题的全局解已经得到,合成的最优控制法则是状态反馈形式的。在这方面,考虑了两个目标集:光滑的圆形目标和带角的方形目标。合成的闭环状态反馈控制法则能在最短时间内驱动双积分器工厂从任意初始状态到达目标集。这需要使用艾萨克斯的微分博弈求解方法,即动态编程(DP),从目标集的可用部分向后求解,而不是使用庞特里亚金最大原则(PMP)提供的最优化必要条件来获得最佳轨迹。本文通过将 PMP 和 DP 方法并列的方式,证明艾萨克斯的微分博弈求解方法可应用于最优控制问题的求解。
{"title":"The Synthesis of Optimal Control Laws Using Isaacs’ Method for the Solution of Differential Games","authors":"Meir Pachter, Isaac E. Weintraub","doi":"10.1007/s10957-024-02490-7","DOIUrl":"https://doi.org/10.1007/s10957-024-02490-7","url":null,"abstract":"<p>In this paper we advocate for Isaacs’ method for the solution of differential games to be applied to the solution of optimal control problems. To make the argument, the vehicle employed is Pontryagin’s canonical optimal control example, which entails a double integrator plant. However, rather than controlling the state to the origin, we require the end state to reach a terminal set that contains the origin in its interior. Indeed, in practice, it is required to control to a prescribed tolerance rather than reach a desired end state; constraining the end state to a terminal manifold of co-dimension n − 1 renders the optimal control problem easier to solve. The global solution of the optimal control problem is obtained and the synthesized optimal control law is in state feedback form. In this respect, two target sets are considered: a smooth circular target and a square target with corners. Closed-loop state-feedback control laws are synthesized that drive the double integrator plant from an arbitrary initial state to the target set in minimum time. This is accomplished using Isaacs’ method for the solution of differential games, which entails dynamic programming (DP), working backward from the usable part of the target set, as opposed to obtaining the optimal trajectories using the necessary conditions for optimality provided by Pontryagin’s Maximum Principle (PMP). In this paper, the case is made for Isaacs’ method for the solution of differential games to be applied to the solution of optimal control problems by way of the juxtaposition of the PMP and DP methods.\u0000</p>","PeriodicalId":50100,"journal":{"name":"Journal of Optimization Theory and Applications","volume":"212 1","pages":""},"PeriodicalIF":1.9,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141868140","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In this paper we study a system of decoupled forward-backward stochastic differential equations driven by a G-Brownian motion (G-FBSDEs) with non-degenerate diffusion. Our objective is to establish the existence of a relaxed optimal control for a non-smooth stochastic optimal control problem. The latter is given in terms of a decoupled G-FBSDE. The cost functional is the solution of the backward stochastic differential equation at the initial time. The key idea to establish existence of a relaxed optimal control is to replace the original control problem by a suitably regularised problem with mollified coefficients, prove the existence of a relaxed control, and then pass to the limit.
在本文中,我们研究了一个由 G 布朗运动(G-FBSDE)驱动的非退化扩散的解耦前向后向随机微分方程系统。我们的目标是为一个非光滑随机最优控制问题建立一个松弛最优控制。后者以解耦 G-FBSDE 的形式给出。成本函数是初始时间后向随机微分方程的解。建立松弛最优控制存在性的关键思路是将原始控制问题替换为一个适当正则化的问题,该问题的系数被修正,证明松弛控制的存在性,然后求极限。
{"title":"Optimal Relaxed Control for a Decoupled G-FBSDE","authors":"Hafida Bouanani, Omar Kebiri, Carsten Hartmann, Amel Redjil","doi":"10.1007/s10957-024-02495-2","DOIUrl":"https://doi.org/10.1007/s10957-024-02495-2","url":null,"abstract":"<p>In this paper we study a system of decoupled forward-backward stochastic differential equations driven by a <i>G</i>-Brownian motion (<i>G</i>-FBSDEs) with non-degenerate diffusion. Our objective is to establish the existence of a relaxed optimal control for a non-smooth stochastic optimal control problem. The latter is given in terms of a decoupled <i>G</i>-FBSDE. The cost functional is the solution of the backward stochastic differential equation at the initial time. The key idea to establish existence of a relaxed optimal control is to replace the original control problem by a suitably regularised problem with mollified coefficients, prove the existence of a relaxed control, and then pass to the limit.\u0000</p>","PeriodicalId":50100,"journal":{"name":"Journal of Optimization Theory and Applications","volume":"1 1","pages":""},"PeriodicalIF":1.9,"publicationDate":"2024-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141771832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}