Pub Date : 2024-09-07DOI: 10.1016/j.automatica.2024.111893
To investigate objects driven by external input and players’ interests, the game-based control system (GBCS) was established. In this system, the high-level leader does not participate directly but regulates the low-level game to make followers’ Nash equilibrium(NE) “better”. This article focuses on a specific type of GBCSs with rational players, where the open-loop NE is unique under any given initial state and macro-regulation. We discuss two kinds of regulations on NEs: achieving reachability among NEs through macro-regulation and Pareto improvability on NEs that can benefit at least one follower without harming anyone else. These regulations help reduce the widespread inconsistency between individual and collective rationality. Moreover, conditions are provided to determine controllability and Pareto improvability on NEs. Finally, an example on opinion dynamics is presented to demonstrate the effectiveness of obtained theoretical results.
为了研究由外部输入和参与者利益驱动的对象,建立了基于游戏的控制系统(GBCS)。在这个系统中,高层领导者并不直接参与,而是调节低层博弈,使追随者的纳什均衡(NE)"更好"。本文重点研究一种特定类型的理性博弈者的 GBCS,在任何给定的初始状态和宏观调控下,其开环 NE 都是唯一的。我们讨论了对 NE 的两种规定:通过宏观调控实现 NE 之间的可达性,以及对 NE 的帕累托改进性,即在不损害其他任何人的情况下,使至少一个追随者受益。这些规定有助于减少个人理性与集体理性之间普遍存在的不一致性。此外,我们还提供了一些条件来确定网络的可控性和帕累托改进性。最后,介绍了一个关于舆论动态的例子,以证明所获得的理论结果的有效性。
{"title":"Controllability and Pareto improvability on Nash equilibriums in game-based control systems","authors":"","doi":"10.1016/j.automatica.2024.111893","DOIUrl":"10.1016/j.automatica.2024.111893","url":null,"abstract":"<div><p>To investigate objects driven by external input and players’ interests, the game-based control system (GBCS) was established. In this system, the high-level leader does not participate directly but regulates the low-level game to make followers’ Nash equilibrium(NE) “better”. This article focuses on a specific type of GBCSs with rational players, where the open-loop NE is unique under any given initial state and macro-regulation. We discuss two kinds of regulations on NEs: achieving reachability among NEs through macro-regulation and Pareto improvability on NEs that can benefit at least one follower without harming anyone else. These regulations help reduce the widespread inconsistency between individual and collective rationality. Moreover, conditions are provided to determine controllability and Pareto improvability on NEs. Finally, an example on opinion dynamics is presented to demonstrate the effectiveness of obtained theoretical results.</p></div>","PeriodicalId":55413,"journal":{"name":"Automatica","volume":null,"pages":null},"PeriodicalIF":4.8,"publicationDate":"2024-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S000510982400387X/pdfft?md5=cef51a4d64d7ae2b1e84f254902d78bc&pid=1-s2.0-S000510982400387X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142162501","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-07DOI: 10.1016/j.automatica.2024.111881
Optimal control strategies are studied through the application of the Pontryagin’s Maximum Principle for a class of non-linear differential systems that are commonly used to describe resource allocation during bacterial growth. The approach is inspired by the optimality of numerous regulatory mechanisms in bacterial cells. In this context, we aim to predict natural feedback loops as optimal control solutions so as to gain insight on the behavior of microorganisms from a control-theoretical perspective. The problem is posed in terms of a control function representing the fraction of the cell dedicated to protein synthesis, and additional controls modeling the fraction of the cell responsible for the consumption of the available nutrient sources in the medium. By studying the necessary conditions for optimality, it is possible to prove that the solutions follow a bang–singular–bang structure, and that they are characterized by a sequential uptake pattern known as diauxic growth, which prioritizes the consumption of richer substrates over poor nutrients. Numerical simulations obtained through an optimal control solver confirm the theoretical results. Finally, we provide an application to batch cultivation of E. coli growing on glucose and lactose. For that, we propose a state feedback law that is based on the optimal control, and we calibrate the obtained closed-loop model to experimental data.
通过对一类常用于描述细菌生长过程中资源分配的非线性微分系统应用庞特里亚金最大原则,研究了最优控制策略。这种方法的灵感来自细菌细胞中众多调节机制的最优性。在这种情况下,我们的目标是将自然反馈回路预测为最优控制方案,从而从控制理论的角度深入了解微生物的行为。该问题由一个控制函数 u0(t)和 n 个附加控制函数 uui(t)构成,前者代表细胞中专门用于蛋白质合成的部分,后者则代表细胞中负责消耗培养基中可用营养源的部分。通过研究最优化的必要条件,可以证明解遵循 "砰-砰-砰 "结构,其特征是一种被称为 "双氧生长 "的顺序吸收模式,即优先消耗较丰富的基质而不是较贫乏的营养物质。通过最优控制求解器获得的数值模拟证实了理论结果。最后,我们提供了在葡萄糖和乳糖上批量培养大肠杆菌的应用。为此,我们提出了基于最优控制的状态反馈法,并根据实验数据对所获得的闭环模型进行了校准。
{"title":"Optimal control strategies in a generic class of bacterial growth models with multiple substrates","authors":"","doi":"10.1016/j.automatica.2024.111881","DOIUrl":"10.1016/j.automatica.2024.111881","url":null,"abstract":"<div><p>Optimal control strategies are studied through the application of the Pontryagin’s Maximum Principle for a class of non-linear differential systems that are commonly used to describe resource allocation during bacterial growth. The approach is inspired by the optimality of numerous regulatory mechanisms in bacterial cells. In this context, we aim to predict natural feedback loops as optimal control solutions so as to gain insight on the behavior of microorganisms from a control-theoretical perspective. The problem is posed in terms of a control function <span><math><mrow><msub><mrow><mi>u</mi></mrow><mrow><mn>0</mn></mrow></msub><mrow><mo>(</mo><mi>t</mi><mo>)</mo></mrow></mrow></math></span> representing the fraction of the cell dedicated to protein synthesis, and <span><math><mi>n</mi></math></span> additional controls <span><math><mrow><msub><mrow><mi>u</mi></mrow><mrow><mi>i</mi></mrow></msub><mrow><mo>(</mo><mi>t</mi><mo>)</mo></mrow></mrow></math></span> modeling the fraction of the cell responsible for the consumption of the available nutrient sources in the medium. By studying the necessary conditions for optimality, it is possible to prove that the solutions follow a bang–singular–bang structure, and that they are characterized by a sequential uptake pattern known as diauxic growth, which prioritizes the consumption of richer substrates over poor nutrients. Numerical simulations obtained through an optimal control solver confirm the theoretical results. Finally, we provide an application to batch cultivation of E. coli growing on glucose and lactose. For that, we propose a state feedback law that is based on the optimal control, and we calibrate the obtained closed-loop model to experimental data.</p></div>","PeriodicalId":55413,"journal":{"name":"Automatica","volume":null,"pages":null},"PeriodicalIF":4.8,"publicationDate":"2024-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0005109824003753/pdfft?md5=4b5a18dc60bfd905fc283f6a25f01d94&pid=1-s2.0-S0005109824003753-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142162577","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-07DOI: 10.1016/j.automatica.2024.111897
Recent years have witnessed a booming interest in data-driven control of dynamical systems. However, the implicit data-driven output predictors are vulnerable to uncertainty such as process disturbance and measurement noise, causing unreliable predictions and unexpected control actions. In this brief, we put forward a new data-driven approach to output prediction of stochastic linear time-invariant (LTI) systems. By utilizing the innovation form, the uncertainty in stochastic LTI systems is recast as innovations that can be readily estimated from input–output data without knowing system matrices. In this way, by applying the fundamental lemma to the innovation form, we propose a new innovation-based data-driven output predictor (OP) of stochastic LTI systems, which bypasses the need for identifying state–space matrices explicitly and building a state estimator. The boundedness of the second moment of prediction errors in closed-loop is established under mild conditions. The proposed data-driven OP can be integrated into optimal control design for better performance. Numerical simulations demonstrate the outperformance of the proposed innovation-based methods in output prediction and control design over existing formulations.
近年来,人们对数据驱动的动态系统控制产生了浓厚的兴趣。然而,隐式数据驱动输出预测器容易受到过程干扰和测量噪声等不确定性的影响,导致预测不可靠和控制行动出乎意料。在本简介中,我们提出了一种新的数据驱动方法,用于随机线性时变(LTI)系统的输出预测。通过利用创新形式,随机 LTI 系统中的不确定性被重塑为创新,无需了解系统矩阵,即可从输入输出数据中轻松估算出创新。这样,通过将基本定理应用于创新形式,我们提出了一种新的基于创新的数据驱动的随机 LTI 系统输出预测器(OP),它绕过了明确识别状态空间矩阵和建立状态估计器的需要。在温和条件下,闭环预测误差第二矩的有界性得以确立。所提出的数据驱动 OP 可以集成到优化控制设计中,以获得更好的性能。数值模拟证明了所提出的基于创新的方法在输出预测和控制设计方面的性能优于现有公式。
{"title":"Data-driven output prediction and control of stochastic systems: An innovation-based approach","authors":"","doi":"10.1016/j.automatica.2024.111897","DOIUrl":"10.1016/j.automatica.2024.111897","url":null,"abstract":"<div><p>Recent years have witnessed a booming interest in data-driven control of dynamical systems. However, the implicit data-driven output predictors are vulnerable to uncertainty such as process disturbance and measurement noise, causing unreliable predictions and unexpected control actions. In this brief, we put forward a new data-driven approach to output prediction of stochastic linear time-invariant (LTI) systems. By utilizing the innovation form, the uncertainty in stochastic LTI systems is recast as innovations that can be readily estimated from input–output data without knowing system matrices. In this way, by applying the fundamental lemma to the innovation form, we propose a new innovation-based data-driven output predictor (OP) of stochastic LTI systems, which bypasses the need for identifying state–space matrices explicitly and building a state estimator. The boundedness of the second moment of prediction errors in closed-loop is established under mild conditions. The proposed data-driven OP can be integrated into optimal control design for better performance. Numerical simulations demonstrate the outperformance of the proposed innovation-based methods in output prediction and control design over existing formulations.</p></div>","PeriodicalId":55413,"journal":{"name":"Automatica","volume":null,"pages":null},"PeriodicalIF":4.8,"publicationDate":"2024-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0005109824003911/pdfft?md5=d01ded28d040db1c663391881e4966c0&pid=1-s2.0-S0005109824003911-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142162503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-07DOI: 10.1016/j.automatica.2024.111894
The problem of online change point detection is to detect abrupt changes in properties of time series, ideally as soon as possible after those changes occur. Existing work on online change point detection either assumes i.i.d. data, focuses on asymptotic analysis, does not present theoretical guarantees on the trade-off between detection accuracy and detection delay, or is only suitable for detecting single change points. In this work, we study the online change point detection problem for linear dynamical systems with unknown dynamics, where the data exhibits temporal correlations and the system could have multiple change points. We develop a data-dependent threshold that can be used in our test that allows one to achieve a pre-specified upper bound on the probability of making a false alarm. We further provide a finite-sample-based bound for the probability of detecting a change point. Our bound demonstrates how parameters used in our algorithm affect the detection probability and delay, and provides guidance on the minimum required time between changes to guarantee detection.
{"title":"Online change points detection for linear dynamical systems with finite sample guarantees","authors":"","doi":"10.1016/j.automatica.2024.111894","DOIUrl":"10.1016/j.automatica.2024.111894","url":null,"abstract":"<div><p>The problem of online change point detection is to detect abrupt changes in properties of time series, ideally as soon as possible after those changes occur. Existing work on online change point detection either assumes i.i.d. data, focuses on asymptotic analysis, does not present theoretical guarantees on the trade-off between detection accuracy and detection delay, or is only suitable for detecting single change points. In this work, we study the online change point detection problem for linear dynamical systems with unknown dynamics, where the data exhibits temporal correlations and the system could have multiple change points. We develop a data-dependent threshold that can be used in our test that allows one to achieve a pre-specified upper bound on the probability of making a false alarm. We further provide a finite-sample-based bound for the probability of detecting a change point. Our bound demonstrates how parameters used in our algorithm affect the detection probability and delay, and provides guidance on the minimum required time between changes to guarantee detection.</p></div>","PeriodicalId":55413,"journal":{"name":"Automatica","volume":null,"pages":null},"PeriodicalIF":4.8,"publicationDate":"2024-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0005109824003881/pdfft?md5=253252efe9e72628a7809ea795bb855e&pid=1-s2.0-S0005109824003881-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142162579","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-07DOI: 10.1016/j.automatica.2024.111882
In this paper, the distributed time-varying optimization problem is investigated for networked Lagrangian systems with parametric uncertainties. Usually, in the literature, to address some distributed control problems for nonlinear systems, a networked virtual system is constructed, and a tracking algorithm is designed such that the agents’ physical states track the virtual states. It is worth pointing out that such an idea requires the exchange of the virtual states and hence necessitates communication among the group. In addition, due to the complexities of the Lagrangian dynamics and the distributed time-varying optimization problem, there exist significant challenges. This paper proposes distributed time-varying optimization algorithms that achieve zero optimum-tracking errors for the networked Lagrangian agents without the communication requirement. The main idea behind the proposed algorithms is to construct a reference system for each agent to generate a reference velocity using absolute and relative physical state measurements with no exchange of virtual states needed, and to design adaptive controllers for Lagrangian systems such that the physical states are able to track the reference velocities and hence the optimal trajectory. The algorithms introduce mutual feedback between the reference systems and the local controllers via physical states/measurements and are amenable to implementation via local onboard sensing in a communication unfriendly environment. Specifically, first, a base algorithm is proposed to solve the distributed time-varying optimization problem for networked Lagrangian systems under switching graphs. Then, based on the base algorithm, a continuous function is introduced to approximate the signum function, forming a continuous distributed optimization algorithm and hence removing the chattering. Such a continuous algorithm is convergent with bounded ultimate optimum-tracking errors, which are proportion to approximation errors. Finally, numerical simulations are provided to illustrate the validity of the proposed algorithms.
{"title":"Distributed continuous-time time-varying optimization for networked Lagrangian systems with quadratic cost functions","authors":"","doi":"10.1016/j.automatica.2024.111882","DOIUrl":"10.1016/j.automatica.2024.111882","url":null,"abstract":"<div><p>In this paper, the distributed time-varying optimization problem is investigated for networked Lagrangian systems with parametric uncertainties. Usually, in the literature, to address some distributed control problems for nonlinear systems, a networked virtual system is constructed, and a tracking algorithm is designed such that the agents’ physical states track the virtual states. It is worth pointing out that such an idea requires the exchange of the virtual states and hence necessitates communication among the group. In addition, due to the complexities of the Lagrangian dynamics and the distributed time-varying optimization problem, there exist significant challenges. This paper proposes distributed time-varying optimization algorithms that achieve zero optimum-tracking errors for the networked Lagrangian agents without the communication requirement. The main idea behind the proposed algorithms is to construct a reference system for each agent to generate a reference velocity using absolute and relative physical state measurements with no exchange of virtual states needed, and to design adaptive controllers for Lagrangian systems such that the physical states are able to track the reference velocities and hence the optimal trajectory. The algorithms introduce mutual feedback between the reference systems and the local controllers via physical states/measurements and are amenable to implementation via local onboard sensing in a communication unfriendly environment. Specifically, first, a base algorithm is proposed to solve the distributed time-varying optimization problem for networked Lagrangian systems under switching graphs. Then, based on the base algorithm, a continuous function is introduced to approximate the signum function, forming a continuous distributed optimization algorithm and hence removing the chattering. Such a continuous algorithm is convergent with bounded ultimate optimum-tracking errors, which are proportion to approximation errors. Finally, numerical simulations are provided to illustrate the validity of the proposed algorithms.</p></div>","PeriodicalId":55413,"journal":{"name":"Automatica","volume":null,"pages":null},"PeriodicalIF":4.8,"publicationDate":"2024-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0005109824003765/pdfft?md5=268cd3ac8c9394f35de6171bde25b9be&pid=1-s2.0-S0005109824003765-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142162502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-06DOI: 10.1016/j.automatica.2024.111883
For -qubit stochastic open quantum systems, an online quantum state estimation (OQSE) algorithm and the associated online estimated-based-state feedback control (OQSE-FC) are proposed in this paper. The proposed OQSE algorithm integrates the online alternating direction multiplier method (OADM) to the continuous weak measurement (CWM). The quantum state feedback control laws are designed based on the Lyapunov stability theorem, and the states for feedback control are online estimated by OQSE algorithm. The convergence of OQSE algorithm and the asymptotic stability of the state feedback control laws are proved.
{"title":"Online estimated-based-state feedback control of n-qubit stochastic open quantum systems","authors":"","doi":"10.1016/j.automatica.2024.111883","DOIUrl":"10.1016/j.automatica.2024.111883","url":null,"abstract":"<div><p>For <span><math><mi>n</mi></math></span>-qubit stochastic open quantum systems, an online quantum state estimation (OQSE) algorithm and the associated online estimated-based-state feedback control (OQSE-FC) are proposed in this paper. The proposed OQSE algorithm integrates the online alternating direction multiplier method (OADM) to the continuous weak measurement (CWM). The quantum state feedback control laws are designed based on the Lyapunov stability theorem, and the states for feedback control are online estimated by OQSE algorithm. The convergence of OQSE algorithm and the asymptotic stability of the state feedback control laws are proved.</p></div>","PeriodicalId":55413,"journal":{"name":"Automatica","volume":null,"pages":null},"PeriodicalIF":4.8,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0005109824003777/pdfft?md5=e20a03851b75b5f56ef9ad5bc20ad096&pid=1-s2.0-S0005109824003777-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142162498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-05DOI: 10.1016/j.automatica.2024.111880
We propose a scalable, distributed algorithm for the optimal transport of large-scale multi-agent systems. We formulate the problem as one of steering the collective towards a target probability measure while minimizing the total cost of transport, with the additional constraint of distributed implementation. Using optimal transport theory, we realize the solution as an iterative transport based on a stochastic proximal descent scheme. At each stage of the transport, the agents implement an online, distributed primal–dual algorithm to obtain local estimates of the Kantorovich potential for optimal transport from the current distribution of the collective to the target distribution. Using these estimates as their local objective functions, the agents then implement the transport by stochastic proximal descent. This two-step process is carried out recursively by the agents to converge asymptotically to the target distribution. We rigorously establish the underlying theoretical framework and convergence of the algorithm and test its behavior in numerical experiments.
{"title":"Distributed online optimization for multi-agent optimal transport","authors":"","doi":"10.1016/j.automatica.2024.111880","DOIUrl":"10.1016/j.automatica.2024.111880","url":null,"abstract":"<div><p>We propose a scalable, distributed algorithm for the optimal transport of large-scale multi-agent systems. We formulate the problem as one of steering the collective towards a target probability measure while minimizing the total cost of transport, with the additional constraint of distributed implementation. Using optimal transport theory, we realize the solution as an iterative transport based on a stochastic proximal descent scheme. At each stage of the transport, the agents implement an online, distributed primal–dual algorithm to obtain local estimates of the Kantorovich potential for optimal transport from the current distribution of the collective to the target distribution. Using these estimates as their local objective functions, the agents then implement the transport by stochastic proximal descent. This two-step process is carried out recursively by the agents to converge asymptotically to the target distribution. We rigorously establish the underlying theoretical framework and convergence of the algorithm and test its behavior in numerical experiments.</p></div>","PeriodicalId":55413,"journal":{"name":"Automatica","volume":null,"pages":null},"PeriodicalIF":4.8,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0005109824003741/pdfft?md5=279afaa21d8e82d5ccefa238244d0e47&pid=1-s2.0-S0005109824003741-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142162505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-05DOI: 10.1016/j.automatica.2024.111884
This paper presents a switching control strategy as a criterion for policy selection in stochastic Dynamic Programming problems over an infinite time horizon. In particular, the Bellman operator, applied iteratively to solve such problems, is generalized to the case of stochastic policies, and formulated as a discrete-time switched affine system. Then, a Lyapunov-based policy selection strategy is designed to ensure the practical convergence of the resulting closed-loop system trajectories towards an appropriately chosen reference value function. This way, it is possible to verify how the chosen reference value function can be approached by using a stabilizing switching signal, the latter defined on a given finite set of stationary stochastic policies. Finally, the presented method is applied to the Value Iteration algorithm, and an illustrative example of a recycling robot is provided to demonstrate its effectiveness in terms of convergence performance.
{"title":"A switching control strategy for policy selection in stochastic Dynamic Programming problems","authors":"","doi":"10.1016/j.automatica.2024.111884","DOIUrl":"10.1016/j.automatica.2024.111884","url":null,"abstract":"<div><p>This paper presents a switching control strategy as a criterion for policy selection in stochastic Dynamic Programming problems over an infinite time horizon. In particular, the Bellman operator, applied iteratively to solve such problems, is generalized to the case of stochastic policies, and formulated as a discrete-time switched affine system. Then, a Lyapunov-based policy selection strategy is designed to ensure the practical convergence of the resulting closed-loop system trajectories towards an appropriately chosen reference value function. This way, it is possible to verify how the chosen reference value function can be approached by using a stabilizing switching signal, the latter defined on a given finite set of stationary stochastic policies. Finally, the presented method is applied to the Value Iteration algorithm, and an illustrative example of a recycling robot is provided to demonstrate its effectiveness in terms of convergence performance.</p></div>","PeriodicalId":55413,"journal":{"name":"Automatica","volume":null,"pages":null},"PeriodicalIF":4.8,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0005109824003789/pdfft?md5=6c3de146a3dd2eb4416dc66a045345cd&pid=1-s2.0-S0005109824003789-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142162500","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-31DOI: 10.1016/j.automatica.2024.111879
In this paper, we propose a new property of quantitative nonblockingness of an automaton with respect to a given cover on its set of marker states. This property quantifies the standard nonblocking property by capturing the practical requirement that every subset (i.e. cell) of marker states can be reached within a prescribed number of steps from any reachable state and following any trajectory of the system. Accordingly, we formulate a new problem of quantitatively nonblocking supervisory control, and characterize its solvability in terms of a new concept of quantitative language completability. It is proven that there exists the unique supremal quantitatively completable sublanguage of a given language, and we develop an effective algorithm to compute the supremal sublanguage. Finally, combining with the algorithm of computing the supremal controllable sublanguage, we design an algorithm to compute the maximally permissive solution to the formulated quantitatively nonblocking supervisory control problem.
{"title":"Quantitatively nonblocking supervisory control of discrete-event systems","authors":"","doi":"10.1016/j.automatica.2024.111879","DOIUrl":"10.1016/j.automatica.2024.111879","url":null,"abstract":"<div><p>In this paper, we propose a new property of <em>quantitative nonblockingness</em> of an automaton with respect to a given cover on its set of marker states. This property <em>quantifies</em> the standard nonblocking property by capturing the practical requirement that every subset (i.e. cell) of marker states can be reached within a prescribed number of steps from any reachable state and following any trajectory of the system. Accordingly, we formulate a new problem of quantitatively nonblocking supervisory control, and characterize its solvability in terms of a new concept of quantitative language completability. It is proven that there exists the unique supremal quantitatively completable sublanguage of a given language, and we develop an effective algorithm to compute the supremal sublanguage. Finally, combining with the algorithm of computing the supremal controllable sublanguage, we design an algorithm to compute the maximally permissive solution to the formulated quantitatively nonblocking supervisory control problem.</p></div>","PeriodicalId":55413,"journal":{"name":"Automatica","volume":null,"pages":null},"PeriodicalIF":4.8,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142095236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-31DOI: 10.1016/j.automatica.2024.111876
This study focuses on the problem of optimal mismatched disturbance rejection control for uncontrollable linear discrete-time systems. In contrast to previous studies, by introducing a quadratic performance index such that the regulated state can track a reference trajectory and minimize the effects of disturbances, mismatched disturbance rejection control is transformed into a linear quadratic tracking problem. The necessary and sufficient conditions for the solvability of this problem over a finite horizon and a disturbance rejection controller are derived by solving a forward–backward difference equation. In the case of an infinite horizon, a sufficient condition for the stabilization of the system is obtained under the detectable condition. Additionally, in combination with the generalized extended state observer, a controller design method is proposed, and the stability analysis of the system under this controller is presented. This paper details our novel approach to disturbance rejection. Finally, four examples are provided to demonstrate the effectiveness of the proposed method.
{"title":"An approach to mismatched disturbance rejection control for uncontrollable systems","authors":"","doi":"10.1016/j.automatica.2024.111876","DOIUrl":"10.1016/j.automatica.2024.111876","url":null,"abstract":"<div><p>This study focuses on the problem of optimal mismatched disturbance rejection control for uncontrollable linear discrete-time systems. In contrast to previous studies, by introducing a quadratic performance index such that the regulated state can track a reference trajectory and minimize the effects of disturbances, mismatched disturbance rejection control is transformed into a linear quadratic tracking problem. The necessary and sufficient conditions for the solvability of this problem over a finite horizon and a disturbance rejection controller are derived by solving a forward–backward difference equation. In the case of an infinite horizon, a sufficient condition for the stabilization of the system is obtained under the detectable condition. Additionally, in combination with the generalized extended state observer, a controller design method is proposed, and the stability analysis of the system under this controller is presented. This paper details our novel approach to disturbance rejection. Finally, four examples are provided to demonstrate the effectiveness of the proposed method.</p></div>","PeriodicalId":55413,"journal":{"name":"Automatica","volume":null,"pages":null},"PeriodicalIF":4.8,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142095237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}