首页 > 最新文献

IEEE Control Systems Letters最新文献

英文 中文
A Trade-Off Between Time and Energy of Stochastic Higher-Order Networks on Multi-Directed Hypergraphs 多向超图上随机高阶网络的时间与能量权衡
IF 2 Q2 AUTOMATION & CONTROL SYSTEMS Pub Date : 2026-01-12 DOI: 10.1109/LCSYS.2026.3652578
Sasa Chen;Dan Liu;Binrui Wang;Jun Mao;Xiaohang Li
This letter concentrates on the trade-off between control time and energy consumption in achieving finite-time synchronization of stochastic higher-order networks. Different from the existing results that either only consider pairwise interactions between nodes or restrict higher-order interactions to 1-directed hyperedges, multi-directed hypergraphs are employed to better describe group interactions in real networks. A switching control strategy is proposed to estimate the control time and energy consumption in achieving finite-time synchronization of higher-order networks. In addition, to discuss the trade-off between control time and energy consumption, an evaluation index function with adjustable weights is constructed to analyze the optimal control gains in finite time. Finally, an example for higher-order complex networks composed of Chua’s circuits is provided to validate the effectiveness of the proposed results.
本文主要研究随机高阶网络在实现有限时间同步时控制时间与能量消耗之间的权衡。不同于现有的结果只考虑节点间的成对交互或将高阶交互限制在1向超边上,多向超图能够更好地描述真实网络中的群体交互。为了估计高阶网络有限时间同步的控制时间和能量消耗,提出了一种切换控制策略。此外,为了讨论控制时间和能量消耗之间的权衡,构造了一个权值可调的评价指标函数来分析有限时间内的最优控制增益。最后,以蔡氏电路组成的高阶复杂网络为例,验证了所提结果的有效性。
{"title":"A Trade-Off Between Time and Energy of Stochastic Higher-Order Networks on Multi-Directed Hypergraphs","authors":"Sasa Chen;Dan Liu;Binrui Wang;Jun Mao;Xiaohang Li","doi":"10.1109/LCSYS.2026.3652578","DOIUrl":"https://doi.org/10.1109/LCSYS.2026.3652578","url":null,"abstract":"This letter concentrates on the trade-off between control time and energy consumption in achieving finite-time synchronization of stochastic higher-order networks. Different from the existing results that either only consider pairwise interactions between nodes or restrict higher-order interactions to 1-directed hyperedges, multi-directed hypergraphs are employed to better describe group interactions in real networks. A switching control strategy is proposed to estimate the control time and energy consumption in achieving finite-time synchronization of higher-order networks. In addition, to discuss the trade-off between control time and energy consumption, an evaluation index function with adjustable weights is constructed to analyze the optimal control gains in finite time. Finally, an example for higher-order complex networks composed of Chua’s circuits is provided to validate the effectiveness of the proposed results.","PeriodicalId":37235,"journal":{"name":"IEEE Control Systems Letters","volume":"9 ","pages":"3143-3148"},"PeriodicalIF":2.0,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stochastic Coordinate Descent via Regret Minimization 基于遗憾最小化的随机坐标下降
IF 2 Q2 AUTOMATION & CONTROL SYSTEMS Pub Date : 2026-01-12 DOI: 10.1109/LCSYS.2026.3653302
Yankai Lin;Xiaopuwen Wang
In this letter, we investigate the performance of stochastic coordinate descent algorithms for convex optimization problems from the novel perspective of regret minimization. Specifically, we consider a stochastic coordinate selection rule that might vary over time, whereas existing results commonly focus on stochastic selection rules that remain fixed. Moreover, unlike most existing results, which assume that the expectation of algorithm updates follows the exact negative gradient direction, our framework is applicable to any stochastic gradient descent (SGD) like algorithms with estimated gradients that satisfy our main assumption, thus making our approach more general than existing results. We show that the updates of the stochastic algorithm to solve a standard convex optimization problem also serve as an online algorithm for an online convex optimization problem that has time-varying cost functions at each time step. We prove that, under relatively weak assumptions, the online algorithm achieves sublinear regret against the solution of the problem, therefore ensuring the average performance of the algorithm within a finite time window. Finally, we verify our theoretical results by solving a linear regression problem using the proposed algorithm.
在这封信中,我们研究了随机坐标下降算法的性能凸优化问题从后悔最小化的新角度。具体来说,我们考虑了一个随机坐标选择规则,它可能随着时间的推移而变化,而现有的结果通常集中在保持固定的随机选择规则上。此外,与大多数现有结果假设算法更新的期望遵循确切的负梯度方向不同,我们的框架适用于任何具有满足我们主要假设的估计梯度的随机梯度下降(SGD)类算法,从而使我们的方法比现有结果更通用。我们证明了解决标准凸优化问题的随机算法的更新也可以作为在每个时间步具有时变代价函数的在线凸优化问题的在线算法。我们证明了在相对弱的假设下,在线算法对问题的解实现了次线性遗憾,从而保证了算法在有限时间窗内的平均性能。最后,我们通过求解一个线性回归问题来验证我们的理论结果。
{"title":"Stochastic Coordinate Descent via Regret Minimization","authors":"Yankai Lin;Xiaopuwen Wang","doi":"10.1109/LCSYS.2026.3653302","DOIUrl":"https://doi.org/10.1109/LCSYS.2026.3653302","url":null,"abstract":"In this letter, we investigate the performance of stochastic coordinate descent algorithms for convex optimization problems from the novel perspective of regret minimization. Specifically, we consider a stochastic coordinate selection rule that might vary over time, whereas existing results commonly focus on stochastic selection rules that remain fixed. Moreover, unlike most existing results, which assume that the expectation of algorithm updates follows the exact negative gradient direction, our framework is applicable to any stochastic gradient descent (SGD) like algorithms with estimated gradients that satisfy our main assumption, thus making our approach more general than existing results. We show that the updates of the stochastic algorithm to solve a standard convex optimization problem also serve as an online algorithm for an online convex optimization problem that has time-varying cost functions at each time step. We prove that, under relatively weak assumptions, the online algorithm achieves sublinear regret against the solution of the problem, therefore ensuring the average performance of the algorithm within a finite time window. Finally, we verify our theoretical results by solving a linear regression problem using the proposed algorithm.","PeriodicalId":37235,"journal":{"name":"IEEE Control Systems Letters","volume":"9 ","pages":"3155-3160"},"PeriodicalIF":2.0,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Closing the Loop Inside Neural Networks: Causality-Guided Layer Adaptation for Fault Recovery Control 神经网络内部闭环:故障恢复控制的因果导层自适应
IF 2 Q2 AUTOMATION & CONTROL SYSTEMS Pub Date : 2026-01-12 DOI: 10.1109/LCSYS.2026.3652898
Mahdi Taheri;Soon-Jo Chung;Fred Y. Hadaegh
This letter studies the problem of real-time fault recovery control for nonlinear control-affine systems subject to actuator loss of effectiveness faults and external disturbances. We develop a two-stage framework that combines causal inference with selective online adaptation to achieve an effective learning-based recovery control method. In the offline phase, we introduce a causal layer attribution technique based on the average causal effect (ACE) to evaluate the relative importance of each layer in a pretrained deep neural network (DNN) controller compensating for faults. This provides a principled approach to select the most causally influential layer for fault recovery control in the sense of ACE, and goes beyond the widely used last-layer adaptation approach. In the online phase, we deploy a Lyapunov-based gradient update to adapt only the ACE-selected layer to circumvent the need for full-network or last-layer only updates. The proposed adaptive controller guarantees uniform ultimate boundedness (UUB) with exponential convergence of the closed-loop system in the presence of actuator faults and external disturbances. Compared to conventional adaptive DNN controllers with full-network adaptation, our methodology has a reduced computational overhead in the online phase. To demonstrate the effectiveness of our proposed methodology, a case study is provided on a 3-axis attitude control system of a spacecraft with four reaction wheels.
本文研究了非线性仿射控制系统在执行器失效故障和外界干扰下的实时故障恢复控制问题。我们开发了一个两阶段框架,将因果推理与选择性在线适应相结合,以实现有效的基于学习的恢复控制方法。在离线阶段,我们引入了一种基于平均因果效应(ACE)的因果层归因技术来评估预训练深度神经网络(DNN)控制器中各层的相对重要性,以补偿故障。这为ACE意义上的故障恢复控制选择因果影响最大的层提供了一种原则性的方法,超越了广泛使用的最后一层自适应方法。在在线阶段,我们部署了一个基于lyapunov的梯度更新,以仅适应ace选择的层,以规避对全网络或仅最后一层更新的需求。所提出的自适应控制器在存在执行器故障和外部干扰的情况下保证闭环系统的均匀最终有界性和指数收敛性。与具有全网络适应性的传统自适应DNN控制器相比,我们的方法在在线阶段的计算开销减少。为了证明所提出方法的有效性,以具有四个反作用轮的航天器三轴姿态控制系统为例进行了研究。
{"title":"Closing the Loop Inside Neural Networks: Causality-Guided Layer Adaptation for Fault Recovery Control","authors":"Mahdi Taheri;Soon-Jo Chung;Fred Y. Hadaegh","doi":"10.1109/LCSYS.2026.3652898","DOIUrl":"https://doi.org/10.1109/LCSYS.2026.3652898","url":null,"abstract":"This letter studies the problem of real-time fault recovery control for nonlinear control-affine systems subject to actuator loss of effectiveness faults and external disturbances. We develop a two-stage framework that combines causal inference with selective online adaptation to achieve an effective learning-based recovery control method. In the offline phase, we introduce a causal layer attribution technique based on the average causal effect (ACE) to evaluate the relative importance of each layer in a pretrained deep neural network (DNN) controller compensating for faults. This provides a principled approach to select the most causally influential layer for fault recovery control in the sense of ACE, and goes beyond the widely used last-layer adaptation approach. In the online phase, we deploy a Lyapunov-based gradient update to adapt only the ACE-selected layer to circumvent the need for full-network or last-layer only updates. The proposed adaptive controller guarantees uniform ultimate boundedness (UUB) with exponential convergence of the closed-loop system in the presence of actuator faults and external disturbances. Compared to conventional adaptive DNN controllers with full-network adaptation, our methodology has a reduced computational overhead in the online phase. To demonstrate the effectiveness of our proposed methodology, a case study is provided on a 3-axis attitude control system of a spacecraft with four reaction wheels.","PeriodicalId":37235,"journal":{"name":"IEEE Control Systems Letters","volume":"9 ","pages":"3137-3142"},"PeriodicalIF":2.0,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026337","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Encoding High-Level Knowledge in Offline Multi-Agent Reinforcement Learning Using Reward Machines 基于奖励机的离线多智能体强化学习高级知识编码
IF 2 Q2 AUTOMATION & CONTROL SYSTEMS Pub Date : 2026-01-06 DOI: 10.1109/LCSYS.2026.3651656
Shayan Meshkat Alsadat;Zhe Xu
Offline reinforcement learning (RL) learns policies from fixed-size datasets without interacting with the environment, while multi-agent reinforcement learning (MARL) faces challenges from large joint state-action spaces and agent interdependencies. Most offline MARL methods apply regularizations, ignoring system-wide dependencies, risking extrapolation errors. We propose Automata-Guided Multi-Agent Offline RL with Reward Machine (AGMORL), a novel framework extending automata learning to offline MARL with reward machines. AGMORL uses a deterministic finite automaton to learn the reward machine from a dataset, capturing team dynamics and agent interactions, while guiding individual policies to avoid out-of-distribution actions by encoding dataset high-level knowledge. Unlike other methods, it avoids secondary components like generative models. We provide convergence guarantees to an optimal policy and show AGMORL outperforms state-of-the-art offline MARL methods.
离线强化学习(RL)在不与环境交互的情况下从固定大小的数据集中学习策略,而多智能体强化学习(MARL)面临来自大型联合状态-动作空间和智能体相互依赖的挑战。大多数脱机MARL方法应用正则化,忽略系统范围的依赖关系,冒着外推错误的风险。我们提出了一种将自动机学习扩展到带奖励机的离线多智能体强化学习(AGMORL)的新框架。AGMORL使用确定性有限自动机从数据集中学习奖励机,捕捉团队动态和代理交互,同时通过编码数据集高级知识指导个体策略避免偏离分布的行为。与其他方法不同,它避免了生成模型等次要组件。我们提供了最优策略的收敛性保证,并表明AGMORL优于最先进的离线MARL方法。
{"title":"Encoding High-Level Knowledge in Offline Multi-Agent Reinforcement Learning Using Reward Machines","authors":"Shayan Meshkat Alsadat;Zhe Xu","doi":"10.1109/LCSYS.2026.3651656","DOIUrl":"https://doi.org/10.1109/LCSYS.2026.3651656","url":null,"abstract":"Offline reinforcement learning (RL) learns policies from fixed-size datasets without interacting with the environment, while multi-agent reinforcement learning (MARL) faces challenges from large joint state-action spaces and agent interdependencies. Most offline MARL methods apply regularizations, ignoring system-wide dependencies, risking extrapolation errors. We propose Automata-Guided Multi-Agent Offline RL with Reward Machine (AGMORL), a novel framework extending automata learning to offline MARL with reward machines. AGMORL uses a deterministic finite automaton to learn the reward machine from a dataset, capturing team dynamics and agent interactions, while guiding individual policies to avoid out-of-distribution actions by encoding dataset high-level knowledge. Unlike other methods, it avoids secondary components like generative models. We provide convergence guarantees to an optimal policy and show AGMORL outperforms state-of-the-art offline MARL methods.","PeriodicalId":37235,"journal":{"name":"IEEE Control Systems Letters","volume":"9 ","pages":"3149-3154"},"PeriodicalIF":2.0,"publicationDate":"2026-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Formation Tracking for Nonlinear Systems via Prescribed-Time Noncooperative Game 基于规定时间非合作对策的非线性系统编队跟踪
IF 2 Q2 AUTOMATION & CONTROL SYSTEMS Pub Date : 2025-12-31 DOI: 10.1109/LCSYS.2025.3650099
Huan Li;Qian Dong
A prescribed-time non-cooperative game control scheme is investigated for formation tracking of nonlinear multi-agent systems in this letter. Within the framework of non-cooperative game theory, followers pursue self-interested objectives, which may conflict both among themselves and with the collective formation goal, leading to a compromised formation geometry determined by the dynamic Nash equilibrium. Meanwhile, a prescribed-time control scheme is proposed such that the states of formation system can converge to the dynamic Nash equilibrium of formation tracking problem. Furthermore, by integrating a robust term and leveraging the Lipschitz condition, the proposed controller effectively handles the uncertainties in the system. The effectiveness of the method is further confirmed by numerical examples.
本文研究了非线性多智能体系统编队跟踪的一种规定时间非合作博弈控制方案。在非合作博弈论的框架下,follower追求的是自利目标,这可能会导致follower之间以及follower与集体编队目标之间的冲突,从而导致由动态纳什均衡决定的队形折衷。同时,提出了一种规定时间的控制方案,使编队系统的状态收敛到编队跟踪问题的动态纳什均衡。此外,通过积分鲁棒项和利用Lipschitz条件,该控制器有效地处理了系统中的不确定性。数值算例进一步验证了该方法的有效性。
{"title":"Formation Tracking for Nonlinear Systems via Prescribed-Time Noncooperative Game","authors":"Huan Li;Qian Dong","doi":"10.1109/LCSYS.2025.3650099","DOIUrl":"https://doi.org/10.1109/LCSYS.2025.3650099","url":null,"abstract":"A prescribed-time non-cooperative game control scheme is investigated for formation tracking of nonlinear multi-agent systems in this letter. Within the framework of non-cooperative game theory, followers pursue self-interested objectives, which may conflict both among themselves and with the collective formation goal, leading to a compromised formation geometry determined by the dynamic Nash equilibrium. Meanwhile, a prescribed-time control scheme is proposed such that the states of formation system can converge to the dynamic Nash equilibrium of formation tracking problem. Furthermore, by integrating a robust term and leveraging the Lipschitz condition, the proposed controller effectively handles the uncertainties in the system. The effectiveness of the method is further confirmed by numerical examples.","PeriodicalId":37235,"journal":{"name":"IEEE Control Systems Letters","volume":"9 ","pages":"3107-3112"},"PeriodicalIF":2.0,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Delay-Free Adaptive Stepsize for the Incremental Aggregated Gradient Method 增量聚合梯度法的无延迟自适应步长
IF 2 Q2 AUTOMATION & CONTROL SYSTEMS Pub Date : 2025-12-31 DOI: 10.1109/LCSYS.2025.3650105
Zhicheng Deng;Xuyang Wu;Jie Lu
In most existing asynchronous methods, the stepsize depends on an upper bound on the delays and decreases as this bound increases. However, since the upper bound is usually unknown and large, the resulting stepsizes are not only difficult to determine in practice but also overly conservative, which leads to slow convergence. To solve this issue, we propose an adaptive stepsize strategy for a typical asynchronous optimization method—the Incremental Aggregated Gradient (IAG) method. Unlike existing methods in the literature, our stepsize does not rely on any delay information and is less conservative, which leads to easier stepsize determination and faster convergence. Under standard assumptions, we provide the convergence rate of IAG with the proposed stepsize. Numerical experiments demonstrate the superior performance of our stepsize over alternative stepsize strategies.
在大多数现有的异步方法中,步长取决于延迟的上界,并随着该上界的增加而减小。然而,由于上界通常是未知的,而且很大,因此得到的步长不仅在实践中难以确定,而且过于保守,导致收敛缓慢。为了解决这一问题,我们针对典型的异步优化方法-增量聚合梯度(IAG)方法提出了一种自适应步长策略。与文献中已有的方法不同,我们的步长不依赖于任何延迟信息,保守性更低,从而更容易确定步长,收敛速度更快。在标准假设下,我们给出了IAG的收敛速度和所提出的步长。数值实验证明了我们的步长策略优于其他步长策略。
{"title":"A Delay-Free Adaptive Stepsize for the Incremental Aggregated Gradient Method","authors":"Zhicheng Deng;Xuyang Wu;Jie Lu","doi":"10.1109/LCSYS.2025.3650105","DOIUrl":"https://doi.org/10.1109/LCSYS.2025.3650105","url":null,"abstract":"In most existing asynchronous methods, the stepsize depends on an upper bound on the delays and decreases as this bound increases. However, since the upper bound is usually unknown and large, the resulting stepsizes are not only difficult to determine in practice but also overly conservative, which leads to slow convergence. To solve this issue, we propose an adaptive stepsize strategy for a typical asynchronous optimization method—the Incremental Aggregated Gradient (IAG) method. Unlike existing methods in the literature, our stepsize does not rely on any delay information and is less conservative, which leads to easier stepsize determination and faster convergence. Under standard assumptions, we provide the convergence rate of IAG with the proposed stepsize. Numerical experiments demonstrate the superior performance of our stepsize over alternative stepsize strategies.","PeriodicalId":37235,"journal":{"name":"IEEE Control Systems Letters","volume":"9 ","pages":"3095-3100"},"PeriodicalIF":2.0,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929531","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Safe Online Control-Informed Learning 安全的在线控制信息学习
IF 2 Q2 AUTOMATION & CONTROL SYSTEMS Pub Date : 2025-12-26 DOI: 10.1109/LCSYS.2025.3648637
Tianyu Zhou;Zihao Liang;Zehui Lu;Shaoshuai Mou
This letter proposes a Safe Online Control-Informed Learning framework for safety-critical autonomous systems. The framework unifies optimal control, parameter estimation, and safety constraints into an online learning process. It employs an extended Kalman filter to incrementally update system parameters in real time, enabling robust and data-efficient adaptation under uncertainty. A softplus barrier function enforces constraint satisfaction during learning and control while eliminating the dependence on high-quality initial guesses. Theoretical analysis establishes convergence and safety guarantees, and the framework’s effectiveness is demonstrated on cart-pole and robot-arm systems.
这封信为安全关键的自主系统提出了一个安全的在线控制知情学习框架。该框架将最优控制、参数估计和安全约束统一到在线学习过程中。它采用扩展的卡尔曼滤波器实时增量更新系统参数,实现了不确定性下的鲁棒和数据高效自适应。softplus障碍函数在学习和控制过程中加强了约束满足,同时消除了对高质量初始猜测的依赖。理论分析证明了该框架的收敛性和安全性,并在车杆和机械臂系统中验证了该框架的有效性。
{"title":"Safe Online Control-Informed Learning","authors":"Tianyu Zhou;Zihao Liang;Zehui Lu;Shaoshuai Mou","doi":"10.1109/LCSYS.2025.3648637","DOIUrl":"https://doi.org/10.1109/LCSYS.2025.3648637","url":null,"abstract":"This letter proposes a Safe Online Control-Informed Learning framework for safety-critical autonomous systems. The framework unifies optimal control, parameter estimation, and safety constraints into an online learning process. It employs an extended Kalman filter to incrementally update system parameters in real time, enabling robust and data-efficient adaptation under uncertainty. A softplus barrier function enforces constraint satisfaction during learning and control while eliminating the dependence on high-quality initial guesses. Theoretical analysis establishes convergence and safety guarantees, and the framework’s effectiveness is demonstrated on cart-pole and robot-arm systems.","PeriodicalId":37235,"journal":{"name":"IEEE Control Systems Letters","volume":"9 ","pages":"3083-3088"},"PeriodicalIF":2.0,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929353","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Safe Control Synthesis for Neural Network Control Systems via Constrained Zonotopes 约束带拓扑神经网络控制系统的安全控制综合
IF 2 Q2 AUTOMATION & CONTROL SYSTEMS Pub Date : 2025-12-26 DOI: 10.1109/LCSYS.2025.3648776
Hang Zhang;Xiangru Xu
This letter addresses the safe control synthesis problem for neural network control systems subject to bounded unknown disturbances and known exogenous inputs. A forward reachability analysis method is developed to over-approximate the system’s forward reachable sets using constrained zonotopes, where the control sequence appears linearly in both the zonotope center and the right-hand side of the associated equality constraints. Based on these over-approximations, a quadratically constrained program and its convexification are formulated to synthesize control sequences that guarantee safety. A numerical example demonstrates the effectiveness of the proposed approach.
这封信解决了受有界未知干扰和已知外源输入的神经网络控制系统的安全控制综合问题。提出了一种前向可达性分析方法,利用约束带拓扑过度逼近系统的前向可达集,其中控制序列在带拓扑中心和相关等式约束的右侧均线性出现。在这些过逼近的基础上,构造了一个二次约束规划及其凸化,以合成保证安全的控制序列。数值算例验证了该方法的有效性。
{"title":"Safe Control Synthesis for Neural Network Control Systems via Constrained Zonotopes","authors":"Hang Zhang;Xiangru Xu","doi":"10.1109/LCSYS.2025.3648776","DOIUrl":"https://doi.org/10.1109/LCSYS.2025.3648776","url":null,"abstract":"This letter addresses the safe control synthesis problem for neural network control systems subject to bounded unknown disturbances and known exogenous inputs. A forward reachability analysis method is developed to over-approximate the system’s forward reachable sets using constrained zonotopes, where the control sequence appears linearly in both the zonotope center and the right-hand side of the associated equality constraints. Based on these over-approximations, a quadratically constrained program and its convexification are formulated to synthesize control sequences that guarantee safety. A numerical example demonstrates the effectiveness of the proposed approach.","PeriodicalId":37235,"journal":{"name":"IEEE Control Systems Letters","volume":"9 ","pages":"3071-3076"},"PeriodicalIF":2.0,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929277","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Liveness, Reachability, and Reversibility of Signal Interpreted Petri Nets 信号解释Petri网的活动性、可达性和可逆性
IF 2 Q2 AUTOMATION & CONTROL SYSTEMS Pub Date : 2025-12-26 DOI: 10.1109/LCSYS.2025.3648432
Andreas Köhler;Ping Zhang
This letter proposes a set of novel sufficient conditions that guarantee three behavioral properties in signal interpreted Petri nets (SIPNs), i.e., liveness, reachability, and reversibility. SIPNs provide a modeling formalism for representing the control algorithm of discrete manufacturing systems. The liveness, reachability, and reversibility properties ensure that the desired control actions remain perpetually executable, the system states are reachable, and that the system can always return to its initial state, respectively. The sufficient conditions are derived based on the Petri net state equation and the enabling rules of the transitions in SIPNs. Moreover, it is shown how the reachability and reversibility can be computationally verified based on an integer linear programming problem. The computational complexity for verifying the properties is polynomial with respect to the number of markings in the SIPN when the reachable set is already available.
这封信提出了一组新的充分条件,以保证信号解释Petri网(sip)的三个行为特性,即活动性,可达性和可逆性。sipn为离散制造系统的控制算法提供了一种建模形式。活动性、可达性和可逆性属性分别确保所需的控制操作永远保持可执行,系统状态是可达的,并且系统总是可以返回到其初始状态。根据Petri网状态方程和sip中转换的使能规则,导出了充分条件。此外,给出了基于整数线性规划问题的可达性和可逆性的计算验证方法。当可达集已经可用时,验证属性的计算复杂度是相对于SIPN中标记数量的多项式。
{"title":"Liveness, Reachability, and Reversibility of Signal Interpreted Petri Nets","authors":"Andreas Köhler;Ping Zhang","doi":"10.1109/LCSYS.2025.3648432","DOIUrl":"https://doi.org/10.1109/LCSYS.2025.3648432","url":null,"abstract":"This letter proposes a set of novel sufficient conditions that guarantee three behavioral properties in signal interpreted Petri nets (SIPNs), i.e., liveness, reachability, and reversibility. SIPNs provide a modeling formalism for representing the control algorithm of discrete manufacturing systems. The liveness, reachability, and reversibility properties ensure that the desired control actions remain perpetually executable, the system states are reachable, and that the system can always return to its initial state, respectively. The sufficient conditions are derived based on the Petri net state equation and the enabling rules of the transitions in SIPNs. Moreover, it is shown how the reachability and reversibility can be computationally verified based on an integer linear programming problem. The computational complexity for verifying the properties is polynomial with respect to the number of markings in the SIPN when the reachable set is already available.","PeriodicalId":37235,"journal":{"name":"IEEE Control Systems Letters","volume":"9 ","pages":"3131-3136"},"PeriodicalIF":2.0,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Data-Driven Virtual Setpoint P-Type Control With Prescribed Performance Function 具有规定性能功能的数据驱动虚拟设定值p型控制
IF 2 Q2 AUTOMATION & CONTROL SYSTEMS Pub Date : 2025-12-26 DOI: 10.1109/LCSYS.2025.3649131
Guojie Li;Ping Zhou
In this letter, a prescribed performance function-based data-driven virtual setpoint P-type controller (PPF-DDVSPC) is proposed for single-input single-output (SISO) systems with nonlinear nonaffine dynamics. First, the original model with error constraint is converted into an unconstrained form using the prescribed performance function and error transformation technique. A virtual setpoint updating law, nested within the outer layer of the traditional P-type controller, is developed based on the newly defined unconstrained variable to limit the tracking error. Then, the unconstrained model and virtual setpoint law are converted into the available equivalent linear data models through dynamic linearization technology. The unknown pseudo-partial derivatives in the two models are estimated utilizing the modified projection algorithm. Finally, the P-type controller with prescribed performance is obtained by replacing the actual setpoint signal with the resulting virtual setpoint law. The bounded input and bounded output (BIBO) stability of the system is demonstrated by the contraction mapping principle, which ensures that the constraint conditions are satisfied. The effectiveness and robustness of the PPF-DDVSPC method are validated through a data-driven simulation of the blast furnace ironmaking process.
在这封信中,提出了一种基于规定性能函数的数据驱动虚拟设定值p型控制器(PPF-DDVSPC),用于具有非线性非仿射动力学的单输入单输出(SISO)系统。首先,利用规定的性能函数和误差变换技术,将具有误差约束的原始模型转换为无约束的模型;基于新定义的无约束变量,建立了嵌套在传统p型控制器外层的虚拟设定值更新律,以限制跟踪误差。然后,通过动态线性化技术将无约束模型和虚拟设定值律转换为可用的等效线性数据模型。利用改进的投影算法对两个模型中的未知伪偏导数进行估计。最后,用得到的虚设定值律代替实际设定值信号,得到具有规定性能的p型控制器。利用收缩映射原理证明了系统的有界输入有界输出(BIBO)稳定性,保证了约束条件的满足。通过对高炉炼铁过程的数据驱动仿真,验证了PPF-DDVSPC方法的有效性和鲁棒性。
{"title":"Data-Driven Virtual Setpoint P-Type Control With Prescribed Performance Function","authors":"Guojie Li;Ping Zhou","doi":"10.1109/LCSYS.2025.3649131","DOIUrl":"https://doi.org/10.1109/LCSYS.2025.3649131","url":null,"abstract":"In this letter, a prescribed performance function-based data-driven virtual setpoint P-type controller (PPF-DDVSPC) is proposed for single-input single-output (SISO) systems with nonlinear nonaffine dynamics. First, the original model with error constraint is converted into an unconstrained form using the prescribed performance function and error transformation technique. A virtual setpoint updating law, nested within the outer layer of the traditional P-type controller, is developed based on the newly defined unconstrained variable to limit the tracking error. Then, the unconstrained model and virtual setpoint law are converted into the available equivalent linear data models through dynamic linearization technology. The unknown pseudo-partial derivatives in the two models are estimated utilizing the modified projection algorithm. Finally, the P-type controller with prescribed performance is obtained by replacing the actual setpoint signal with the resulting virtual setpoint law. The bounded input and bounded output (BIBO) stability of the system is demonstrated by the contraction mapping principle, which ensures that the constraint conditions are satisfied. The effectiveness and robustness of the PPF-DDVSPC method are validated through a data-driven simulation of the blast furnace ironmaking process.","PeriodicalId":37235,"journal":{"name":"IEEE Control Systems Letters","volume":"9 ","pages":"3089-3094"},"PeriodicalIF":2.0,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929371","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Control Systems Letters
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1