首页 > 最新文献

IEEE open journal of control systems最新文献

英文 中文
On the Representation of Piecewise Quadratic Functions by Neural Networks 用神经网络表示分段二次函数
Pub Date : 2025-09-09 DOI: 10.1109/OJCSYS.2025.3607844
Dieter Teichrib;Moritz Schulze Darup
Neural networks (NNs) are commonly used to approximate functions based on data samples, as they are a universal function approximator for a large class of functions. However, choosing a suitable topology in terms of depth, width and activation function for NNs that allow for low error approximations is a non-trivial task. For the approximation of continuous piecewise affine (PWA) functions, this task has been solved by showing that for every PWA function, there exist NNs with rectified linear unit (relu) and maxout activation that allow an exact representation of the PWA function. This connection between PWA functions and NNs has led to some valuable insights into the representation capabilities of NNs. Moreover, the connection was used in control for approximating the PWA optimal control law of model predictive control (MPC) for linear systems. We show that a similar connection exists between NNs and continuous piecewise quadratic (PWQ) functions by deriving topologies for NNs that allow an exact representation of arbitrary PWQ functions with a polyhedral domain partition. Furthermore, we demonstrate that the proposed NNs can efficiently approximate the PWQ optimal value function for linear MPC.
神经网络通常用于基于数据样本的函数近似,因为它们是一类大函数的通用函数近似器。然而,在深度、宽度和激活函数方面为允许低误差近似的神经网络选择合适的拓扑是一项重要的任务。对于连续分段仿射(PWA)函数的逼近,通过证明对于每个PWA函数,存在具有整流线性单元(relu)和maxout激活的神经网络,可以精确表示PWA函数,从而解决了该任务。PWA函数和神经网络之间的这种联系对神经网络的表示能力产生了一些有价值的见解。并将该连接用于线性系统模型预测控制(MPC)的PWA最优控制律的逼近。我们通过推导神经网络的拓扑,证明了神经网络和连续分段二次(PWQ)函数之间存在类似的联系,该拓扑允许用多面体域划分精确表示任意PWQ函数。此外,我们证明了所提出的神经网络可以有效地近似线性MPC的PWQ最优值函数。
{"title":"On the Representation of Piecewise Quadratic Functions by Neural Networks","authors":"Dieter Teichrib;Moritz Schulze Darup","doi":"10.1109/OJCSYS.2025.3607844","DOIUrl":"https://doi.org/10.1109/OJCSYS.2025.3607844","url":null,"abstract":"Neural networks (NNs) are commonly used to approximate functions based on data samples, as they are a universal function approximator for a large class of functions. However, choosing a suitable topology in terms of depth, width and activation function for NNs that allow for low error approximations is a non-trivial task. For the approximation of continuous piecewise affine (PWA) functions, this task has been solved by showing that for every PWA function, there exist NNs with rectified linear unit (relu) and maxout activation that allow an exact representation of the PWA function. This connection between PWA functions and NNs has led to some valuable insights into the representation capabilities of NNs. Moreover, the connection was used in control for approximating the PWA optimal control law of model predictive control (MPC) for linear systems. We show that a similar connection exists between NNs and continuous piecewise quadratic (PWQ) functions by deriving topologies for NNs that allow an exact representation of arbitrary PWQ functions with a polyhedral domain partition. Furthermore, we demonstrate that the proposed NNs can efficiently approximate the PWQ optimal value function for linear MPC.","PeriodicalId":73299,"journal":{"name":"IEEE open journal of control systems","volume":"4 ","pages":"447-462"},"PeriodicalIF":0.0,"publicationDate":"2025-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11153811","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145352169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Leveraging Analytic Gradients in Provably Safe Reinforcement Learning 在可证明安全的强化学习中利用分析梯度
Pub Date : 2025-09-09 DOI: 10.1109/OJCSYS.2025.3607845
Tim Walter;Hannah Markgraf;Jonathan Külz;Matthias Althoff
The deployment of autonomous robots in safety-critical applications requires safety guarantees. Provably safe reinforcement learning is an active field of research that aims to provide such guarantees using safeguards. These safeguards should be integrated during training to reduce the sim-to-real gap. While there are several approaches for safeguarding sampling-based reinforcement learning, analytic gradient-based reinforcement learning often achieves superior performance from fewer environment interactions. However, there is no safeguarding approach for this learning paradigm yet. Our work addresses this gap by developing the first effective safeguard for analytic gradient-based reinforcement learning. We analyse existing, differentiable safeguards, adapt them through modified mappings and gradient formulations, and integrate them into a state-of-the-art learning algorithm and a differentiable simulation. Using numerical experiments on three control tasks, we evaluate how different safeguards affect learning. The results demonstrate safeguarded training without compromising performance. Additional visuals are provided at timwalter.github.io/safe-agb-rl.github.io.
在安全关键应用中部署自主机器人需要安全保障。可证明的安全强化学习是一个活跃的研究领域,旨在使用安全措施提供这样的保证。这些保障措施应在培训期间结合起来,以减少模拟与实际的差距。虽然有几种方法可以保护基于采样的强化学习,但基于分析梯度的强化学习通常可以从更少的环境交互中获得更好的性能。然而,目前还没有针对这种学习范式的保护方法。我们的工作通过开发基于分析梯度的强化学习的第一个有效保障来解决这一差距。我们分析了现有的可微保障措施,通过修改映射和梯度公式对其进行调整,并将其集成到最先进的学习算法和可微模拟中。通过对三个控制任务的数值实验,我们评估了不同的保护措施如何影响学习。结果表明,在不影响性能的情况下,有保障的训练。额外的视觉效果提供在timwalter.github.io/safe-agb-rl.github.io。
{"title":"Leveraging Analytic Gradients in Provably Safe Reinforcement Learning","authors":"Tim Walter;Hannah Markgraf;Jonathan Külz;Matthias Althoff","doi":"10.1109/OJCSYS.2025.3607845","DOIUrl":"https://doi.org/10.1109/OJCSYS.2025.3607845","url":null,"abstract":"The deployment of autonomous robots in safety-critical applications requires safety guarantees. Provably safe reinforcement learning is an active field of research that aims to provide such guarantees using safeguards. These safeguards should be integrated during training to reduce the sim-to-real gap. While there are several approaches for safeguarding sampling-based reinforcement learning, analytic gradient-based reinforcement learning often achieves superior performance from fewer environment interactions. However, there is no safeguarding approach for this learning paradigm yet. Our work addresses this gap by developing the first effective safeguard for analytic gradient-based reinforcement learning. We analyse existing, differentiable safeguards, adapt them through modified mappings and gradient formulations, and integrate them into a state-of-the-art learning algorithm and a differentiable simulation. Using numerical experiments on three control tasks, we evaluate how different safeguards affect learning. The results demonstrate safeguarded training without compromising performance. Additional visuals are provided at timwalter.github.io/safe-agb-rl.github.io.","PeriodicalId":73299,"journal":{"name":"IEEE open journal of control systems","volume":"4 ","pages":"463-481"},"PeriodicalIF":0.0,"publicationDate":"2025-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11154003","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145351918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cooperative Pursuit-Evasion Games With a Flat Sphere Condition 具有平球条件的追捕-逃避合作博弈
Pub Date : 2025-09-01 DOI: 10.1109/OJCSYS.2025.3604640
Dejan Milutinović;Alexander Von Moll;Satyanarayana Gupta Manyam;David W. Casbeer;Isaac E. Weintraub;Meir Pachter
In planar pursuit-evasion differential games considering a faster pursuer and a slower evader, the interception points resulting from equilibrium strategies lie on the Apollonius circle. This property is instrumental for leveraging geometric approaches for solving multiple pursuit-evasion scenarios in the plane. In this paper, we study a pursuit-evasion differential game on a sphere and generalize the planar Apollonius circle to the spherical domain. For the differential game, we provide equilibrium strategies for all initial positions of the pursuer and evader, including a special case when they are on the opposite sides of the sphere and on the same line with the center of the sphere when there are infinitely many geodesics between the two players. In contrast to planar scenarios, on the sphere we find that the interception point from the equilibrium strategies can leave the Apollonius domain boundary. We present a condition to ensure the intercept point remains on the boundary of the Apollonius domain. This condition allows for generalizing planar pursuit-evasion strategies to the sphere, and we show how these results are applied by analyzing the scenarios of target guarding and two-pursuer, single evader differential games on the sphere.
在考虑快速追逃者和慢速追逃者的平面追逃微分对策中,均衡策略产生的拦截点位于阿波罗尼乌斯圆上。这个性质有助于利用几何方法来解决平面上的多个追捕逃避场景。本文研究了球面上的追-避微分对策,并将平面阿波罗圆推广到球面上。对于微分对策,我们提供了追捕者和逃避者的所有初始位置的均衡策略,包括当他们在球体的相对两侧并且与球体中心在同一条线上时,当两个参与者之间存在无限多条测地线时的特殊情况。与平面场景相比,在球面上,我们发现平衡策略的拦截点可以离开阿波罗尼乌斯域边界。我们提出了一个保证截点保持在阿波罗尼乌斯域边界上的条件。这个条件允许将平面追捕-逃避策略推广到球体,我们通过分析球体上的目标守卫和两个追捕者,一个逃避者微分博弈的场景来展示这些结果是如何应用的。
{"title":"Cooperative Pursuit-Evasion Games With a Flat Sphere Condition","authors":"Dejan Milutinović;Alexander Von Moll;Satyanarayana Gupta Manyam;David W. Casbeer;Isaac E. Weintraub;Meir Pachter","doi":"10.1109/OJCSYS.2025.3604640","DOIUrl":"https://doi.org/10.1109/OJCSYS.2025.3604640","url":null,"abstract":"In planar pursuit-evasion differential games considering a faster pursuer and a slower evader, the interception points resulting from equilibrium strategies lie on the Apollonius circle. This property is instrumental for leveraging geometric approaches for solving multiple pursuit-evasion scenarios in the plane. In this paper, we study a pursuit-evasion differential game on a sphere and generalize the planar Apollonius circle to the spherical domain. For the differential game, we provide equilibrium strategies for all initial positions of the pursuer and evader, including a special case when they are on the opposite sides of the sphere and on the same line with the center of the sphere when there are infinitely many geodesics between the two players. In contrast to planar scenarios, on the sphere we find that the interception point from the equilibrium strategies can leave the Apollonius domain boundary. We present a condition to ensure the intercept point remains on the boundary of the Apollonius domain. This condition allows for generalizing planar pursuit-evasion strategies to the sphere, and we show how these results are applied by analyzing the scenarios of target guarding and two-pursuer, single evader differential games on the sphere.","PeriodicalId":73299,"journal":{"name":"IEEE open journal of control systems","volume":"4 ","pages":"360-372"},"PeriodicalIF":0.0,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11145945","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145141640","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gaussian Process Supported Stochastic MPC for Distribution Grids 配电网高斯过程支持的随机MPC算法
Pub Date : 2025-08-22 DOI: 10.1109/OJCSYS.2025.3601836
Moritz Wenzel;Edoardo De Din;Marcel Zimmer;Andrea Benigni
The efficacy of control systems for distribution grids can be influenced by different sources of uncertainty. Stochastic Model Predictive Control (SMPC) can be employed to compensate for such uncertainties by integrating their probability distribution into the control problem. An efficient SMPC algorithm for online control applications is the stochastic tube SMPC, which is able to treat the evaluation of the chance constraints analytically. However, this approach is efficient only when the calculation of the constraint back-off is applied to a linear model. To address this issue, this work employs Gaussian Processes to approximate the nonlinear part of the power flow equations based on offline training, which is integrated into the SMPC formulation. The resulting SMPC is first validated and then tested on a benchmark system, comparing the results with Deterministic MPC and SMPC that excludes Gaussian Processes. The proposed SMPC proves to be more efficient in terms of cost minimization, reference tracking and voltage violationreduction.
配电网控制系统的有效性会受到不同不确定性来源的影响。随机模型预测控制(SMPC)可以通过将这些不确定性的概率分布整合到控制问题中来补偿这些不确定性。对于在线控制应用,一种有效的SMPC算法是随机管式SMPC算法,它能够解析地处理机会约束的评估。然而,这种方法只有当约束退离的计算应用于线性模型时才有效。为了解决这一问题,本文采用基于离线训练的高斯过程来近似功率流方程的非线性部分,并将其集成到SMPC公式中。首先对所得的SMPC进行验证,然后在基准系统上进行测试,并将结果与Deterministic MPC和排除高斯过程的SMPC进行比较。所提出的SMPC在成本最小化、参考跟踪和电压冲突减少方面具有更高的效率。
{"title":"Gaussian Process Supported Stochastic MPC for Distribution Grids","authors":"Moritz Wenzel;Edoardo De Din;Marcel Zimmer;Andrea Benigni","doi":"10.1109/OJCSYS.2025.3601836","DOIUrl":"https://doi.org/10.1109/OJCSYS.2025.3601836","url":null,"abstract":"The efficacy of control systems for distribution grids can be influenced by different sources of uncertainty. Stochastic Model Predictive Control (SMPC) can be employed to compensate for such uncertainties by integrating their probability distribution into the control problem. An efficient SMPC algorithm for online control applications is the stochastic tube SMPC, which is able to treat the evaluation of the chance constraints analytically. However, this approach is efficient only when the calculation of the constraint back-off is applied to a linear model. To address this issue, this work employs Gaussian Processes to approximate the nonlinear part of the power flow equations based on offline training, which is integrated into the SMPC formulation. The resulting SMPC is first validated and then tested on a benchmark system, comparing the results with Deterministic MPC and SMPC that excludes Gaussian Processes. The proposed SMPC proves to be more efficient in terms of cost minimization, reference tracking and voltage violationreduction.","PeriodicalId":73299,"journal":{"name":"IEEE open journal of control systems","volume":"4 ","pages":"332-348"},"PeriodicalIF":0.0,"publicationDate":"2025-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11134570","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145027992","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integrating Reinforcement Learning and Model Predictive Control for Mixed- Logical Dynamical Systems 混合逻辑动力系统强化学习与模型预测控制的集成
Pub Date : 2025-08-21 DOI: 10.1109/OJCSYS.2025.3601435
Caio Fabio Oliveira da Silva;Azita Dabiri;Bart De Schutter
This work proposes an approach that integrates reinforcement learning (RL) and model predictive control (MPC) to solve finite-horizon optimal control problems in mixed-logical dynamical systems efficiently. Optimization-based control of such systems with discrete and continuous decision variables entails the online solution of mixed-integer linear programs, which suffer from the curse of dimensionality. In the proposed approach, by repeated interaction with a simulator of the system, a reinforcement learning agent is trained to provide a policy for the discrete decision variables. During online operation, the RL policy simplifies the online optimization problem of the MPC controller from a mixed-integer linear program to a linear program, significantly reducing the computation time. A fundamental contribution of this work is the definition of the decoupled Q-function, which plays a crucial role in making the learning problem tractable in a combinatorial action space. We motivate the use of recurrent neural networks to approximate the decoupled Q-function and show how they can be employed in a reinforcement learning setting. A microgrid system is used as an illustrative example where real-world data is used to demonstrate that the proposed method substantially reduces the maximum online computation time of MPC (up to $20times$) while maintaining high feasibility and average optimality gap lower than 1.1% .
本文提出了一种集成强化学习(RL)和模型预测控制(MPC)的方法来有效地解决混合逻辑动态系统中的有限视界最优控制问题。这类具有离散和连续决策变量的系统的优化控制需要在线求解受维数诅咒的混合整数线性规划。在提出的方法中,通过与系统模拟器的重复交互,训练强化学习代理为离散决策变量提供策略。在线运行时,RL策略将MPC控制器的在线优化问题从混合整数线性规划简化为线性规划,大大减少了计算时间。这项工作的一个基本贡献是解耦q函数的定义,它在使学习问题在组合动作空间中易于处理方面起着至关重要的作用。我们鼓励使用循环神经网络来近似解耦的q函数,并展示如何在强化学习设置中使用它们。以微电网系统为例,使用实际数据证明所提出的方法大大减少了MPC的最大在线计算时间(高达$20times$),同时保持高可行性和平均最优性差距低于1.1%。
{"title":"Integrating Reinforcement Learning and Model Predictive Control for Mixed- Logical Dynamical Systems","authors":"Caio Fabio Oliveira da Silva;Azita Dabiri;Bart De Schutter","doi":"10.1109/OJCSYS.2025.3601435","DOIUrl":"https://doi.org/10.1109/OJCSYS.2025.3601435","url":null,"abstract":"This work proposes an approach that integrates reinforcement learning (RL) and model predictive control (MPC) to solve finite-horizon optimal control problems in mixed-logical dynamical systems efficiently. Optimization-based control of such systems with discrete and continuous decision variables entails the online solution of mixed-integer linear programs, which suffer from the curse of dimensionality. In the proposed approach, by repeated interaction with a simulator of the system, a reinforcement learning agent is trained to provide a policy for the discrete decision variables. During online operation, the RL policy simplifies the online optimization problem of the MPC controller from a mixed-integer linear program to a linear program, significantly reducing the computation time. A fundamental contribution of this work is the definition of the decoupled Q-function, which plays a crucial role in making the learning problem tractable in a combinatorial action space. We motivate the use of recurrent neural networks to approximate the decoupled Q-function and show how they can be employed in a reinforcement learning setting. A microgrid system is used as an illustrative example where real-world data is used to demonstrate that the proposed method substantially reduces the maximum online computation time of MPC (up to <inline-formula><tex-math>$20times$</tex-math></inline-formula>) while maintaining high feasibility and average optimality gap lower than 1.1% .","PeriodicalId":73299,"journal":{"name":"IEEE open journal of control systems","volume":"4 ","pages":"316-331"},"PeriodicalIF":0.0,"publicationDate":"2025-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11134093","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145027930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MDP-Based High-Level Decision-Making for Combining Safety and Optimality: Autonomous Overtaking 基于mdp的安全性与最优性相结合的高层决策:自动超车
Pub Date : 2025-08-20 DOI: 10.1109/OJCSYS.2025.3600925
Xue-Fang Wang;Jingjing Jiang;Wen-Hua Chen
This paper presents a novel solution for optimal high-level decision-making in autonomous overtaking on two-lane roads, considering both opposite-direction and same-direction traffic. The proposed solutionaccounts for key factors such as safety and optimality, while also ensuring recursive feasibility and stability.To safely complete overtaking maneuvers, the solution is built on a constrained Markov decision process (MDP) that generates optimal decisions for path planners. By combining MDP with model predictive control (MPC), the approach guarantees recursive feasibility and stability through a baseline control policy that calculates the terminal cost and is incorporated into a constructed Lyapunov function. The proposed solution is validated through five simulated driving scenarios, demonstrating its robustness in handling diverse interactions within dynamic and complex traffic conditions.
提出了一种考虑反向和同向车辆的双车道自动超车高层最优决策的新方法。该方案兼顾了安全性和最优性等关键因素,同时保证了递归的可行性和稳定性。为了安全地完成超车操作,该解决方案建立在约束马尔可夫决策过程(MDP)的基础上,该决策过程为路径规划者生成最优决策。通过将MDP与模型预测控制(MPC)相结合,该方法通过计算终端成本的基线控制策略来保证递归的可行性和稳定性,并将其纳入构建的Lyapunov函数中。通过五个模拟驾驶场景验证了该方案的有效性,证明了其在处理动态和复杂交通条件下的各种交互方面的鲁棒性。
{"title":"MDP-Based High-Level Decision-Making for Combining Safety and Optimality: Autonomous Overtaking","authors":"Xue-Fang Wang;Jingjing Jiang;Wen-Hua Chen","doi":"10.1109/OJCSYS.2025.3600925","DOIUrl":"https://doi.org/10.1109/OJCSYS.2025.3600925","url":null,"abstract":"This paper presents a novel solution for optimal high-level decision-making in autonomous overtaking on two-lane roads, considering both opposite-direction and same-direction traffic. The proposed solutionaccounts for key factors such as safety and optimality, while also ensuring recursive feasibility and stability.To safely complete overtaking maneuvers, the solution is built on a constrained Markov decision process (MDP) that generates optimal decisions for path planners. By combining MDP with model predictive control (MPC), the approach guarantees recursive feasibility and stability through a baseline control policy that calculates the terminal cost and is incorporated into a constructed Lyapunov function. The proposed solution is validated through five simulated driving scenarios, demonstrating its robustness in handling diverse interactions within dynamic and complex traffic conditions.","PeriodicalId":73299,"journal":{"name":"IEEE open journal of control systems","volume":"4 ","pages":"299-315"},"PeriodicalIF":0.0,"publicationDate":"2025-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11130904","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145021272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feature Construction Using Network Control Theory and Rank Encoding for Graph Machine Learning 基于网络控制理论和秩编码的图机学习特征构建
Pub Date : 2025-08-15 DOI: 10.1109/OJCSYS.2025.3599371
Anwar Said;Yifan Wei;Obaid Ullah Ahmad;Mudassir Shabbir;Waseem Abbas;Xenofon Koutsoukos
In this article, we utilize the concept of average controllability in graphs, along with a novel rank encoding method, to enhance the performance of Graph Neural Networks (GNNs) in social network classification tasks. GNNs have proven highly effective in various network-based learning applications and require some form of node features to function. However, their performance is heavily influenced by the expressiveness of these features. In social networks, node features are often unavailable due to privacy constraints or the absence of inherent attributes, making it challenging for GNNs to achieve optimal performance. To address this limitation, we propose two strategies for constructing expressive node features. First, we introduce average controllability along with other centrality metrics (denoted as NCT-EFA) as node-level metrics that capture critical aspects of network topology. Building on this, we develop a rank encoding method that transforms average controllability—or any other graph-theoretic metric—into a fixed-dimensional feature space, thereby improving feature representation. We conduct extensive numerical evaluations using six benchmark GNN models across four social network datasets to compare different node feature construction methods. Our results demonstrate that incorporating average controllability into the feature space significantly improves GNN performance. Moreover, the proposed rank encoding method outperforms traditional one-hot degree encoding, improving the ROC AUC from 68.7% to 73.9% using GraphSAGE on the GitHub Stargazers dataset, underscoring its effectiveness in generating expressive and efficient node representations.
在本文中,我们利用图中平均可控性的概念,以及一种新的秩编码方法,来提高图神经网络(gnn)在社交网络分类任务中的性能。gnn已被证明在各种基于网络的学习应用中非常有效,并且需要某种形式的节点特征才能发挥作用。然而,它们的性能在很大程度上受到这些特征的表现力的影响。在社交网络中,由于隐私约束或缺乏固有属性,节点特征通常不可用,这给gnn实现最佳性能带来了挑战。为了解决这一限制,我们提出了两种构建表达性节点特征的策略。首先,我们将平均可控性与其他中心性指标(表示为NCT-EFA)一起引入,作为捕获网络拓扑关键方面的节点级指标。在此基础上,我们开发了一种秩编码方法,将平均可控性(或任何其他图论度量)转换为固定维特征空间,从而改进特征表示。我们在四个社交网络数据集上使用六个基准GNN模型进行了广泛的数值评估,以比较不同的节点特征构建方法。我们的研究结果表明,将平均可控性纳入特征空间可以显著提高GNN的性能。此外,本文提出的秩编码方法优于传统的单热度编码,在GitHub Stargazers数据集上使用GraphSAGE将ROC AUC从68.7%提高到73.9%,强调了其在生成富有表现力和高效的节点表示方面的有效性。
{"title":"Feature Construction Using Network Control Theory and Rank Encoding for Graph Machine Learning","authors":"Anwar Said;Yifan Wei;Obaid Ullah Ahmad;Mudassir Shabbir;Waseem Abbas;Xenofon Koutsoukos","doi":"10.1109/OJCSYS.2025.3599371","DOIUrl":"https://doi.org/10.1109/OJCSYS.2025.3599371","url":null,"abstract":"In this article, we utilize the concept of average controllability in graphs, along with a novel rank encoding method, to enhance the performance of Graph Neural Networks (GNNs) in social network classification tasks. GNNs have proven highly effective in various network-based learning applications and require some form of node features to function. However, their performance is heavily influenced by the expressiveness of these features. In social networks, node features are often unavailable due to privacy constraints or the absence of inherent attributes, making it challenging for GNNs to achieve optimal performance. To address this limitation, we propose two strategies for constructing expressive node features. First, we introduce average controllability along with other centrality metrics (denoted as NCT-EFA) as node-level metrics that capture critical aspects of network topology. Building on this, we develop a rank encoding method that transforms average controllability—or any other graph-theoretic metric—into a fixed-dimensional feature space, thereby improving feature representation. We conduct extensive numerical evaluations using six benchmark GNN models across four social network datasets to compare different node feature construction methods. Our results demonstrate that incorporating average controllability into the feature space significantly improves GNN performance. Moreover, the proposed rank encoding method outperforms traditional one-hot degree encoding, improving the ROC AUC from 68.7% to 73.9% using GraphSAGE on the GitHub Stargazers dataset, underscoring its effectiveness in generating expressive and efficient node representations.","PeriodicalId":73299,"journal":{"name":"IEEE open journal of control systems","volume":"4 ","pages":"288-298"},"PeriodicalIF":0.0,"publicationDate":"2025-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11126872","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145021271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Capture-the-Flag Differential Game: Attack, Transition and Retreat 夺旗微分博弈:进攻、转换和撤退
Pub Date : 2025-08-14 DOI: 10.1109/OJCSYS.2025.3599473
Eloy Garcia;David W. Casbeer
This paper analyzes the classic game of capture-the-flag, modeled as a conflict between an Attacker and a Defender. The game unfolds in distinct phases with changing objectives: first, the Attacker tries to capture a flag while the Defender attempts to intercept; second, if successful, the Attacker tries to reach a safe zone while the Defender again seeks interception. We mathematically derive the optimal state-feedback strategies for both players and the associated Value function for each phase, rigorously proving their correctness. A key contribution is introducing the transition phase, where we analyze the Defender’s optimal repositioning strategy when flag capture becomes inevitable, preparing it for the game’s second phase. This novel transition connects the game’s stages, critically enabling us to solve the overall Game of Kind – determining the winner from any starting condition – and define the precise circumstances under which the Attacker can both capture the flag and successfully escape to the safe zone.
本文分析了经典的夺旗游戏,将其建模为攻击者和防御者之间的冲突。游戏以不同的阶段展开,目标不断变化:首先,攻击者试图夺取一面旗帜,而防守者试图拦截;其次,如果成功,攻击者尝试到达安全区域,而防御者再次寻求拦截。我们从数学上推导出每个阶段参与者的最佳状态反馈策略和相关的价值函数,严格证明了它们的正确性。一个关键的贡献是引入过渡阶段,在这个阶段,我们分析了当夺旗不可避免时防守者的最佳重新定位策略,为游戏的第二阶段做准备。这种新颖的过渡连接了游戏的各个阶段,使我们能够解决整个“同类游戏”(game of Kind)——在任何起始条件下决定获胜者——并定义攻击者既能夺取旗帜又能成功逃到安全区的精确环境。
{"title":"The Capture-the-Flag Differential Game: Attack, Transition and Retreat","authors":"Eloy Garcia;David W. Casbeer","doi":"10.1109/OJCSYS.2025.3599473","DOIUrl":"https://doi.org/10.1109/OJCSYS.2025.3599473","url":null,"abstract":"This paper analyzes the classic game of capture-the-flag, modeled as a conflict between an Attacker and a Defender. The game unfolds in distinct phases with changing objectives: first, the Attacker tries to capture a flag while the Defender attempts to intercept; second, if successful, the Attacker tries to reach a safe zone while the Defender again seeks interception. We mathematically derive the optimal state-feedback strategies for both players and the associated Value function for each phase, rigorously proving their correctness. A key contribution is introducing the transition phase, where we analyze the Defender’s optimal repositioning strategy when flag capture becomes inevitable, preparing it for the game’s second phase. This novel transition connects the game’s stages, critically enabling us to solve the overall Game of Kind – determining the winner from any starting condition – and define the precise circumstances under which the Attacker can both capture the flag and successfully escape to the safe zone.","PeriodicalId":73299,"journal":{"name":"IEEE open journal of control systems","volume":"4 ","pages":"271-287"},"PeriodicalIF":0.0,"publicationDate":"2025-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11125922","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144998218","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Vision-Based Collision Avoidance for Multi-Agent Systems With Intermittent Measurements 基于视觉的间歇测量多智能体系统避碰
Pub Date : 2025-08-13 DOI: 10.1109/OJCSYS.2025.3598673
Mia Scoblic;Camilla Tabasso;Venanzio Cichella;Isaac Kaminer
Collision avoidance is a fundamental aspect of many applications involving autonomous vehicles. Solving this problem becomes especially challenging when the agents involved cannot communicate. In these scenarios, onboard sensors are essential for detecting and avoiding other vehicles or obstacles. However, in many practical applications, sensors have limited range and measurements may be intermittent due to external factors. With this in mind, in this work, we present a novel decentralized vision-based collision avoidance algorithm which does not require communication among the agents and has mild assumptions on the sensing capabilities of the vehicles. Once a collision is detected, the agents replan their trajectories to follow a circular path centered at the point of collision. A feedback control law is designed so that the vehicles can maintain a predefined phase shift along this circle and therefore are able to avoid collisions. A Lyapunov analysis is performed to provide performance bounds and the efficacy of the proposed method is demonstrated through both simulated and experimental results.
避免碰撞是许多自动驾驶汽车应用的一个基本方面。当涉及的代理无法通信时,解决这个问题变得特别具有挑战性。在这些情况下,车载传感器对于探测和避开其他车辆或障碍物至关重要。然而,在许多实际应用中,传感器的测量范围有限,并且由于外部因素的影响,测量可能是间歇性的。考虑到这一点,在这项工作中,我们提出了一种新的基于分散视觉的避碰算法,该算法不需要智能体之间的通信,并且对车辆的感知能力有轻微的假设。一旦检测到碰撞,智能体就会重新规划它们的轨迹,沿着以碰撞点为中心的圆形路径前进。设计了一种反馈控制律,使车辆能够沿此圆保持预定的相移,从而避免碰撞。李雅普诺夫分析提供了性能界限,并通过模拟和实验结果证明了所提出方法的有效性。
{"title":"Vision-Based Collision Avoidance for Multi-Agent Systems With Intermittent Measurements","authors":"Mia Scoblic;Camilla Tabasso;Venanzio Cichella;Isaac Kaminer","doi":"10.1109/OJCSYS.2025.3598673","DOIUrl":"https://doi.org/10.1109/OJCSYS.2025.3598673","url":null,"abstract":"Collision avoidance is a fundamental aspect of many applications involving autonomous vehicles. Solving this problem becomes especially challenging when the agents involved cannot communicate. In these scenarios, onboard sensors are essential for detecting and avoiding other vehicles or obstacles. However, in many practical applications, sensors have limited range and measurements may be intermittent due to external factors. With this in mind, in this work, we present a novel decentralized vision-based collision avoidance algorithm which does not require communication among the agents and has mild assumptions on the sensing capabilities of the vehicles. Once a collision is detected, the agents replan their trajectories to follow a circular path centered at the point of collision. A feedback control law is designed so that the vehicles can maintain a predefined phase shift along this circle and therefore are able to avoid collisions. A Lyapunov analysis is performed to provide performance bounds and the efficacy of the proposed method is demonstrated through both simulated and experimental results.","PeriodicalId":73299,"journal":{"name":"IEEE open journal of control systems","volume":"4 ","pages":"349-359"},"PeriodicalIF":0.0,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11123838","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778185","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generalized Concentration-Based Performance Guarantees on Sensor Selection for State Estimation 基于广义浓度的状态估计传感器选择性能保证
Pub Date : 2025-08-13 DOI: 10.1109/OJCSYS.2025.3598626
Christopher I. Calle;Shaunak D. Bopardikar
In this work, we apply concentration-based results to the problem of sensor selection for state estimation to provide us with meaningful guarantees on the properties of our selection. We consider a selection of sensors that is randomly chosen with replacement for a stochastic linear dynamical system, and we utilize the Kalman filter to perform state estimation. Our main contributions are four-fold. First, we derive novel matrix concentration inequalities (CIs) for a sum of positive semi-definite random matrices. Second, we provide two algorithms for specifying the parameters required to apply our matrix CIs, a novel statistical tool. Third, we propose two avenues for improving the sample complexity of this statistical tool. Fourth, we provide a procedure for optimizing the semi-definite bounds of our matrix CIs. When our matrix CIs are applied to the problem of sensor selection for state estimation, our final contribution is a procedure for optimizing the filtered state estimation error covariance matrix of the Kalman filter. Finally, we show through simulations that our bounds significantly outperform those of an existing matrix CI and are applicable for a larger parameter regime. Also, we demonstrate the applicability of our matrix CIs for the state estimation of nonlinear dynamical systems.
在这项工作中,我们将基于浓度的结果应用于状态估计的传感器选择问题,为我们提供了对我们选择的属性的有意义的保证。我们考虑用随机线性动力系统替换随机选择的传感器的选择,并利用卡尔曼滤波器进行状态估计。我们的主要贡献有四方面。首先,我们导出了一类正半定随机矩阵和的矩阵浓度不等式。其次,我们提供了两种算法来指定应用我们的矩阵CIs(一种新的统计工具)所需的参数。第三,我们提出了两种方法来提高该统计工具的样本复杂度。第四,我们提供了一个优化矩阵ci的半定界的过程。当我们的矩阵ci应用于状态估计的传感器选择问题时,我们的最终贡献是优化卡尔曼滤波器的滤波状态估计误差协方差矩阵的过程。最后,我们通过模拟表明,我们的边界明显优于现有矩阵CI的边界,并且适用于更大的参数范围。同时,我们也证明了矩阵ci对非线性动力系统状态估计的适用性。
{"title":"Generalized Concentration-Based Performance Guarantees on Sensor Selection for State Estimation","authors":"Christopher I. Calle;Shaunak D. Bopardikar","doi":"10.1109/OJCSYS.2025.3598626","DOIUrl":"https://doi.org/10.1109/OJCSYS.2025.3598626","url":null,"abstract":"In this work, we apply concentration-based results to the problem of sensor selection for state estimation to provide us with meaningful guarantees on the properties of our selection. We consider a selection of sensors that is randomly chosen with replacement for a stochastic linear dynamical system, and we utilize the Kalman filter to perform state estimation. Our main contributions are four-fold. First, we derive novel matrix concentration inequalities (CIs) for a sum of positive semi-definite random matrices. Second, we provide two algorithms for specifying the parameters required to apply our matrix CIs, a novel statistical tool. Third, we propose two avenues for improving the sample complexity of this statistical tool. Fourth, we provide a procedure for optimizing the semi-definite bounds of our matrix CIs. When our matrix CIs are applied to the problem of sensor selection for state estimation, our final contribution is a procedure for optimizing the filtered state estimation error covariance matrix of the Kalman filter. Finally, we show through simulations that our bounds significantly outperform those of an existing matrix CI and are applicable for a larger parameter regime. Also, we demonstrate the applicability of our matrix CIs for the state estimation of nonlinear dynamical systems.","PeriodicalId":73299,"journal":{"name":"IEEE open journal of control systems","volume":"4 ","pages":"250-270"},"PeriodicalIF":0.0,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11123730","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144990018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE open journal of control systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1