首页 > 最新文献

Expert Systems with Applications最新文献

英文 中文
Scheduling a Constrained Hybrid Flowshop Using a Variable Representation Cooperative Co-Evolutionary Algorithm 基于变量表示协同进化算法的约束混合流水车间调度
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-13 DOI: 10.1016/j.eswa.2026.131136
Bing-Tao Wang , Quan-Ke Pan , Shengxiang Yang , Xue-Lei Jing , WeiMin Li
The research addresses a hybrid flowshop scheduling problem incorporating worker competency constraints. Unlike most existing studies that assume workers can operate all machines, our work accounts for the absence of certain worker skills. The added constraints substantially increase the problem’s complexity, rendering traditional algorithms inadequate for obtaining feasible solutions. Therefore, a mixed-integer programming model is formulated, and a variable representation cooperative co-evolutionary algorithm (VRCCEA) is designed to achieve makespan minimization. Based on the decomposition idea, we use two populations to address the multi-coupled problem and implement a cooperative mechanism by introducing a solution archive to promote the co-evolution of populations. Given the limitations of a single encoding–decoding strategy, a variable representation mechanism is provided to balance the exploration scale and search efficiency. To prevent the failures of worker assignment, we design a heuristic based on resource constraint matrix (RCM), which conducts a greedy search within the feasible region. For the problem-specific knowledge, a reduced insertion neighborhood and an accelerated evaluation strategy are proposed to swiftly identify the best neighborhood solution. Finally, analytical experiments show the practical value of the algorithmic components and demonstrate that VRCCEA significantly outperforms five advanced metaheuristics.
研究了一个包含工人能力约束的混合型流水车间调度问题。与大多数假设工人可以操作所有机器的现有研究不同,我们的研究解释了某些工人技能的缺乏。增加的约束大大增加了问题的复杂性,使得传统的算法无法获得可行的解。为此,建立了混合整数规划模型,并设计了一种变量表示协同进化算法(VRCCEA)来实现最大时间跨度的最小化。在分解思想的基础上,采用双种群解决多耦合问题,并通过引入解决方案存档实现种群协同进化的合作机制。针对单一编解码策略的局限性,提出了一种变量表示机制来平衡搜索规模和搜索效率。为了防止工人分配失败,设计了一种基于资源约束矩阵(RCM)的启发式算法,在可行区域内进行贪婪搜索。针对问题特定知识,提出了一种减少插入邻域和加速评估策略来快速识别最佳邻域解。最后,分析实验证明了算法组件的实用价值,并表明VRCCEA显著优于五种先进的元启发式算法。
{"title":"Scheduling a Constrained Hybrid Flowshop Using a Variable Representation Cooperative Co-Evolutionary Algorithm","authors":"Bing-Tao Wang ,&nbsp;Quan-Ke Pan ,&nbsp;Shengxiang Yang ,&nbsp;Xue-Lei Jing ,&nbsp;WeiMin Li","doi":"10.1016/j.eswa.2026.131136","DOIUrl":"10.1016/j.eswa.2026.131136","url":null,"abstract":"<div><div>The research addresses a hybrid flowshop scheduling problem incorporating worker competency constraints. Unlike most existing studies that assume workers can operate all machines, our work accounts for the absence of certain worker skills. The added constraints substantially increase the problem’s complexity, rendering traditional algorithms inadequate for obtaining feasible solutions. Therefore, a mixed-integer programming model is formulated, and a variable representation cooperative co-evolutionary algorithm (VRCCEA) is designed to achieve makespan minimization. Based on the decomposition idea, we use two populations to address the multi-coupled problem and implement a cooperative mechanism by introducing a solution archive to promote the co-evolution of populations. Given the limitations of a single encoding–decoding strategy, a variable representation mechanism is provided to balance the exploration scale and search efficiency. To prevent the failures of worker assignment, we design a heuristic based on resource constraint matrix (RCM), which conducts a greedy search within the feasible region. For the problem-specific knowledge, a reduced insertion neighborhood and an accelerated evaluation strategy are proposed to swiftly identify the best neighborhood solution. Finally, analytical experiments show the practical value of the algorithmic components and demonstrate that VRCCEA significantly outperforms five advanced metaheuristics.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"308 ","pages":"Article 131136"},"PeriodicalIF":7.5,"publicationDate":"2026-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145980534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust UAV trajectory prediction under diverse disturbances via teacher-student framework 基于师生框架的多干扰下无人机弹道鲁棒预测
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-12 DOI: 10.1016/j.eswa.2026.131185
Haoxiang Lei, Jianbo Su
Accurate trajectory prediction of unmanned aerial vehicles (UAVs) is crucial for effective anti-UAV defense. However, existing methods are typically developed under ideal conditions and fail to maintain robustness under diverse disturbances. To address this challenge, we propose a Teacher-Student framework for UAV state estimation and trajectory forecasting that enhances reliability across diverse disturbances. The framework integrates diffusion-based denoising and audio-visual feature fusion to extract robust motion states, while pseudo-state supervision is derived from kinematic modeling and CAD-guided pose estimation. Experimental results demonstrate that our method consistently outperforms state-of-the-art baselines across both ideal and disturbed scenarios, achieving accurate long-horizon predictions essential for real-world anti-UAV applications. Code will be released to support future research in robust anti-UAV systems at https://github.com/hxlei0827/Robust-Anti-UAV-Under-Diverse-Disturbances.
准确的无人机弹道预测是有效的反无人机防御的关键。然而,现有的方法通常是在理想条件下开发的,不能在各种干扰下保持鲁棒性。为了应对这一挑战,我们提出了一种用于无人机状态估计和轨迹预测的师生框架,该框架提高了不同干扰下的可靠性。该框架集成了基于扩散的去噪和视听特征融合来提取鲁棒运动状态,同时通过运动学建模和cad引导的姿态估计派生出伪状态监督。实验结果表明,我们的方法在理想和干扰情况下始终优于最先进的基线,实现了对现实世界反无人机应用至关重要的准确的长期预测。代码将在https://github.com/hxlei0827/Robust-Anti-UAV-Under-Diverse-Disturbances上发布,以支持鲁棒反无人机系统的未来研究。
{"title":"Robust UAV trajectory prediction under diverse disturbances via teacher-student framework","authors":"Haoxiang Lei,&nbsp;Jianbo Su","doi":"10.1016/j.eswa.2026.131185","DOIUrl":"10.1016/j.eswa.2026.131185","url":null,"abstract":"<div><div>Accurate trajectory prediction of unmanned aerial vehicles (UAVs) is crucial for effective anti-UAV defense. However, existing methods are typically developed under ideal conditions and fail to maintain robustness under diverse disturbances. To address this challenge, we propose a Teacher-Student framework for UAV state estimation and trajectory forecasting that enhances reliability across diverse disturbances. The framework integrates diffusion-based denoising and audio-visual feature fusion to extract robust motion states, while pseudo-state supervision is derived from kinematic modeling and CAD-guided pose estimation. Experimental results demonstrate that our method consistently outperforms state-of-the-art baselines across both ideal and disturbed scenarios, achieving accurate long-horizon predictions essential for real-world anti-UAV applications. Code will be released to support future research in robust anti-UAV systems at <span><span>https://github.com/hxlei0827/Robust-Anti-UAV-Under-Diverse-Disturbances</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"308 ","pages":"Article 131185"},"PeriodicalIF":7.5,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145980574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning and predicting traffic conflicts in mixed traffic: A spatiotemporal graph neural network with manifold similarity learning 混合交通中交通冲突的学习与预测:一种具有流形相似学习的时空图神经网络
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-12 DOI: 10.1016/j.eswa.2026.131183
Zongshi Liu , Guojian Zou , Ting Wang , Meiting Tu , Hongwei Wang , Ye Li
The coexistence of connected and automated vehicles (CAVs) and human-driven vehicles (HDVs) introduces complex non-linear dynamics, characterized by stop-and-go wave noise and velocity separation, making real-time safety risk assessment difficult. Current research on crash/conflict prediction in mixed CAV-HDV traffic remains limited, existing risk assessment models, which predominantly rely on linear Euclidean distances or instantaneous feature similarity, often misinterpret non-conflict fluctuations as crash precursors, resulting in unstable performance and high false alarm rates. To address this, we propose a Manifold Similarity Spatiotemporal Graph Network (MS-STGNet) tailored for robust real-time conflict prediction in mixed freeway traffic. Unlike distinguishing traffic states in a linear space, this model constructs a manifold-based traffic-state similarity graph to capture the intrinsic geometric structure of traffic evolution. It integrates physical adjacency with semantic neighbors and combines residual feature extraction, temporal convolution, and an adaptive fusion gate to learn spatiotemporal risk patterns. We evaluated the framework’s performance under mixed traffic scenarios with varying penetration rates of CAVs and HDVs. The experimental results demonstrate that MS-STGNet achieves consistently exceptional and stable performance across varying market penetration levels and traffic scenarios. Compared to state-of-the-art baseline models, it delivers higher predictive accuracy and substantially lower false alarm rates. The methodologies and outcomes presented in this study have the potential to be used for real-time mixed traffic control on intelligent highways and crash prevention in real-time crash risk warnings at high-risk locations.
网联自动驾驶汽车(cav)和人类驾驶汽车(HDVs)共存,引入了复杂的非线性动力学,其特征是走走停停的波噪声和速度分离,使得实时安全风险评估变得困难。目前对混合交通中碰撞/冲突预测的研究仍然有限,现有的风险评估模型主要依赖于线性欧几里得距离或瞬时特征相似性,往往将非冲突波动误认为碰撞前兆,导致性能不稳定,虚警率高。为了解决这个问题,我们提出了一个流形相似时空图网络(MS-STGNet),为混合高速公路交通的鲁棒实时冲突预测量身定制。与在线性空间中区分交通状态不同,该模型构建了一个基于流形的交通状态相似图来捕捉交通演化的内在几何结构。该方法将物理邻接与语义邻接相结合,结合残差特征提取、时间卷积和自适应融合门来学习时空风险模式。我们评估了该框架在混合流量场景下的性能,这些场景具有不同的cav和hdv渗透率。实验结果表明,MS-STGNet在不同的市场渗透水平和流量场景下都能实现持续的卓越和稳定的性能。与最先进的基线模型相比,它提供了更高的预测准确性和更低的误报率。本研究提出的方法和结果有可能用于智能高速公路的实时混合交通控制和高风险地点的实时碰撞风险预警中的碰撞预防。
{"title":"Learning and predicting traffic conflicts in mixed traffic: A spatiotemporal graph neural network with manifold similarity learning","authors":"Zongshi Liu ,&nbsp;Guojian Zou ,&nbsp;Ting Wang ,&nbsp;Meiting Tu ,&nbsp;Hongwei Wang ,&nbsp;Ye Li","doi":"10.1016/j.eswa.2026.131183","DOIUrl":"10.1016/j.eswa.2026.131183","url":null,"abstract":"<div><div>The coexistence of connected and automated vehicles (CAVs) and human-driven vehicles (HDVs) introduces complex non-linear dynamics, characterized by stop-and-go wave noise and velocity separation, making real-time safety risk assessment difficult. Current research on crash/conflict prediction in mixed CAV-HDV traffic remains limited, existing risk assessment models, which predominantly rely on linear Euclidean distances or instantaneous feature similarity, often misinterpret non-conflict fluctuations as crash precursors, resulting in unstable performance and high false alarm rates. To address this, we propose a Manifold Similarity Spatiotemporal Graph Network (MS-STGNet) tailored for robust real-time conflict prediction in mixed freeway traffic. Unlike distinguishing traffic states in a linear space, this model constructs a manifold-based traffic-state similarity graph to capture the intrinsic geometric structure of traffic evolution. It integrates physical adjacency with semantic neighbors and combines residual feature extraction, temporal convolution, and an adaptive fusion gate to learn spatiotemporal risk patterns. We evaluated the framework’s performance under mixed traffic scenarios with varying penetration rates of CAVs and HDVs. The experimental results demonstrate that MS-STGNet achieves consistently exceptional and stable performance across varying market penetration levels and traffic scenarios. Compared to state-of-the-art baseline models, it delivers higher predictive accuracy and substantially lower false alarm rates. The methodologies and outcomes presented in this study have the potential to be used for real-time mixed traffic control on intelligent highways and crash prevention in real-time crash risk warnings at high-risk locations.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"309 ","pages":"Article 131183"},"PeriodicalIF":7.5,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145981844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A cross-channel color image encryption scheme based on a novel 4D hyperchaotic system and 3-layer Peano curve 基于新型四维超混沌系统和三层Peano曲线的跨通道彩色图像加密方案
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-12 DOI: 10.1016/j.eswa.2026.131145
Zhuo Chen , Zhong Chen , Bofeng Long , Tongzhe Liu , Ming Yao
Currently, many color image encryption algorithms suffer from weak inter-channel interactions and lack sensitivity in their permutation and diffusion mechanisms. Therefore, this paper presents a novel 4D hyperchaotic system along with two types of 3-layer Peano curves (3LPC), based on which an efficient and secure cross-channel color image encryption scheme is proposed. Firstly, we incorporate the absolute value term and implement nonlinear coupling within the hyperchaotic system, which is proven to have complex dynamic behavior and a wide hyperchaotic interval through attractor diagrams, Lyapunov exponent spectrums, bifurcation diagrams, Poincaré section diagrams, and NIST test. Secondly, to overcome the challenges of cross-channel coverage and low utilization in current space-filling curves used for image encryption, we creatively propose the 3LPC that corresponds to the three-channel physical structure of color images. Based on different application scenarios, we design two distinct adaptation strategies to efficiently meet the encryption requirements of various stages in the encryption process. Finally, we propose a cross-channel technique with cross-channel random permutation and bidirectional spatial diffusion. By optimizing the 3LPC, the RGB channels of the color image are interconnected, ensuring that the influence of pixel encryption extends beyond individual channels to encompass the entire image structure. Meanwhile, the chaotic matrix generated by the 4D hyperchaotic system guarantees randomness in the encryption process, significantly increasing key sensitivity and key space. Simulation results and security analyses demonstrate that the proposed encryption scheme exhibits excellent permutation and diffusion properties, effectively resisting a range of illegal attacks.
目前,许多彩色图像加密算法的通道间交互作用较弱,排列和扩散机制缺乏敏感性。为此,本文提出了一种新颖的4D超混沌系统和两种3层Peano曲线(3LPC),并在此基础上提出了一种高效、安全的跨通道彩色图像加密方案。首先,通过吸引子图、李雅普诺夫指数谱、分岔图、poincar截面图和NIST测试,证明了超混沌系统具有复杂的动态行为和宽的超混沌区间。其次,为了克服当前用于图像加密的空间填充曲线的跨通道覆盖和低利用率的挑战,我们创造性地提出了对应于彩色图像的三通道物理结构的3LPC。针对不同的应用场景,我们设计了两种不同的自适应策略,以有效满足加密过程中各个阶段的加密需求。最后,我们提出了一种跨通道随机排列和双向空间扩散的跨通道技术。通过优化3LPC,彩色图像的RGB通道相互连接,确保像素加密的影响从单个通道扩展到整个图像结构。同时,4D超混沌系统生成的混沌矩阵保证了加密过程的随机性,显著提高了密钥灵敏度和密钥空间。仿真结果和安全性分析表明,该加密方案具有良好的排列和扩散特性,能够有效抵御各种非法攻击。
{"title":"A cross-channel color image encryption scheme based on a novel 4D hyperchaotic system and 3-layer Peano curve","authors":"Zhuo Chen ,&nbsp;Zhong Chen ,&nbsp;Bofeng Long ,&nbsp;Tongzhe Liu ,&nbsp;Ming Yao","doi":"10.1016/j.eswa.2026.131145","DOIUrl":"10.1016/j.eswa.2026.131145","url":null,"abstract":"<div><div>Currently, many color image encryption algorithms suffer from weak inter-channel interactions and lack sensitivity in their permutation and diffusion mechanisms. Therefore, this paper presents a novel 4D hyperchaotic system along with two types of 3-layer Peano curves (3LPC), based on which an efficient and secure cross-channel color image encryption scheme is proposed. Firstly, we incorporate the absolute value term and implement nonlinear coupling within the hyperchaotic system, which is proven to have complex dynamic behavior and a wide hyperchaotic interval through attractor diagrams, Lyapunov exponent spectrums, bifurcation diagrams, Poincaré section diagrams, and NIST test. Secondly, to overcome the challenges of cross-channel coverage and low utilization in current space-filling curves used for image encryption, we creatively propose the 3LPC that corresponds to the three-channel physical structure of color images. Based on different application scenarios, we design two distinct adaptation strategies to efficiently meet the encryption requirements of various stages in the encryption process. Finally, we propose a cross-channel technique with cross-channel random permutation and bidirectional spatial diffusion. By optimizing the 3LPC, the RGB channels of the color image are interconnected, ensuring that the influence of pixel encryption extends beyond individual channels to encompass the entire image structure. Meanwhile, the chaotic matrix generated by the 4D hyperchaotic system guarantees randomness in the encryption process, significantly increasing key sensitivity and key space. Simulation results and security analyses demonstrate that the proposed encryption scheme exhibits excellent permutation and diffusion properties, effectively resisting a range of illegal attacks.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"309 ","pages":"Article 131145"},"PeriodicalIF":7.5,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145981845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal self-powered sensing and deep learning for posture and attention assessment in classrooms 教室中姿态和注意力评估的多模态自供电传感和深度学习
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-12 DOI: 10.1016/j.eswa.2026.131176
Junjie Tang, Manyun Zhang, Yutong Wang, Zhiyuan Zhu
Maintaining healthy posture and attention is essential for effective learning, yet conventional monitoring systems are often costly and limited to single modalities. This study presents a multimodal expert system integrating self-powered triboelectric sensors and visual information for real-time classroom monitoring. A smart seat cushion generates an open-circuit voltage of 168 V under 20 N, enabling stable, energy-autonomous signal acquisition. The triboelectric branch employs a CNN–LSTM–STAN architecture to capture temporal patterns and salient features, achieving 96.4% accuracy in classifying seven sitting postures. The visual branch utilizes MobileNetV3 integrated with CBAM to efficiently extract facial features. Subsequently, features from both the triboelectric and visual branches are fused via concatenation for joint posture and attention assessment, achieving an attention recognition accuracy of 93.2%. These results demonstrate the effectiveness of combining self-powered sensing with deep learning-based multimodal analysis for robust, real-time classroom behavioral assessment.
保持健康的姿势和注意力对有效学习至关重要,但传统的监测系统往往成本高昂,且仅限于单一模式。本研究提出了一种集成自供电摩擦电传感器和视觉信息的多模态专家系统,用于教室实时监控。智能坐垫在20 N下产生168v的开路电压,实现稳定的能量自主信号采集。摩擦电分支采用CNN-LSTM-STAN架构捕捉时间模式和显著特征,对7种坐姿进行分类,准确率达到96.4%。视觉分支利用MobileNetV3与CBAM集成,有效提取面部特征。随后,通过连接将摩擦电和视觉分支的特征融合在一起,用于关节姿势和注意力评估,实现了93.2%的注意力识别准确率。这些结果证明了将自供电传感与基于深度学习的多模态分析相结合,用于鲁棒的实时课堂行为评估的有效性。
{"title":"Multimodal self-powered sensing and deep learning for posture and attention assessment in classrooms","authors":"Junjie Tang,&nbsp;Manyun Zhang,&nbsp;Yutong Wang,&nbsp;Zhiyuan Zhu","doi":"10.1016/j.eswa.2026.131176","DOIUrl":"10.1016/j.eswa.2026.131176","url":null,"abstract":"<div><div>Maintaining healthy posture and attention is essential for effective learning, yet conventional monitoring systems are often costly and limited to single modalities. This study presents a multimodal expert system integrating self-powered triboelectric sensors and visual information for real-time classroom monitoring. A smart seat cushion generates an open-circuit voltage of 168 V under 20 N, enabling stable, energy-autonomous signal acquisition. The triboelectric branch employs a CNN–LSTM–STAN architecture to capture temporal patterns and salient features, achieving 96.4% accuracy in classifying seven sitting postures. The visual branch utilizes MobileNetV3 integrated with CBAM to efficiently extract facial features. Subsequently, features from both the triboelectric and visual branches are fused via concatenation for joint posture and attention assessment, achieving an attention recognition accuracy of 93.2%. These results demonstrate the effectiveness of combining self-powered sensing with deep learning-based multimodal analysis for robust, real-time classroom behavioral assessment.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"308 ","pages":"Article 131176"},"PeriodicalIF":7.5,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145980570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D shape recognition and interpretability model based on the fusion of real visual and tactile point clouds 基于真实视觉与触觉点云融合的三维形状识别与可解释性模型
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-12 DOI: 10.1016/j.eswa.2026.131200
Feihong Ma , Jia Ma , Meng Chen , Yuliang Li , JunLiang Wang
Most research on 3D object shape recognition is unimodal and lacks interpretability. To address issues of missing point cloud and poor interpretability of visual perception when acquiring the object shape information under occlusion, a novel multi-modal model based on the fusion of visual and tactile point cloud information for 3D shape recognition and interpretability is proposed. First, the experimental acquisition platform for visual and tactile point clouds is constructed, which facilitates the collection of visual and tactile point clouds of objects under self-occlusion conditions. Second, a shape recognition model based on the fusion of multiple attention mechanisms for visual and tactile point clouds has been established, which is used to extract features from the preprocessed visual and tactile point clouds. The instance and class accuracies of the Dual-VT-Multi-attention model on the self-built dataset are 80.32% and 83.32%, respectively, which are significantly higher than single visual or tactile modal. Finally, to provide an intuitive interpretation of the classification decision process of the Dual-VT-Multi-attention model in each modal, an interpretable method based on the recognition model of visual and tactile point clouds is proposed. The contribution of each point can be calculated by weighting the summation of its feature vectors, which allows the generation of the Class Attention Response Map to visualize the points that are important for the model’s classification decision. The Class Attention Response Map makes the shape recognition result of Dual-VT-Multi-attention model in each modal more transparent and interpretable.
大多数关于三维物体形状识别的研究都是单模的,缺乏可解释性。针对遮挡下物体形状信息获取过程中存在点云缺失和视觉感知可解释性差的问题,提出了一种基于视觉和触觉点云信息融合的三维形状识别和可解释性多模态模型。首先,构建了视觉触觉点云实验采集平台,实现了自遮挡条件下物体视觉触觉点云的采集。其次,建立了基于多注意机制融合的视觉触觉点云形状识别模型,用于从预处理后的视觉触觉点云中提取特征;双vt -多注意模型在自建数据集上的实例和类别准确率分别为80.32%和83.32%,显著高于单一视觉或触觉模态。最后,为了直观地解释双vt -多注意模型在各个模态下的分类决策过程,提出了一种基于视觉和触觉点云识别模型的可解释方法。每个点的贡献可以通过加权其特征向量的总和来计算,这允许生成类注意响应图来可视化对模型分类决策重要的点。类注意反应图使得双vt -多注意模型在每个模态下的形状识别结果更加透明和可解释。
{"title":"3D shape recognition and interpretability model based on the fusion of real visual and tactile point clouds","authors":"Feihong Ma ,&nbsp;Jia Ma ,&nbsp;Meng Chen ,&nbsp;Yuliang Li ,&nbsp;JunLiang Wang","doi":"10.1016/j.eswa.2026.131200","DOIUrl":"10.1016/j.eswa.2026.131200","url":null,"abstract":"<div><div>Most research on 3D object shape recognition is unimodal and lacks interpretability. To address issues of missing point cloud and poor interpretability of visual perception when acquiring the object shape information under occlusion, a novel multi-modal model based on the fusion of visual and tactile point cloud information for 3D shape recognition and interpretability is proposed. First, the experimental acquisition platform for visual and tactile point clouds is constructed, which facilitates the collection of visual and tactile point clouds of objects under self-occlusion conditions. Second, a shape recognition model based on the fusion of multiple attention mechanisms for visual and tactile point clouds has been established, which is used to extract features from the preprocessed visual and tactile point clouds. The instance and class accuracies of the Dual-VT-Multi-attention model on the self-built dataset are 80.32% and 83.32%, respectively, which are significantly higher than single visual or tactile modal. Finally, to provide an intuitive interpretation of the classification decision process of the Dual-VT-Multi-attention model in each modal, an interpretable method based on the recognition model of visual and tactile point clouds is proposed. The contribution of each point can be calculated by weighting the summation of its feature vectors, which allows the generation of the Class Attention Response Map to visualize the points that are important for the model’s classification decision. The Class Attention Response Map makes the shape recognition result of Dual-VT-Multi-attention model in each modal more transparent and interpretable.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"308 ","pages":"Article 131200"},"PeriodicalIF":7.5,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145980532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A time domain compound attention neural network for direction perception with vestibular model verification 基于前庭模型验证的方向感知时域复合注意神经网络
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-11 DOI: 10.1016/j.eswa.2026.131186
Yixin Liu , Zhihao Zhang , Lingling Wang , Li Fu , Xiaohong Liu
Vestibular perception is essential for human spatial navigation, providing vital information about motion and orientation. However, existing vestibular research overlooks how the brain dynamically interprets self-motion. We present the Time Domain Compound Attention (TDCA) network, a model that decodes directional states from electroencephalography (EEG). TDCA employs multiscale temporal convolutions to capture both transient and sustained neural dynamics. A self-attention module highlights informative spatial-feature representations, while a temporal convolution module integrates them over time. Using a dataset of vestibular direction perception with synchronized EEG and inertial measurement unit (IMU) recordings from 20 participants performing five motion states (left, right, forward, backward, and stationary), TDCA achieved 93.97% accuracy under subject-independent tenfold cross-validation. Beyond its high decoding accuracy, TDCA’s temporal predictions exhibit strong alignment with an IMU-driven vestibular model, providing biophysically grounded, dual-path validation and supporting its physiological plausibility. These findings advance brain-inspired navigation research and demonstrate the feasibility of online brain-computer interfaces under natural vestibular stimulation.
前庭知觉对于人类的空间导航是必不可少的,它提供关于运动和方向的重要信息。然而,现有的前庭研究忽略了大脑如何动态地解释自我运动。我们提出了时域复合注意(TDCA)网络,这是一个从脑电图(EEG)解码方向状态的模型。TDCA采用多尺度时间卷积来捕获瞬态和持续的神经动力学。自关注模块强调信息空间特征表示,而时间卷积模块则随时间集成它们。利用20名被试进行5种运动状态(左、右、前、后、静止)的前庭方向感知数据集和同步脑电和惯性测量单元(IMU)记录,在与被试无关的10倍交叉验证下,TDCA的准确率达到93.97%。除了高解码精度之外,TDCA的时间预测与imu驱动的前庭模型表现出强烈的一致性,提供生物物理基础,双路径验证并支持其生理合理性。这些发现推动了脑启发导航研究,并证明了在自然前庭刺激下在线脑机接口的可行性。
{"title":"A time domain compound attention neural network for direction perception with vestibular model verification","authors":"Yixin Liu ,&nbsp;Zhihao Zhang ,&nbsp;Lingling Wang ,&nbsp;Li Fu ,&nbsp;Xiaohong Liu","doi":"10.1016/j.eswa.2026.131186","DOIUrl":"10.1016/j.eswa.2026.131186","url":null,"abstract":"<div><div>Vestibular perception is essential for human spatial navigation, providing vital information about motion and orientation. However, existing vestibular research overlooks how the brain dynamically interprets self-motion. We present the Time Domain Compound Attention (TDCA) network, a model that decodes directional states from electroencephalography (EEG). TDCA employs multiscale temporal convolutions to capture both transient and sustained neural dynamics. A self-attention module highlights informative spatial-feature representations, while a temporal convolution module integrates them over time. Using a dataset of vestibular direction perception with synchronized EEG and inertial measurement unit (IMU) recordings from 20 participants performing five motion states (left, right, forward, backward, and stationary), TDCA achieved 93.97% accuracy under subject-independent tenfold cross-validation. Beyond its high decoding accuracy, TDCA’s temporal predictions exhibit strong alignment with an IMU-driven vestibular model, providing biophysically grounded, dual-path validation and supporting its physiological plausibility. These findings advance brain-inspired navigation research and demonstrate the feasibility of online brain-computer interfaces under natural vestibular stimulation.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"309 ","pages":"Article 131186"},"PeriodicalIF":7.5,"publicationDate":"2026-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145969396","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FTdasc: A frequency-Time domain approach with stationarity correction for multivariate time series forecasting FTdasc:一种多变量时间序列预测的频率-时域平稳性校正方法
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-11 DOI: 10.1016/j.eswa.2026.131164
Fei Hao , Xiaofeng Zhang , Yepeng Liu , Yujuan Sun , Hua Wang , Lin Yang , Ren Wang
Predicting future trends based on historical data is essential in real-world applications such as industrial energy planning and urban Traffic planning. However, due to the inherent complexity of real-world data, it often exhibits non-stationary, making it difficult for models to capture latent features and leading to a decline in forecasting performance. In this study, FTdasc is proposed to address this challenge. FTdasc combines frequency- and time-domain information for decomposition, effectively capturing long-range and short-range dependencies within the time series. Additionally, it integrates inter-channel with intra-channel information to provide a more comprehensive feature representation. More importantly, FTdasc introduces a Stationarity correction method based on temporal dependencies, which restores non-Stationary information by constraining the data distribution. Experimental results on ten benchmark datasets demonstrate that FTdasc is highly robust and effective for both long- and short-term time series forecasting. Code availability: https://github.com/hao-fei-hub/FTdasc.
在工业能源规划和城市交通规划等实际应用中,基于历史数据预测未来趋势至关重要。然而,由于现实世界数据固有的复杂性,它往往表现为非平稳,这使得模型难以捕捉潜在特征并导致预测性能下降。在本研究中,FTdasc被提出来解决这一挑战。FTdasc结合了频率和时域信息进行分解,有效地捕获了时间序列中的远程和短程依赖关系。此外,它还集成了通道间和通道内的信息,以提供更全面的特征表示。更重要的是,FTdasc引入了一种基于时间依赖性的平稳性校正方法,通过约束数据分布来恢复非平稳性信息。在10个基准数据集上的实验结果表明,FTdasc对长期和短期时间序列预测都具有高度的鲁棒性和有效性。代码可用性:https://github.com/hao-fei-hub/FTdasc。
{"title":"FTdasc: A frequency-Time domain approach with stationarity correction for multivariate time series forecasting","authors":"Fei Hao ,&nbsp;Xiaofeng Zhang ,&nbsp;Yepeng Liu ,&nbsp;Yujuan Sun ,&nbsp;Hua Wang ,&nbsp;Lin Yang ,&nbsp;Ren Wang","doi":"10.1016/j.eswa.2026.131164","DOIUrl":"10.1016/j.eswa.2026.131164","url":null,"abstract":"<div><div>Predicting future trends based on historical data is essential in real-world applications such as industrial energy planning and urban Traffic planning. However, due to the inherent complexity of real-world data, it often exhibits non-stationary, making it difficult for models to capture latent features and leading to a decline in forecasting performance. In this study, FTdasc is proposed to address this challenge. FTdasc combines frequency- and time-domain information for decomposition, effectively capturing long-range and short-range dependencies within the time series. Additionally, it integrates inter-channel with intra-channel information to provide a more comprehensive feature representation. More importantly, FTdasc introduces a Stationarity correction method based on temporal dependencies, which restores non-Stationary information by constraining the data distribution. Experimental results on ten benchmark datasets demonstrate that FTdasc is highly robust and effective for both long- and short-term time series forecasting. Code availability: <span><span>https://github.com/hao-fei-hub/FTdasc</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"308 ","pages":"Article 131164"},"PeriodicalIF":7.5,"publicationDate":"2026-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145980576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DDR-YOLO: An efficient and accurate object detection algorithm for distracted driving behaviors DDR-YOLO:一种高效、准确的分心驾驶行为目标检测算法
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-11 DOI: 10.1016/j.eswa.2026.131170
Qian Shen , Lei Zhang , Yan Zhang , Yuxiang Zhang , Shihao Liu , Yi Li
<div><div>In recent years, researchers have employed image classification and object detection methods to recognize distracted driving behaviors (DDB). Nevertheless, a comprehensive comparative analysis of these two methods within the realm of distracted driving behavior recognition (DDR) remains underexplored, resulting in most existing algorithms struggling to balance efficiency and accuracy. Therefore, based on a comparative analysis of these two methods, this paper proposes a novel DDR algorithm named DDR-YOLO inspired by YOLO11. Initially, this paper explores the method that performs better in DDR using 250,000 manually labeled images from the 100-Drivers dataset. Furthermore, the lightweight DDR-YOLO algorithm that achieves high accuracy while improving efficiency is introduced. To accurately capture both the local details and overall postural features of DDB, an innovative Neck structure called MHMS is designed along with a new feature extraction module referred to as SGHCB. To further optimize model efficiency, this paper presents an efficient spatial-reorganization upsampling (ESU) module and a novel Shared Convolution Detection head (SCDetection). ESU restructures feature information across channel and spatial dimensions through channel shuffle and spatial shift, with a significant reduction in computational complexity and loss of feature information. By introducing a dedicated detection head branch for huge targets and sharing convolutional parameters across all four branches, SCDetection achieves enhanced detection capability for oversized objects and greater computational efficiency. Additionally, an adaptive dynamic label assignment strategy is developed to enhance the discriminative ability of both high-confidence class predictions and precisely regressed bounding box coordinates, thereby improving recognition accuracy. Moreover, a novel channel pruning method termed DG-LAMP is proposed to significantly reduce the computational cost of the model. Then knowledge distillation is implemented to compensate for the accuracy loss. Experimental results reveal that on the 100-Drivers dataset, most existing lightweight classification algorithms underperform, achieving classification accuracies of only 70% to 80%, and fail to classify multiple DDB occurring at the same time. The DDR-YOLO achieves accuracies of 91.6% and 88.8% on RGB and near-infrared modalities with a computational cost of 1.2 GFLOPs, a parameter count of 0.45M and approximately 2000 FPS. In addition, generalization experiments conducted on the StateFarm dataset and our self-collected dataset achieve accuracies of 44.3% and 87.6%, respectively. Furthermore, the proposed algorithm is deployed on an NVIDIA Jetson Orin Nano 8GB platform for practical validation. In high-power mode, DDR-YOLO runs stably for extended periods with the FPS remaining at around 29, and the operating temperature stays within a normal range. These results confirm that the proposed algorithm shows outst
近年来,研究人员采用图像分类和目标检测方法来识别分心驾驶行为。然而,对这两种方法在分心驾驶行为识别(DDR)领域的全面比较分析仍未得到充分探讨,导致大多数现有算法难以平衡效率和准确性。因此,本文在对这两种方法进行比较分析的基础上,提出了一种受YOLO11启发的新型DDR算法DDR- yolo。最初,本文使用来自100-Drivers数据集的25万张手动标记的图像探索了在DDR中表现更好的方法。在此基础上,介绍了在提高效率的同时实现高精度的轻量级DDR-YOLO算法。为了准确地捕捉DDB的局部细节和整体姿势特征,设计了一种名为MHMS的创新颈部结构,以及一种名为SGHCB的新特征提取模块。为了进一步优化模型效率,本文提出了一种高效的空间重组上采样(ESU)模块和一种新的共享卷积检测头(SCDetection)。ESU通过通道shuffle和空间移位跨通道和空间维度重构特征信息,显著降低了计算复杂度和特征信息的损失。通过为大型目标引入专用检测头分支,并在所有四个分支之间共享卷积参数,SCDetection实现了对超大目标的增强检测能力和更高的计算效率。此外,开发了一种自适应动态标签分配策略,以增强高置信度类预测和精确回归的边界框坐标的判别能力,从而提高识别精度。此外,提出了一种新的通道剪枝方法DG-LAMP,大大降低了模型的计算成本。在此基础上,通过知识精馏来弥补精度损失。实验结果表明,在100-Drivers数据集上,大多数现有的轻量级分类算法表现不佳,分类准确率仅为70% ~ 80%,并且无法对同时发生的多个DDB进行分类。DDR-YOLO在RGB和近红外模式下的精度分别为91.6%和88.8%,计算成本为1.2 GFLOPs,参数计数为0.45M,约为2000 FPS。此外,在StateFarm数据集和我们自己收集的数据集上进行的概化实验,准确率分别达到44.3%和87.6%。并在NVIDIA Jetson Orin Nano 8GB平台上进行了实际验证。在高功率模式下,DDR-YOLO长时间稳定运行,FPS保持在29左右,工作温度保持在正常范围内。这些结果证实了该算法在保持较高精度的同时,在模型大小和实时性方面表现出了出色的性能。
{"title":"DDR-YOLO: An efficient and accurate object detection algorithm for distracted driving behaviors","authors":"Qian Shen ,&nbsp;Lei Zhang ,&nbsp;Yan Zhang ,&nbsp;Yuxiang Zhang ,&nbsp;Shihao Liu ,&nbsp;Yi Li","doi":"10.1016/j.eswa.2026.131170","DOIUrl":"10.1016/j.eswa.2026.131170","url":null,"abstract":"&lt;div&gt;&lt;div&gt;In recent years, researchers have employed image classification and object detection methods to recognize distracted driving behaviors (DDB). Nevertheless, a comprehensive comparative analysis of these two methods within the realm of distracted driving behavior recognition (DDR) remains underexplored, resulting in most existing algorithms struggling to balance efficiency and accuracy. Therefore, based on a comparative analysis of these two methods, this paper proposes a novel DDR algorithm named DDR-YOLO inspired by YOLO11. Initially, this paper explores the method that performs better in DDR using 250,000 manually labeled images from the 100-Drivers dataset. Furthermore, the lightweight DDR-YOLO algorithm that achieves high accuracy while improving efficiency is introduced. To accurately capture both the local details and overall postural features of DDB, an innovative Neck structure called MHMS is designed along with a new feature extraction module referred to as SGHCB. To further optimize model efficiency, this paper presents an efficient spatial-reorganization upsampling (ESU) module and a novel Shared Convolution Detection head (SCDetection). ESU restructures feature information across channel and spatial dimensions through channel shuffle and spatial shift, with a significant reduction in computational complexity and loss of feature information. By introducing a dedicated detection head branch for huge targets and sharing convolutional parameters across all four branches, SCDetection achieves enhanced detection capability for oversized objects and greater computational efficiency. Additionally, an adaptive dynamic label assignment strategy is developed to enhance the discriminative ability of both high-confidence class predictions and precisely regressed bounding box coordinates, thereby improving recognition accuracy. Moreover, a novel channel pruning method termed DG-LAMP is proposed to significantly reduce the computational cost of the model. Then knowledge distillation is implemented to compensate for the accuracy loss. Experimental results reveal that on the 100-Drivers dataset, most existing lightweight classification algorithms underperform, achieving classification accuracies of only 70% to 80%, and fail to classify multiple DDB occurring at the same time. The DDR-YOLO achieves accuracies of 91.6% and 88.8% on RGB and near-infrared modalities with a computational cost of 1.2 GFLOPs, a parameter count of 0.45M and approximately 2000 FPS. In addition, generalization experiments conducted on the StateFarm dataset and our self-collected dataset achieve accuracies of 44.3% and 87.6%, respectively. Furthermore, the proposed algorithm is deployed on an NVIDIA Jetson Orin Nano 8GB platform for practical validation. In high-power mode, DDR-YOLO runs stably for extended periods with the FPS remaining at around 29, and the operating temperature stays within a normal range. These results confirm that the proposed algorithm shows outst","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"308 ","pages":"Article 131170"},"PeriodicalIF":7.5,"publicationDate":"2026-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145980589","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SeisMind: A domain-knowledge-informed reinforcement learning framework for intelligent control of structural seismic response SeisMind:用于结构地震反应智能控制的领域知识强化学习框架
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-11 DOI: 10.1016/j.eswa.2026.131171
Zhen Wang , Yuqing Gao , Yuzhen Chen , Zheng Lu
Active and intelligent control of structural seismic response has become a key approach for enhancing the safety and resilience of civil infrastructure. However, traditional model-based active controllers typically rely on accurate structural models and fixed control laws, which may perform poorly under complex real-world conditions. To address these limitations, this study proposes SeisMind, an end-to-end seismic control framework in which control policies are optimized using deep reinforcement learning. To handle unpredictable seismic excitations, a stochastic training environment is established by systematically sampling a wide range of near-field and far-field earthquake records, capturing diverse frequency contents, intensity levels, and seismic characteristics. To improve the robustness of the control strategy, structural uncertainties, particularly stiffness degradation, are incorporated via domain randomization during training. In addition, a physically interpretable reward function is designed to integrate structural response indicators capturing both structural and non-structural damage, as well as actuator effort. Numerical experiments demonstrate that SeisMind achieves effective control performance across linear and nonlinear structural systems, maintaining stable performance even under structural degradation. Across multiple seismic intensity levels, SeisMind exhibits more stable, self-adaptive, and less variable control performance across a range of ground motion records, outperforming conventional Linear Quadratic Regulator and H robust controllers, thereby highlighting its potential as a scalable and generalizable solution for next-generation intelligent seismic control of civil infrastructure.
主动智能控制结构地震反应已成为提高民用基础设施安全性和抗灾能力的重要途径。然而,传统的基于模型的主动控制器通常依赖于精确的结构模型和固定的控制律,在复杂的现实条件下可能表现不佳。为了解决这些限制,本研究提出了SeisMind,这是一个端到端地震控制框架,其中使用深度强化学习优化控制策略。为了处理不可预测的地震激励,通过系统地采样广泛的近场和远场地震记录,捕获不同的频率内容,强度水平和地震特征,建立随机训练环境。为了提高控制策略的鲁棒性,在训练过程中通过域随机化纳入结构不确定性,特别是刚度退化。此外,设计了一个物理可解释的奖励函数,以整合结构响应指标,捕获结构和非结构损坏,以及执行器的努力。数值实验表明,SeisMind在线性和非线性结构系统中都具有有效的控制性能,即使在结构退化的情况下也能保持稳定的性能。在多个地震烈度级别中,SeisMind在一系列地面运动记录中表现出更稳定、自适应和更少变化的控制性能,优于传统的线性二次型调节器和H∞鲁棒控制器,从而突出了其作为下一代民用基础设施智能地震控制的可扩展和可推广解决方案的潜力。
{"title":"SeisMind: A domain-knowledge-informed reinforcement learning framework for intelligent control of structural seismic response","authors":"Zhen Wang ,&nbsp;Yuqing Gao ,&nbsp;Yuzhen Chen ,&nbsp;Zheng Lu","doi":"10.1016/j.eswa.2026.131171","DOIUrl":"10.1016/j.eswa.2026.131171","url":null,"abstract":"<div><div>Active and intelligent control of structural seismic response has become a key approach for enhancing the safety and resilience of civil infrastructure. However, traditional model-based active controllers typically rely on accurate structural models and fixed control laws, which may perform poorly under complex real-world conditions. To address these limitations, this study proposes SeisMind, an end-to-end seismic control framework in which control policies are optimized using deep reinforcement learning. To handle unpredictable seismic excitations, a stochastic training environment is established by systematically sampling a wide range of near-field and far-field earthquake records, capturing diverse frequency contents, intensity levels, and seismic characteristics. To improve the robustness of the control strategy, structural uncertainties, particularly stiffness degradation, are incorporated via domain randomization during training. In addition, a physically interpretable reward function is designed to integrate structural response indicators capturing both structural and non-structural damage, as well as actuator effort. Numerical experiments demonstrate that SeisMind achieves effective control performance across linear and nonlinear structural systems, maintaining stable performance even under structural degradation. Across multiple seismic intensity levels, SeisMind exhibits more stable, self-adaptive, and less variable control performance across a range of ground motion records, outperforming conventional Linear Quadratic Regulator and <em>H</em><sub>∞</sub> robust controllers, thereby highlighting its potential as a scalable and generalizable solution for next-generation intelligent seismic control of civil infrastructure.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"309 ","pages":"Article 131171"},"PeriodicalIF":7.5,"publicationDate":"2026-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145981726","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Expert Systems with Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1