首页 > 最新文献

IET Cybersystems and Robotics最新文献

英文 中文
Adaptive Safe Braking and Distance Prediction for Overhead Cranes With Multivariation Using MLP 基于MLP的多变量桥式起重机自适应安全制动与距离预测
IF 1.2 Q3 AUTOMATION & CONTROL SYSTEMS Pub Date : 2025-04-01 DOI: 10.1049/csy2.70007
Tenglong Zhang, Guoliang Liu, Huili Chen, Guohui Tian, Qingqiang Guo

The emergency braking and braking distance prediction of an overhead crane pose challenging issues in its safe operation. This paper employs a multilayer perceptron (MLP) to implement an adaptive safe distance prediction functionality for an overhead crane with multiple variations. First, a discrete model of an overhead crane is constructed, and a model predictive control (MPC) model with angle constraints is applied for safe braking. Second, we analysed and selected the input variations of the safe distance prediction model. Subsequently, we permuted the inputs to the MLP and analysed the effect of each input on the accuracy of the MLP in predicting safety distances separately. We constructed a training dataset, and a test dataset and we optimised the safe distance prediction model through the training dataset. Finally, we conducted a comparative analysis between the MLP and nlinfit algorithms, highlighting the superiority of MLP-based adaptive safety distance prediction for bridge cranes. Experiments confirm the method's ability to ensure minimal swing angle during the entire braking process to achieve safe braking. The results underscore the practical utility and novelty of the proposed algorithm.

桥式起重机的紧急制动和制动距离预测是桥式起重机安全运行的重要课题。本文采用多层感知器(MLP)实现了多变量桥式起重机的自适应安全距离预测功能。首先,建立了桥式起重机的离散模型,并将其应用于具有角度约束的模型预测控制(MPC)模型进行安全制动。其次,对安全距离预测模型的输入变量进行了分析和选择。随后,我们对MLP的输入进行了排列,并分别分析了每个输入对MLP预测安全距离准确性的影响。我们构建了训练数据集和测试数据集,并通过训练数据集对安全距离预测模型进行了优化。最后,对MLP算法和nlinfit算法进行了对比分析,突出了基于MLP的桥式起重机自适应安全距离预测的优越性。实验验证了该方法在整个制动过程中能够保证最小的摆角,从而实现安全制动。结果表明了该算法的实用性和新颖性。
{"title":"Adaptive Safe Braking and Distance Prediction for Overhead Cranes With Multivariation Using MLP","authors":"Tenglong Zhang,&nbsp;Guoliang Liu,&nbsp;Huili Chen,&nbsp;Guohui Tian,&nbsp;Qingqiang Guo","doi":"10.1049/csy2.70007","DOIUrl":"10.1049/csy2.70007","url":null,"abstract":"<p>The emergency braking and braking distance prediction of an overhead crane pose challenging issues in its safe operation. This paper employs a multilayer perceptron (MLP) to implement an adaptive safe distance prediction functionality for an overhead crane with multiple variations. First, a discrete model of an overhead crane is constructed, and a model predictive control (MPC) model with angle constraints is applied for safe braking. Second, we analysed and selected the input variations of the safe distance prediction model. Subsequently, we permuted the inputs to the MLP and analysed the effect of each input on the accuracy of the MLP in predicting safety distances separately. We constructed a training dataset, and a test dataset and we optimised the safe distance prediction model through the training dataset. Finally, we conducted a comparative analysis between the MLP and nlinfit algorithms, highlighting the superiority of MLP-based adaptive safety distance prediction for bridge cranes. Experiments confirm the method's ability to ensure minimal swing angle during the entire braking process to achieve safe braking. The results underscore the practical utility and novelty of the proposed algorithm.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70007","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143749682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Move to See More: Approaching Object With Partial Occlusion Using Large Multimodal Model and Active Object Detection 移动以查看更多:使用大型多模态模型和活动对象检测接近部分遮挡的对象
IF 1.2 Q3 AUTOMATION & CONTROL SYSTEMS Pub Date : 2025-03-27 DOI: 10.1049/csy2.70008
Aoqi Wang, Guohui Tian, Yuhao Wang, Zhongyang Li

Active object detection (AOD) is a crucial task in the field of robotics. A key challenge in household environments for AOD is that the target object is often undetectable due to partial occlusion, which leads to the failure of traditional methods. To address the occlusion problem, this paper first proposes a novel occlusion handling method based on the large multimodal model (LMM). The method utilises an LMM to detect and analyse input RGB images and generates adjustment actions to progressively eliminate occlusion. After the occlusion is handled, an improved AOD method based on a deep Q-learning network (DQN) is used to complete the task. We introduce an attention mechanism to process image features, enabling the model to focus on critical regions of the input images. Additionally, a new reward function is proposed that comprehensively considers the bounding box of the target object and the robot's distance to the object, along with the actions performed by the robot. Experiments on the dataset and in real-world scenarios validate the effectiveness of the proposed method in performing AOD tasks under partial occlusion.

主动目标检测(AOD)是机器人领域的一项重要任务。在家庭环境中进行AOD的一个关键挑战是,由于部分遮挡,目标物体往往无法检测到,这导致传统方法的失败。针对遮挡问题,本文首先提出了一种基于大多模态模型(large multimodal model, LMM)的遮挡处理方法。该方法利用LMM检测和分析输入的RGB图像,并生成调整动作以逐步消除遮挡。在遮挡处理后,使用基于深度q -学习网络(DQN)的改进AOD方法来完成任务。我们引入了注意机制来处理图像特征,使模型能够专注于输入图像的关键区域。此外,提出了一种新的奖励函数,该函数综合考虑了目标物体的边界框、机器人与目标物体的距离以及机器人所执行的动作。在数据集和现实场景上的实验验证了该方法在部分遮挡下执行AOD任务的有效性。
{"title":"Move to See More: Approaching Object With Partial Occlusion Using Large Multimodal Model and Active Object Detection","authors":"Aoqi Wang,&nbsp;Guohui Tian,&nbsp;Yuhao Wang,&nbsp;Zhongyang Li","doi":"10.1049/csy2.70008","DOIUrl":"10.1049/csy2.70008","url":null,"abstract":"<p>Active object detection (AOD) is a crucial task in the field of robotics. A key challenge in household environments for AOD is that the target object is often undetectable due to partial occlusion, which leads to the failure of traditional methods. To address the occlusion problem, this paper first proposes a novel occlusion handling method based on the large multimodal model (LMM). The method utilises an LMM to detect and analyse input RGB images and generates adjustment actions to progressively eliminate occlusion. After the occlusion is handled, an improved AOD method based on a deep Q-learning network (DQN) is used to complete the task. We introduce an attention mechanism to process image features, enabling the model to focus on critical regions of the input images. Additionally, a new reward function is proposed that comprehensively considers the bounding box of the target object and the robot's distance to the object, along with the actions performed by the robot. Experiments on the dataset and in real-world scenarios validate the effectiveness of the proposed method in performing AOD tasks under partial occlusion.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70008","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143707626","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bioinspired framework for real-time collision detection with dynamic obstacles in cluttered outdoor environments using event cameras 生物启发的框架,实时碰撞检测与动态障碍物在混乱的室外环境中使用事件相机
IF 1.2 Q3 AUTOMATION & CONTROL SYSTEMS Pub Date : 2025-03-14 DOI: 10.1049/csy2.70006
Meriem Ben Miled, Wenwen Liu, Yuanchang Liu

In the field of robotics and visual-based navigation, event cameras are gaining popularity due to their exceptional dynamic range, low power consumption, and rapid response capabilities. These neuromorphic devices facilitate the efficient detection and avoidance of fast moving obstacles, and address common limitations of traditional hardware. However, the majority of state-of-the-art event-based algorithms still rely on conventional computer vision strategies. The goal is to shift from the standard protocols for dynamic obstacle detection by taking inspiration from the time-computational paradigm of biological vision system. In this paper, the authors present an innovative framework inspired by a biological response mechanism triggered by approaching objects, enabling the perception and identification of potential collision threats. The method, validated through both simulation and real-world experimentation, charts a new path in the application of event cameras for dynamic obstacle detection and avoidance in autonomous unmanned aerial vehicles. When compared to conventional methods, the proposed approach demonstrates a success rate of 97% in detecting obstacles within real-world outdoor settings.

在机器人和基于视觉的导航领域,事件摄像机因其出色的动态范围、低功耗和快速响应能力而越来越受欢迎。这些神经形态装置有助于有效地检测和避免快速移动的障碍物,并解决传统硬件的常见限制。然而,大多数最先进的基于事件的算法仍然依赖于传统的计算机视觉策略。目标是通过从生物视觉系统的时间计算范式中汲取灵感,从标准协议转变为动态障碍物检测。在本文中,作者提出了一个创新的框架,该框架受接近物体触发的生物反应机制的启发,能够感知和识别潜在的碰撞威胁。该方法通过仿真和实际实验验证,为事件相机在自主无人机动态障碍物检测和避障中的应用开辟了新的道路。与传统方法相比,该方法在真实室外环境中检测障碍物的成功率为97%。
{"title":"Bioinspired framework for real-time collision detection with dynamic obstacles in cluttered outdoor environments using event cameras","authors":"Meriem Ben Miled,&nbsp;Wenwen Liu,&nbsp;Yuanchang Liu","doi":"10.1049/csy2.70006","DOIUrl":"10.1049/csy2.70006","url":null,"abstract":"<p>In the field of robotics and visual-based navigation, event cameras are gaining popularity due to their exceptional dynamic range, low power consumption, and rapid response capabilities. These neuromorphic devices facilitate the efficient detection and avoidance of fast moving obstacles, and address common limitations of traditional hardware. However, the majority of state-of-the-art event-based algorithms still rely on conventional computer vision strategies. The goal is to shift from the standard protocols for dynamic obstacle detection by taking inspiration from the time-computational paradigm of biological vision system. In this paper, the authors present an innovative framework inspired by a biological response mechanism triggered by approaching objects, enabling the perception and identification of potential collision threats. The method, validated through both simulation and real-world experimentation, charts a new path in the application of event cameras for dynamic obstacle detection and avoidance in autonomous unmanned aerial vehicles. When compared to conventional methods, the proposed approach demonstrates a success rate of 97% in detecting obstacles within real-world outdoor settings.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70006","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143629855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Novel vision-LiDAR fusion framework for human action recognition based on dynamic lateral connection 基于动态横向连接的人体动作识别视觉-激光雷达融合框架
IF 1.2 Q3 AUTOMATION & CONTROL SYSTEMS Pub Date : 2024-12-31 DOI: 10.1049/csy2.70005
Fei Yan, Guangyao Jin, Zheng Mu, Shouxing Zhang, Yinghao Cai, Tao Lu, Yan Zhuang

In the past decades, substantial progress has been made in human action recognition. However, most existing studies and datasets for human action recognition utilise still images or videos as the primary modality. Image-based approaches can be easily impacted by adverse environmental conditions. In this paper, the authors propose combining RGB images and point clouds from LiDAR sensors for human action recognition. A dynamic lateral convolutional network (DLCN) is proposed to fuse features from multi-modalities. The RGB features and the geometric information from the point clouds closely interact with each other in the DLCN, which is complementary in action recognition. The experimental results on the JRDB-Act dataset demonstrate that the proposed DLCN outperforms the state-of-the-art approaches of human action recognition. The authors show the potential of the proposed DLCN in various complex scenarios, which is highly valuable in real-world applications.

在过去的几十年里,人类行为识别已经取得了实质性的进展。然而,大多数现有的人类动作识别研究和数据集利用静止图像或视频作为主要模式。基于图像的方法容易受到不利环境条件的影响。本文提出将RGB图像与LiDAR传感器的点云相结合用于人体动作识别。提出了一种融合多模态特征的动态横向卷积网络(DLCN)。在DLCN中,RGB特征与点云的几何信息相互作用密切,在动作识别中互为补充。在JRDB-Act数据集上的实验结果表明,所提出的DLCN优于最先进的人类动作识别方法。作者展示了所提出的DLCN在各种复杂场景中的潜力,这在实际应用中具有很高的价值。
{"title":"Novel vision-LiDAR fusion framework for human action recognition based on dynamic lateral connection","authors":"Fei Yan,&nbsp;Guangyao Jin,&nbsp;Zheng Mu,&nbsp;Shouxing Zhang,&nbsp;Yinghao Cai,&nbsp;Tao Lu,&nbsp;Yan Zhuang","doi":"10.1049/csy2.70005","DOIUrl":"10.1049/csy2.70005","url":null,"abstract":"<p>In the past decades, substantial progress has been made in human action recognition. However, most existing studies and datasets for human action recognition utilise still images or videos as the primary modality. Image-based approaches can be easily impacted by adverse environmental conditions. In this paper, the authors propose combining RGB images and point clouds from LiDAR sensors for human action recognition. A dynamic lateral convolutional network (DLCN) is proposed to fuse features from multi-modalities. The RGB features and the geometric information from the point clouds closely interact with each other in the DLCN, which is complementary in action recognition. The experimental results on the JRDB-Act dataset demonstrate that the proposed DLCN outperforms the state-of-the-art approaches of human action recognition. The authors show the potential of the proposed DLCN in various complex scenarios, which is highly valuable in real-world applications.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"6 4","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70005","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143121447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Big2Small: Learning from masked image modelling with heterogeneous self-supervised knowledge distillation Big2Small:基于异构自监督知识蒸馏的蒙面图像建模学习
IF 1.2 Q3 AUTOMATION & CONTROL SYSTEMS Pub Date : 2024-12-31 DOI: 10.1049/csy2.70002
Ziming Wang, Shumin Han, Xiaodi Wang, Jing Hao, Xianbin Cao, Baochang Zhang

Small convolutional neural network (CNN)-based models usually require transferring knowledge from a large model before they are deployed in computationally resource-limited edge devices. Masked image modelling (MIM) methods achieve great success in various visual tasks but remain largely unexplored in knowledge distillation for heterogeneous deep models. The reason is mainly due to the significant discrepancy between the transformer-based large model and the CNN-based small network. In this paper, the authors develop the first heterogeneous self-supervised knowledge distillation (HSKD) based on MIM, which can efficiently transfer knowledge from large transformer models to small CNN-based models in a self-supervised fashion. Our method builds a bridge between transformer-based models and CNNs by training a UNet-style student with sparse convolution, which can effectively mimic the visual representation inferred by a teacher over masked modelling. Our method is a simple yet effective learning paradigm to learn the visual representation and distribution of data from heterogeneous teacher models, which can be pre-trained using advanced self-supervised methods. Extensive experiments show that it adapts well to various models and sizes, consistently achieving state-of-the-art performance in image classification, object detection, and semantic segmentation tasks. For example, in the Imagenet 1K dataset, HSKD improves the accuracy of Resnet-50 (sparse) from 76.98% to 80.01%.

基于卷积神经网络(CNN)的小型模型在部署到计算资源有限的边缘设备之前,通常需要从大型模型中转移知识。掩膜图像建模(MIM)方法在各种视觉任务中取得了巨大的成功,但在异构深度模型的知识提炼方面仍未得到充分的探索。究其原因,主要是由于基于变压器的大模型与基于cnn的小网络存在显著差异。本文首次提出了基于MIM的异构自监督知识蒸馏(HSKD)方法,该方法能够以自监督的方式将知识从大型变压器模型高效地转移到基于cnn的小型模型中。我们的方法通过使用稀疏卷积训练unet风格的学生,在基于变压器的模型和cnn之间建立了一座桥梁,该方法可以有效地模仿教师通过掩模建模推断的视觉表示。我们的方法是一种简单而有效的学习范式,可以从异构教师模型中学习数据的可视化表示和分布,这些模型可以使用先进的自监督方法进行预训练。大量的实验表明,它可以很好地适应各种模型和尺寸,在图像分类、目标检测和语义分割任务中始终如一地实现最先进的性能。例如,在Imagenet 1K数据集中,HSKD将Resnet-50(稀疏)的准确率从76.98%提高到80.01%。
{"title":"Big2Small: Learning from masked image modelling with heterogeneous self-supervised knowledge distillation","authors":"Ziming Wang,&nbsp;Shumin Han,&nbsp;Xiaodi Wang,&nbsp;Jing Hao,&nbsp;Xianbin Cao,&nbsp;Baochang Zhang","doi":"10.1049/csy2.70002","DOIUrl":"10.1049/csy2.70002","url":null,"abstract":"<p>Small convolutional neural network (CNN)-based models usually require transferring knowledge from a large model before they are deployed in computationally resource-limited edge devices. Masked image modelling (MIM) methods achieve great success in various visual tasks but remain largely unexplored in knowledge distillation for heterogeneous deep models. The reason is mainly due to the significant discrepancy between the transformer-based large model and the CNN-based small network. In this paper, the authors develop the first heterogeneous self-supervised knowledge distillation (HSKD) based on MIM, which can efficiently transfer knowledge from large transformer models to small CNN-based models in a self-supervised fashion. Our method builds a bridge between transformer-based models and CNNs by training a UNet-style student with sparse convolution, which can effectively mimic the visual representation inferred by a teacher over masked modelling. Our method is a simple yet effective learning paradigm to learn the visual representation and distribution of data from heterogeneous teacher models, which can be pre-trained using advanced self-supervised methods. Extensive experiments show that it adapts well to various models and sizes, consistently achieving state-of-the-art performance in image classification, object detection, and semantic segmentation tasks. For example, in the Imagenet 1K dataset, HSKD improves the accuracy of Resnet-50 (sparse) from 76.98% to 80.01%.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"6 4","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70002","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143121450","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic feature-based markerless calibration and navigation method for augmented reality assisted dental treatment 基于特征的自动无标记校准和导航方法,用于增强现实辅助牙科治疗
IF 1.2 Q3 AUTOMATION & CONTROL SYSTEMS Pub Date : 2024-12-31 DOI: 10.1049/csy2.70003
Faizan Ahmad, Jing Xiong, Zeyang Xia

Augmented reality (AR) is gaining traction in the field of computer-assisted treatment (CAT). Head-mounted display (HMD)-based AR in CAT provides dentists with enhanced visualisation by directly overlaying a three-dimensional (3D) model on a real patient during dental treatment. However, conventional AR-based treatments rely on optical markers and trackers, which makes them tedious, expensive, and uncomfortable for dentists. Therefore, a markerless image-to-patient tracking system is necessary to overcome these challenges and enhance system efficiency. This paper proposes a novel feature-based markerless calibration and navigation method for an HMD-based AR visualisation system. The authors address three sub-challenges: firstly, synthetic RGB-D data for anatomical landmark detection is generated to train a deep convolutional neural network (DCNN); secondly, the HMD is automatically calibrated using detected anatomical landmarks, eliminating the need for user input or optical trackers; and thirdly, a multi-iterative closest point (ICP) algorithm is developed for effective 3D-3D real-time navigation. The authors conduct several experiments on a commercially available HMD (HoloLens 2). Finally, the authors compare and evaluate the approach against state-of-the-art methods that employ HoloLens. The proposed method achieves a calibration virtual-to-real re-projection distance of (1.09 ± 0.23) mm and navigation projection errors and accuracies of approximately (0.53 ± 0.19) mm and 93.87%, respectively.

增强现实技术(AR)在计算机辅助治疗(CAT)领域正获得越来越多的关注。CAT中基于头戴式显示器(HMD)的AR通过在牙科治疗期间直接将三维(3D)模型叠加在真实患者上,为牙医提供增强的可视化。然而,传统的基于ar的治疗依赖于光学标记和跟踪器,这使得它们对牙医来说既繁琐又昂贵,而且不舒服。因此,一个无标记的图像到患者跟踪系统是必要的,以克服这些挑战,提高系统效率。针对基于hmd的AR可视化系统,提出了一种新的基于特征的无标记标定与导航方法。首先,生成用于解剖地标检测的合成RGB-D数据来训练深度卷积神经网络(DCNN);其次,HMD使用检测到的解剖标志自动校准,无需用户输入或光学跟踪器;第三,提出了一种有效的3D-3D实时导航的多迭代最近点(ICP)算法。作者在商用HMD (HoloLens 2)上进行了几个实验。最后,作者将该方法与采用HoloLens的最先进方法进行了比较和评估。该方法标定虚实重投影距离为(1.09±0.23)mm,导航投影误差和精度分别约为(0.53±0.19)mm和93.87%。
{"title":"Automatic feature-based markerless calibration and navigation method for augmented reality assisted dental treatment","authors":"Faizan Ahmad,&nbsp;Jing Xiong,&nbsp;Zeyang Xia","doi":"10.1049/csy2.70003","DOIUrl":"10.1049/csy2.70003","url":null,"abstract":"<p>Augmented reality (AR) is gaining traction in the field of computer-assisted treatment (CAT). Head-mounted display (HMD)-based AR in CAT provides dentists with enhanced visualisation by directly overlaying a three-dimensional (3D) model on a real patient during dental treatment. However, conventional AR-based treatments rely on optical markers and trackers, which makes them tedious, expensive, and uncomfortable for dentists. Therefore, a markerless image-to-patient tracking system is necessary to overcome these challenges and enhance system efficiency. This paper proposes a novel feature-based markerless calibration and navigation method for an HMD-based AR visualisation system. The authors address three sub-challenges: firstly, synthetic RGB-D data for anatomical landmark detection is generated to train a deep convolutional neural network (DCNN); secondly, the HMD is automatically calibrated using detected anatomical landmarks, eliminating the need for user input or optical trackers; and thirdly, a multi-iterative closest point (ICP) algorithm is developed for effective 3D-3D real-time navigation. The authors conduct several experiments on a commercially available HMD (HoloLens 2). Finally, the authors compare and evaluate the approach against state-of-the-art methods that employ HoloLens. The proposed method achieves a calibration virtual-to-real re-projection distance of (1.09 ± 0.23) mm and navigation projection errors and accuracies of approximately (0.53 ± 0.19) mm and 93.87%, respectively.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"6 4","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70003","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143121449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing stability and safety: A novel multi-constraint model predictive control approach for forklift trajectory 提高稳定性和安全性:一种新的多约束模型预测控制方法的叉车轨迹
IF 1.2 Q3 AUTOMATION & CONTROL SYSTEMS Pub Date : 2024-12-31 DOI: 10.1049/csy2.70004
Yizhen Sun, Junyou Yang, Donghui Zhao, Moses Chukwuka Okonkwo, Jianmin Zhang, Shuoyu Wang, Yang Liu

The advancements in intelligent manufacturing have made high-precision trajectory tracking technology crucial for improving the efficiency and safety of in-factory cargo transportation. This study addresses the limitations of current forklift navigation systems in trajectory control accuracy and stability by proposing the Enhanced Stability and Safety Model Predictive Control (ESS-MPC) method. This approach includes a multi-constraint strategy for improved stability and safety. The kinematic model for a single front steering-wheel forklift vehicle is constructed with all known state quantities, including the steering angle, resulting in a more accurate model description and trajectory prediction. To ensure vehicle safety, the spatial safety boundary obtained from the trajectory planning module is established as a hard constraint for ESS-MPC tracking. The optimisation constraints are also updated with the key kinematic and dynamic parameters of the forklift. The ESS-MPC method improved the position and pose accuracy and stability by 57.93%, 37.83%, and 57.51%, respectively, as demonstrated through experimental validation using simulation and real-world environments. This study provides significant support for the development of autonomous navigation systems for industrial forklifts.

智能制造的进步使得高精度轨迹跟踪技术成为提高工厂内货物运输效率和安全性的关键。针对当前叉车导航系统在轨迹控制精度和稳定性方面的局限性,提出了增强稳定性与安全模型预测控制(ESS-MPC)方法。该方法包括一个多约束策略,以提高稳定性和安全性。利用已知的所有状态量,包括转向角,建立了单前轮叉车的运动学模型,使模型描述和轨迹预测更加准确。为保证车辆安全,建立轨迹规划模块得到的空间安全边界作为ESS-MPC跟踪的硬约束。优化约束还更新了叉车的关键运动学和动力学参数。通过仿真和现实环境的实验验证,ESS-MPC方法的位姿精度和稳定性分别提高了57.93%、37.83%和57.51%。该研究为工业叉车自主导航系统的开发提供了重要的支持。
{"title":"Enhancing stability and safety: A novel multi-constraint model predictive control approach for forklift trajectory","authors":"Yizhen Sun,&nbsp;Junyou Yang,&nbsp;Donghui Zhao,&nbsp;Moses Chukwuka Okonkwo,&nbsp;Jianmin Zhang,&nbsp;Shuoyu Wang,&nbsp;Yang Liu","doi":"10.1049/csy2.70004","DOIUrl":"10.1049/csy2.70004","url":null,"abstract":"<p>The advancements in intelligent manufacturing have made high-precision trajectory tracking technology crucial for improving the efficiency and safety of in-factory cargo transportation. This study addresses the limitations of current forklift navigation systems in trajectory control accuracy and stability by proposing the Enhanced Stability and Safety Model Predictive Control (ESS-MPC) method. This approach includes a multi-constraint strategy for improved stability and safety. The kinematic model for a single front steering-wheel forklift vehicle is constructed with all known state quantities, including the steering angle, resulting in a more accurate model description and trajectory prediction. To ensure vehicle safety, the spatial safety boundary obtained from the trajectory planning module is established as a hard constraint for ESS-MPC tracking. The optimisation constraints are also updated with the key kinematic and dynamic parameters of the forklift. The ESS-MPC method improved the position and pose accuracy and stability by 57.93%, 37.83%, and 57.51%, respectively, as demonstrated through experimental validation using simulation and real-world environments. This study provides significant support for the development of autonomous navigation systems for industrial forklifts.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"6 4","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70004","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143121448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D-printed biomimetic and bioinspired soft actuators 三维打印仿生和生物启发软致动器
IF 1.2 Q3 AUTOMATION & CONTROL SYSTEMS Pub Date : 2024-11-10 DOI: 10.1049/csy2.70001
Sonja S. Sparks, Alejandro G. Obando, Yizong Li, Si Chen, Shanshan Yao, Kaiyan Qiu

A major intent of scientific research is the replication of the behaviour observed in natural spaces. In robotics, these can be through biomimetic movements in devices and inspiration from diverse actions in nature, also known as bioinspired features. An interesting pathway enabling both features is the fabrication of soft actuators. Specifically, 3D-printing has been explored as a potential approach for the development of biomimetic and bioinspired soft actuators. The extent of this method is highlighted through the large array of applications and techniques used to create these devices, as applications from the movement of fern trees to contraction in organs are explored. In this review, different 3D-printing fabrication methods, materials, and types of soft actuators, and their respective applications are discussed in depth. Finally, the extent of their use for present operations and future technological advances are discussed.

科学研究的一个主要目的是复制在自然空间中观察到的行为。在机器人学中,这可以通过设备中的生物仿生运动和从自然界的各种行为中获得灵感来实现,这也被称为生物启发功能。实现这两种功能的一个有趣途径是制造软致动器。具体来说,三维打印技术已被视为开发仿生物和生物启发软致动器的一种潜在方法。从蕨类植物的运动到器官的收缩,大量的应用和技术被用来制造这些设备,从而凸显了这种方法的广泛性。本综述将深入讨论不同的 3D 打印制造方法、材料、软致动器类型及其各自的应用。最后,还讨论了它们在当前操作中的应用范围以及未来的技术进步。
{"title":"3D-printed biomimetic and bioinspired soft actuators","authors":"Sonja S. Sparks,&nbsp;Alejandro G. Obando,&nbsp;Yizong Li,&nbsp;Si Chen,&nbsp;Shanshan Yao,&nbsp;Kaiyan Qiu","doi":"10.1049/csy2.70001","DOIUrl":"10.1049/csy2.70001","url":null,"abstract":"<div>\u0000 \u0000 \u0000 <section>\u0000 \u0000 <p>A major intent of scientific research is the replication of the behaviour observed in natural spaces. In robotics, these can be through biomimetic movements in devices and inspiration from diverse actions in nature, also known as bioinspired features. An interesting pathway enabling both features is the fabrication of soft actuators. Specifically, 3D-printing has been explored as a potential approach for the development of biomimetic and bioinspired soft actuators. The extent of this method is highlighted through the large array of applications and techniques used to create these devices, as applications from the movement of fern trees to contraction in organs are explored. In this review, different 3D-printing fabrication methods, materials, and types of soft actuators, and their respective applications are discussed in depth. Finally, the extent of their use for present operations and future technological advances are discussed.</p>\u0000 </section>\u0000 </div>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"6 4","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-11-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70001","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142664904","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Correction-enabled reversible data hiding with pixel repetition for high embedding rate and quality preservation 利用像素重复校正功能进行可逆数据隐藏,实现高嵌入率和质量保证
IF 1.2 Q3 AUTOMATION & CONTROL SYSTEMS Pub Date : 2024-09-30 DOI: 10.1049/csy2.70000
Mohammad Ali Kawser, Hussain Nyeem, Md Abdul Wahed

A novel correction-enabled Pixel Repetition (PR)-based Reversible Data Hiding (RDH) framework, featuring a new embedding scheme is presented. The proposed RDH scheme uses contextually redundant block pixels, generated via PR, in a two-phase adaptive embedding process, enhancing both image quality and data embedding rates. Specifically, each 2×2 $2times 2$ block encodes 4 bits of data using new mapping conditions that facilitate seed pixel reconstruction from remaining block pixels and provide additional embedding opportunities. Additionally, an innovative post-embedding error correction technique, based on 2k ${2}^{k}$-bit error-correction, minimises post-embedding distortion, further improving image quality. This error correction approach augments data embedding robustness, vital for applications like medical imaging, telemedicine, and digital watermarking that requires high embedding capacity with minimum possible distortion. The proposed scheme surpasses existing state-of-the-art methods in embedding rate-distortion performance, validated through subjective and objective analyses. Furthermore, statistical analysis, including histogram and fragility testing, confirms the scheme's potential for image authentication across diverse multimedia applications. The correction-enabled RDH with PR offers enhanced embedding capacity and image quality preservation, making it particularly advantageous for applications requiring robust data hiding while maintaining visual fidelity.

本文介绍了一种基于像素重复(PR)校正的新型可逆数据隐藏(RDH)框架,它采用了一种新的嵌入方案。所提出的 RDH 方案在两阶段自适应嵌入过程中使用了通过 PR 生成的上下文冗余块像素,从而提高了图像质量和数据嵌入率。具体来说,每个 2 × 2 2 次 2$ 块使用新的映射条件编码 4 比特数据,这有利于从剩余块像素重建种子像素,并提供额外的嵌入机会。此外,基于 2 k ${2}^{k}$ -比特纠错的创新嵌入后纠错技术最大限度地减少了嵌入后失真,进一步提高了图像质量。这种纠错方法增强了数据嵌入的鲁棒性,对于医学成像、远程医疗和数字水印等要求高嵌入容量和最小失真度的应用至关重要。通过主观和客观分析验证,所提出的方案在嵌入率-失真性能方面超越了现有的最先进方法。此外,包括直方图和脆性测试在内的统计分析也证实了该方案在各种多媒体应用中进行图像认证的潜力。带有 PR 的校正 RDH 可提供更强的嵌入能力和图像质量保证,因此特别适用于需要在保持视觉保真度的同时进行稳健数据隐藏的应用。
{"title":"Correction-enabled reversible data hiding with pixel repetition for high embedding rate and quality preservation","authors":"Mohammad Ali Kawser,&nbsp;Hussain Nyeem,&nbsp;Md Abdul Wahed","doi":"10.1049/csy2.70000","DOIUrl":"10.1049/csy2.70000","url":null,"abstract":"<p>A novel correction-enabled Pixel Repetition (PR)-based Reversible Data Hiding (RDH) framework, featuring a new embedding scheme is presented. The proposed RDH scheme uses contextually redundant block pixels, generated via PR, in a two-phase adaptive embedding process, enhancing both image quality and data embedding rates. Specifically, each <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mrow>\u0000 <mrow>\u0000 <mn>2</mn>\u0000 <mo>×</mo>\u0000 <mn>2</mn>\u0000 </mrow>\u0000 </mrow>\u0000 </mrow>\u0000 <annotation> $2times 2$</annotation>\u0000 </semantics></math> block encodes 4 bits of data using new mapping conditions that facilitate seed pixel reconstruction from remaining block pixels and provide additional embedding opportunities. Additionally, an innovative post-embedding error correction technique, based on <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mrow>\u0000 <mrow>\u0000 <msup>\u0000 <mn>2</mn>\u0000 <mi>k</mi>\u0000 </msup>\u0000 </mrow>\u0000 </mrow>\u0000 </mrow>\u0000 <annotation> ${2}^{k}$</annotation>\u0000 </semantics></math>-bit error-correction, minimises post-embedding distortion, further improving image quality. This error correction approach augments data embedding robustness, vital for applications like medical imaging, telemedicine, and digital watermarking that requires high embedding capacity with minimum possible distortion. The proposed scheme surpasses existing state-of-the-art methods in embedding rate-distortion performance, validated through subjective and objective analyses. Furthermore, statistical analysis, including histogram and fragility testing, confirms the scheme's potential for image authentication across diverse multimedia applications. The correction-enabled RDH with PR offers enhanced embedding capacity and image quality preservation, making it particularly advantageous for applications requiring robust data hiding while maintaining visual fidelity.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"6 3","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70000","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142360031","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Anti-sloshing control: Flatness-based trajectory planning and tracking control with an integrated extended state observer 防滑控制:基于平整度的轨迹规划和跟踪控制与综合扩展状态观测器
IF 1.2 Q3 AUTOMATION & CONTROL SYSTEMS Pub Date : 2024-08-25 DOI: 10.1049/csy2.12121
Khanh Nguyen Viet, Minh Do Duc, Thanh Cao Duc, Tung Lam Nguyen

The phenomenon of sloshing causes a significantly negative impact on a wide range of industries. A time-optimal flatness-based trajectory planning and Lyapunov-based model predictive control (LMPC) is proposed for trajectory tracking of a transmitting cylindrical container filled with liquid. Firstly, this research presents an equivalent discrete model based on a mass-spring-damper system. Subsequently, after the flatness of the adopted non-linear model for 2D is established, time-optimal trajectories are introduced. A control method called LMPC is shown to solve the problem of orbital tracking, which allows setting limits for state variables. In addition, to ensure system performance, a linear extended state observer (LESO) is integrated to cope with system uncertainties. Finally, the efficiency of the proposed approach for liquid sloshing suppression and tracking is illustrated by simulations.

荡气现象给各行各业带来了极大的负面影响。针对装满液体的传输圆柱形容器的轨迹跟踪,提出了一种基于时间最优平面度的轨迹规划和基于李亚普诺夫的模型预测控制(LMPC)。首先,本研究提出了一个基于质量-弹簧-阻尼系统的等效离散模型。随后,在建立了所采用的二维非线性模型的平面性之后,引入了时间最优轨迹。一种名为 LMPC 的控制方法被用于解决轨道跟踪问题,它允许为状态变量设置限制。此外,为确保系统性能,还集成了线性扩展状态观测器(LESO),以应对系统的不确定性。最后,通过仿真说明了所提方法在液体荡浮抑制和跟踪方面的效率。
{"title":"Anti-sloshing control: Flatness-based trajectory planning and tracking control with an integrated extended state observer","authors":"Khanh Nguyen Viet,&nbsp;Minh Do Duc,&nbsp;Thanh Cao Duc,&nbsp;Tung Lam Nguyen","doi":"10.1049/csy2.12121","DOIUrl":"10.1049/csy2.12121","url":null,"abstract":"<p>The phenomenon of sloshing causes a significantly negative impact on a wide range of industries. A time-optimal flatness-based trajectory planning and Lyapunov-based model predictive control (LMPC) is proposed for trajectory tracking of a transmitting cylindrical container filled with liquid. Firstly, this research presents an equivalent discrete model based on a mass-spring-damper system. Subsequently, after the flatness of the adopted non-linear model for 2D is established, time-optimal trajectories are introduced. A control method called LMPC is shown to solve the problem of orbital tracking, which allows setting limits for state variables. In addition, to ensure system performance, a linear extended state observer (LESO) is integrated to cope with system uncertainties. Finally, the efficiency of the proposed approach for liquid sloshing suppression and tracking is illustrated by simulations.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"6 3","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.12121","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142077783","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IET Cybersystems and Robotics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1