Pub Date : 2024-11-28DOI: 10.1007/s43684-024-00082-w
Farzad Azizi Zade, Arvin Ebrahimkhanlou
Previous studies have primarily focused on converting point clouds (PC) into a dense mech of 3D finite element models, neglecting the conversion of PCs into as-built wireframe models with two-node elements for line elements such as beams and columns. This study aims to demonstrate the feasibility of this direct conversion, utilizing building framing patterns to create wireframe models. The study also integrates the OpenSeesPy package for modal analysis and double integration for bending estimation to demonstrate the application of the presented method in robotic inspection. Results indicate the successful conversion of a 4-story mass timber building PC to a 3D structural model with an average error of 7.5% under simplified assumptions. Further, two complex mass timber shed PCs were tested, resulting in detailed wireframe models. According to resource monitoring, our method can process ∼593 points/second, mostly affected by the number of neighbors used in the first stage of sparse points removal. Lastly, our method detects beams, columns, ceilings (floors), and walls with their directions. This research can facilitate various structural modeling directly based on PC data for digital twinning and autonomous robotic inspection.
以往的研究主要集中在将点云(PC)转换为密集的三维有限元模型,而忽略了将 PC 转换为带有双节点元素的线性元素(如梁和柱)的竣工线框模型。本研究旨在证明这种直接转换的可行性,利用建筑框架模式创建线框模型。本研究还集成了用于模态分析的 OpenSeesPy 软件包和用于弯曲估算的双重积分,以演示所提出的方法在机器人检测中的应用。结果表明,在简化假设条件下,成功地将一栋 4 层大规模木结构建筑的 PC 转换为三维结构模型,平均误差为 7.5%。此外,我们还测试了两个复杂的大型木结构建筑 PC,并生成了详细的线框模型。根据资源监测,我们的方法每秒可处理 ∼ 593 个点,这主要受到第一阶段稀疏点去除过程中使用的邻接点数量的影响。最后,我们的方法可以检测梁、柱、天花板(地板)和墙壁及其方向。这项研究有助于直接基于 PC 数据进行各种结构建模,从而实现数字结对和自主机器人检测。
{"title":"Point clouds to as-built two-node wireframe digital twin: a novel method to support autonomous robotic inspection","authors":"Farzad Azizi Zade, Arvin Ebrahimkhanlou","doi":"10.1007/s43684-024-00082-w","DOIUrl":"10.1007/s43684-024-00082-w","url":null,"abstract":"<div><p>Previous studies have primarily focused on converting point clouds (PC) into a dense mech of 3D finite element models, neglecting the conversion of PCs into as-built wireframe models with two-node elements for line elements such as beams and columns. This study aims to demonstrate the feasibility of this direct conversion, utilizing building framing patterns to create wireframe models. The study also integrates the OpenSeesPy package for modal analysis and double integration for bending estimation to demonstrate the application of the presented method in robotic inspection. Results indicate the successful conversion of a 4-story mass timber building PC to a 3D structural model with an average error of 7.5% under simplified assumptions. Further, two complex mass timber shed PCs were tested, resulting in detailed wireframe models. According to resource monitoring, our method can process ∼593 points/second, mostly affected by the number of neighbors used in the first stage of sparse points removal. Lastly, our method detects beams, columns, ceilings (floors), and walls with their directions. This research can facilitate various structural modeling directly based on PC data for digital twinning and autonomous robotic inspection.</p></div>","PeriodicalId":71187,"journal":{"name":"自主智能系统(英文)","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43684-024-00082-w.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142737095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-19DOI: 10.1007/s43684-024-00081-x
Haoqi Li, Jiangping Hu, Xiaoming Hu, Bijoy K. Ghosh
Combining safety objectives with stability objectives is crucial for safety-critical systems. Existing studies generally unified these two objectives by constructing Lyapunov-type barrier functions. However, insufficient analysis of key set relationships within the system may render the proposed safety and stability conditions conservative, and these studies also did not provide how to use such conditions to design safety-stability control strategies. This paper proposed a feasible and constructive design to achieve stabilization of safety-critical systems by a relaxed converse Lyapunov-barrier approach. By analyzing the relationships between a series of sets associated with the safety-critical system, the stability and safety conditions can be appropriately relaxed. Then, with the help of relaxed converse control Lyapunov-barrier functions (RCCLBFs), a theoretical result was obtained for the stability of affine nonlinear systems with safety constraints. Subsequently, a constructive method was developed for a second-order strict-feedback system to transform the process of solving RCCLBFs into a Lyapunov-like stabilization problem. Finally, the proposed safety-stability control method is exerted on a robotic system and demonstrated by simulations.
{"title":"Stabilization of nonlinear safety-critical systems by relaxed converse Lyapunov-barrier approach and its applications in robotic systems","authors":"Haoqi Li, Jiangping Hu, Xiaoming Hu, Bijoy K. Ghosh","doi":"10.1007/s43684-024-00081-x","DOIUrl":"10.1007/s43684-024-00081-x","url":null,"abstract":"<div><p>Combining safety objectives with stability objectives is crucial for safety-critical systems. Existing studies generally unified these two objectives by constructing Lyapunov-type barrier functions. However, insufficient analysis of key set relationships within the system may render the proposed safety and stability conditions conservative, and these studies also did not provide how to use such conditions to design safety-stability control strategies. This paper proposed a feasible and constructive design to achieve stabilization of safety-critical systems by a relaxed converse Lyapunov-barrier approach. By analyzing the relationships between a series of sets associated with the safety-critical system, the stability and safety conditions can be appropriately relaxed. Then, with the help of relaxed converse control Lyapunov-barrier functions (RCCLBFs), a theoretical result was obtained for the stability of affine nonlinear systems with safety constraints. Subsequently, a constructive method was developed for a second-order strict-feedback system to transform the process of solving RCCLBFs into a Lyapunov-like stabilization problem. Finally, the proposed safety-stability control method is exerted on a robotic system and demonstrated by simulations.</p></div>","PeriodicalId":71187,"journal":{"name":"自主智能系统(英文)","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43684-024-00081-x.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142672387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-13DOI: 10.1007/s43684-024-00080-y
Tingting Bao, Ding Lin, Xumei Zhang, Zhiguo Zhou, Kejia Wang
As an essential part of modern smart manufacturing, road transport with large and heavy trucks has in-creased dramatically. Due to the inside wheel difference in the process of turning, there is a considerable safety hazard in the blind area of the inside wheel difference. In this paper, multiple cameras combined with deep learning algorithms are introduced to detect pedestrians in the blind area of wheel error. A scheme of vehicle-pedestrian safety alarm detection system is developed via the integration of YOLOv5 and an improved binocular distance measurement method. The system accurately measures the distance between the truck and nearby pedestrians by utilizing multiple cameras and PP Human recognition, providing real-time safety alerts. The experimental results show that this method significantly reduces distance measurement errors, improves the reliability of pedestrian detection, achieves high accuracy and real-time performance, and thus enhances the safety of trucks in complex traffic environments.
{"title":"Pedestrian safety alarm system based on binocular distance measurement for trucks using recognition feature analysis","authors":"Tingting Bao, Ding Lin, Xumei Zhang, Zhiguo Zhou, Kejia Wang","doi":"10.1007/s43684-024-00080-y","DOIUrl":"10.1007/s43684-024-00080-y","url":null,"abstract":"<div><p>As an essential part of modern smart manufacturing, road transport with large and heavy trucks has in-creased dramatically. Due to the inside wheel difference in the process of turning, there is a considerable safety hazard in the blind area of the inside wheel difference. In this paper, multiple cameras combined with deep learning algorithms are introduced to detect pedestrians in the blind area of wheel error. A scheme of vehicle-pedestrian safety alarm detection system is developed via the integration of YOLOv5 and an improved binocular distance measurement method. The system accurately measures the distance between the truck and nearby pedestrians by utilizing multiple cameras and PP Human recognition, providing real-time safety alerts. The experimental results show that this method significantly reduces distance measurement errors, improves the reliability of pedestrian detection, achieves high accuracy and real-time performance, and thus enhances the safety of trucks in complex traffic environments.</p></div>","PeriodicalId":71187,"journal":{"name":"自主智能系统(英文)","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43684-024-00080-y.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600740","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-01DOI: 10.1007/s43684-024-00077-7
Tingting Bao, Zhijun Wu, Jianliang Chen
Feasible, smooth, and time-jerk optimal trajectory is essential for manipulators utilized in manufacturing process. A novel technique to generate trajectories in the joint space for robotic manipulators based on quintic B-spline and constrained multi-objective student psychology based optimization (CMOSPBO) is proposed in this paper. In order to obtain the optimal trajectories, two objective functions including the total travelling time and the integral of the squared jerk along the whole trajectories are considered. The whole trajectories are interpolated by quintic B-spline and then optimized by CMOSPBO, while taking into account kinematic constraints of velocity, acceleration, and jerk. CMOSPBO mainly includes improved student psychology based optimization, archive management, and an adaptive ε-constraint handling method. Lévy flights and differential mutation are adopted to enhance the global exploration capacity of the improved SPBO. The ε value is varied with iterations and feasible solutions to prevent the premature convergence of CMOSPBO. Solution density estimation corresponding to the solution distribution in decision space and objective space is proposed to increase the diversity of solutions. The experimental results show that CMOSPBO outperforms than SQP, and NSGA-II in terms of the motion efficiency and jerk. The comparison results demonstrate the effectiveness of the proposed method to generate time-jerk optimal and jerk-continuous trajectories for manipulators.
{"title":"Multi-objective optimal trajectory planning for manipulators based on CMOSPBO","authors":"Tingting Bao, Zhijun Wu, Jianliang Chen","doi":"10.1007/s43684-024-00077-7","DOIUrl":"10.1007/s43684-024-00077-7","url":null,"abstract":"<div><p>Feasible, smooth, and time-jerk optimal trajectory is essential for manipulators utilized in manufacturing process. A novel technique to generate trajectories in the joint space for robotic manipulators based on quintic B-spline and constrained multi-objective student psychology based optimization (CMOSPBO) is proposed in this paper. In order to obtain the optimal trajectories, two objective functions including the total travelling time and the integral of the squared jerk along the whole trajectories are considered. The whole trajectories are interpolated by quintic B-spline and then optimized by CMOSPBO, while taking into account kinematic constraints of velocity, acceleration, and jerk. CMOSPBO mainly includes improved student psychology based optimization, archive management, and an adaptive <i>ε</i>-constraint handling method. Lévy flights and differential mutation are adopted to enhance the global exploration capacity of the improved SPBO. The <i>ε</i> value is varied with iterations and feasible solutions to prevent the premature convergence of CMOSPBO. Solution density estimation corresponding to the solution distribution in decision space and objective space is proposed to increase the diversity of solutions. The experimental results show that CMOSPBO outperforms than SQP, and NSGA-II in terms of the motion efficiency and jerk. The comparison results demonstrate the effectiveness of the proposed method to generate time-jerk optimal and jerk-continuous trajectories for manipulators.</p></div>","PeriodicalId":71187,"journal":{"name":"自主智能系统(英文)","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43684-024-00077-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142565895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-25DOI: 10.1007/s43684-024-00078-6
Yichen Zhou, Wenhe Han, Heng Zhou
Customer maintenance is of vital importance to the enterprise management. Valuable assessment and efficient prediction for customer ordering behavior can offer better decision-making and reduce business costs significantly. According to existing studies about customer behavior regularity segment and demand prediction most focus on e-commerce and other fields with large amount of data, making them not suitable for small enterprises and data features like sparsity and outliers are not mined when doing regularity quantification. Additionally, more and more complex network structures for demand prediction are proposed, which builds on the assumption that all the samples have predictive value, ignoring the fine-grained analysis of different time series regularity with high cost. To deal with the above issues, a multi-step regularity assessment and joint prediction system for ordering time series is proposed. For extracting features, comprehensive assessment of customer regularity based on entropy weight method with the result of predictability quantification using K-Means clustering algorithm, real entropy, LZW algorithm and anomaly detection adopting Isolation Forest algorithm not only gives an objective result to ‘how high the regularity of customers is’, filling the gap in the field of regularity quantification, but also provides a theoretical basis for demand prediction models selection. Prediction models: Random Forest regression, XGBoost, CNN and LSTM network are experimented with sMAPE and MSLE for performance evaluation to verify the effectiveness of the proposed regularity quantitation method. Moreover, a merged CNN-BiLSTM neural network model is established for predicting those customers with low regularity and difficult to predict by traditional machine leaning algorithms, which performs better on the data set compared to others. Random Forest is still used for prediction of customers with high regularity due to its high training efficiency. Finally, the results of prediction, regularity quantification, and classification are output from the intelligent system, which is capable of providing scientific basis for corporate strategy decision and has highly extendibility in other enterprises and fields for follow-up research.
{"title":"A multi-step regularity assessment and joint prediction system for ordering time series based on entropy and deep learning","authors":"Yichen Zhou, Wenhe Han, Heng Zhou","doi":"10.1007/s43684-024-00078-6","DOIUrl":"10.1007/s43684-024-00078-6","url":null,"abstract":"<div><p>Customer maintenance is of vital importance to the enterprise management. Valuable assessment and efficient prediction for customer ordering behavior can offer better decision-making and reduce business costs significantly. According to existing studies about customer behavior regularity segment and demand prediction most focus on e-commerce and other fields with large amount of data, making them not suitable for small enterprises and data features like sparsity and outliers are not mined when doing regularity quantification. Additionally, more and more complex network structures for demand prediction are proposed, which builds on the assumption that all the samples have predictive value, ignoring the fine-grained analysis of different time series regularity with high cost. To deal with the above issues, a multi-step regularity assessment and joint prediction system for ordering time series is proposed. For extracting features, comprehensive assessment of customer regularity based on entropy weight method with the result of predictability quantification using K-Means clustering algorithm, real entropy, LZW algorithm and anomaly detection adopting Isolation Forest algorithm not only gives an objective result to ‘how high the regularity of customers is’, filling the gap in the field of regularity quantification, but also provides a theoretical basis for demand prediction models selection. Prediction models: Random Forest regression, XGBoost, CNN and LSTM network are experimented with sMAPE and MSLE for performance evaluation to verify the effectiveness of the proposed regularity quantitation method. Moreover, a merged CNN-BiLSTM neural network model is established for predicting those customers with low regularity and difficult to predict by traditional machine leaning algorithms, which performs better on the data set compared to others. Random Forest is still used for prediction of customers with high regularity due to its high training efficiency. Finally, the results of prediction, regularity quantification, and classification are output from the intelligent system, which is capable of providing scientific basis for corporate strategy decision and has highly extendibility in other enterprises and fields for follow-up research.</p></div>","PeriodicalId":71187,"journal":{"name":"自主智能系统(英文)","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43684-024-00078-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142519061","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Metal powder contributes to the environmental burdens of additive manufacturing (AM) substantially. Current life cycle assessments (LCAs) of metal powders present considerable variations of lifecycle environmental inventory due to process divergence, spatial heterogeneity, or temporal fluctuation. Most importantly, the amounts of LCA studies on metal powder are limited and primarily confined to partial material types. To this end, based on the data surveyed from a metal powder supplier, this study conducted an LCA of titanium and nickel alloy produced by electrode-inducted and vacuum-inducted melting gas atomization, respectively. Given that energy consumption dominates the environmental burden of powder production and is influenced by metal materials’ physical properties, we proposed a Bayesian stochastic Kriging model to estimate the energy consumption during the gas atomization process. This model considered the inherent uncertainties of training data and adaptively updated the parameters of interest when new environmental data on gas atomization were available. With the predicted energy use information of specific powder, the corresponding lifecycle environmental impacts can be further autonomously estimated in conjunction with the other surveyed powder production stages. Results indicated the environmental impact of titanium alloy powder is slightly higher than that of nickel alloy powder and their lifecycle carbon emissions are around 20 kg CO2 equivalency. The proposed Bayesian stochastic Kriging model showed more accurate predictions of energy consumption compared with conventional Kriging and stochastic Kriging models. This study enables data imputation of energy consumption during gas atomization given the physical properties and producing technique of powder materials.
{"title":"Life cycle assessment of metal powder production: a Bayesian stochastic Kriging model-based autonomous estimation","authors":"Haibo Xiao, Baoyun Gao, Shoukang Yu, Bin Liu, Sheng Cao, Shitong Peng","doi":"10.1007/s43684-024-00079-5","DOIUrl":"10.1007/s43684-024-00079-5","url":null,"abstract":"<div><p>Metal powder contributes to the environmental burdens of additive manufacturing (AM) substantially. Current life cycle assessments (LCAs) of metal powders present considerable variations of lifecycle environmental inventory due to process divergence, spatial heterogeneity, or temporal fluctuation. Most importantly, the amounts of LCA studies on metal powder are limited and primarily confined to partial material types. To this end, based on the data surveyed from a metal powder supplier, this study conducted an LCA of titanium and nickel alloy produced by electrode-inducted and vacuum-inducted melting gas atomization, respectively. Given that energy consumption dominates the environmental burden of powder production and is influenced by metal materials’ physical properties, we proposed a Bayesian stochastic Kriging model to estimate the energy consumption during the gas atomization process. This model considered the inherent uncertainties of training data and adaptively updated the parameters of interest when new environmental data on gas atomization were available. With the predicted energy use information of specific powder, the corresponding lifecycle environmental impacts can be further autonomously estimated in conjunction with the other surveyed powder production stages. Results indicated the environmental impact of titanium alloy powder is slightly higher than that of nickel alloy powder and their lifecycle carbon emissions are around 20 kg CO<sub>2</sub> equivalency. The proposed Bayesian stochastic Kriging model showed more accurate predictions of energy consumption compared with conventional Kriging and stochastic Kriging models. This study enables data imputation of energy consumption during gas atomization given the physical properties and producing technique of powder materials.</p></div>","PeriodicalId":71187,"journal":{"name":"自主智能系统(英文)","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43684-024-00079-5.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142443184","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Textile dyeing requires optimizing combinations of ingredients and process parameters to achieve target colour properties. Modelling the complex relationships between these factors and the resulting colour is challenging. In this case, a physics-informed approach for multi-output regression to model CIELAB colour values from dyeing ingredient and process inputs is proposed. Leveraging attention mechanisms and multi-task learning, the model outperforms baseline methods at predicting multiple colour outputs jointly. Specifically, the Transformer model’s attention mechanism captures the complex interactions between dyeing ingredients and process parameters, while the multi-task learning framework exploits the intrinsic correlations among the L*, a*, and b* dimensions of the CIELAB colour space. In addition, the incorporation of physical knowledge through a physics-informed loss function integrates the CMC colour difference formula. This loss function, along with the attention mechanisms, enables the model to learn the nuanced relationships between the dyeing process variables and the final colour output, thereby improving the overall prediction accuracy. This reduces trial-and-error costs and resource waste, contributing to environmental sustainability by minimizing water and energy consumption and chemical emissions.
{"title":"Leveraging multi-output modelling for CIELAB using colour difference formula towards sustainable textile dyeing","authors":"Zheyuan Chen, Jian Liu, Jian Li, Mukun Yuan, Guangping Yu","doi":"10.1007/s43684-024-00076-8","DOIUrl":"10.1007/s43684-024-00076-8","url":null,"abstract":"<div><p>Textile dyeing requires optimizing combinations of ingredients and process parameters to achieve target colour properties. Modelling the complex relationships between these factors and the resulting colour is challenging. In this case, a physics-informed approach for multi-output regression to model CIELAB colour values from dyeing ingredient and process inputs is proposed. Leveraging attention mechanisms and multi-task learning, the model outperforms baseline methods at predicting multiple colour outputs jointly. Specifically, the Transformer model’s attention mechanism captures the complex interactions between dyeing ingredients and process parameters, while the multi-task learning framework exploits the intrinsic correlations among the L*, a*, and b* dimensions of the CIELAB colour space. In addition, the incorporation of physical knowledge through a physics-informed loss function integrates the CMC colour difference formula. This loss function, along with the attention mechanisms, enables the model to learn the nuanced relationships between the dyeing process variables and the final colour output, thereby improving the overall prediction accuracy. This reduces trial-and-error costs and resource waste, contributing to environmental sustainability by minimizing water and energy consumption and chemical emissions.</p></div>","PeriodicalId":71187,"journal":{"name":"自主智能系统(英文)","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43684-024-00076-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142413919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-18DOI: 10.1007/s43684-024-00075-9
Gang Huang, Liangzhu Lu, Yifan Zhang, Gangfu Cao, Zhe Zhou
To solve the problem of mobile robots needing to adjust their pose for accurate operation after reaching the target point in the indoor environment, a localization method based on scene modeling and recognition has been designed. Firstly, the offline scene model is created by both handcrafted feature and semantic feature. Then, the scene recognition and location calculation are performed online based on the offline scene model. To improve the accuracy of recognition and location calculation, this paper proposes a method that integrates both semantic features matching and handcrafted features matching. Based on the results of scene recognition, the accurate location is obtained through metric calculation with 3D information. The experimental results show that the accuracy of scene recognition is over 90%, and the average localization error is less than 1 meter. Experimental results demonstrate that the localization has a better performance after using the proposed improved method.
{"title":"Improved vision-only localization method for mobile robots in indoor environments","authors":"Gang Huang, Liangzhu Lu, Yifan Zhang, Gangfu Cao, Zhe Zhou","doi":"10.1007/s43684-024-00075-9","DOIUrl":"10.1007/s43684-024-00075-9","url":null,"abstract":"<div><p>To solve the problem of mobile robots needing to adjust their pose for accurate operation after reaching the target point in the indoor environment, a localization method based on scene modeling and recognition has been designed. Firstly, the offline scene model is created by both handcrafted feature and semantic feature. Then, the scene recognition and location calculation are performed online based on the offline scene model. To improve the accuracy of recognition and location calculation, this paper proposes a method that integrates both semantic features matching and handcrafted features matching. Based on the results of scene recognition, the accurate location is obtained through metric calculation with 3D information. The experimental results show that the accuracy of scene recognition is over 90%, and the average localization error is less than 1 meter. Experimental results demonstrate that the localization has a better performance after using the proposed improved method.</p></div>","PeriodicalId":71187,"journal":{"name":"自主智能系统(英文)","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43684-024-00075-9.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142412349","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-14DOI: 10.1007/s43684-024-00074-w
Julius Bächle, Jakob Häringer, Noah Köhler, Kadir-Kaan Özer, Markus Enzweiler, Reiner Marchthaler
This article introduces an open-source software stack designed for autonomous 1:10 scale model vehicles. Initially developed for the Bosch Future Mobility Challenge (BFMC) student competition, this versatile software stack is applicable to a variety of autonomous driving competitions. The stack comprises perception, planning, and control modules, each essential for precise and reliable scene understanding in complex environments such as a miniature smart city in the context of BFMC. Given the limited computing power of model vehicles and the necessity for low-latency real-time applications, the stack is implemented in C++, employs YOLO Version 5 s for environmental perception, and leverages the state-of-the-art Robot Operating System (ROS) for inter-process communication. We believe that this article and the accompanying open-source software will be a valuable resource for future teams participating in autonomous driving student competitions. Our work can serve as a foundational tool for novice teams and a reference for more experienced participants. The code and data are publicly available on GitHub.
本文介绍了专为 1:10 比例自动驾驶模型车设计的开源软件栈。这款多功能软件堆栈最初是为博世未来交通挑战赛(BFMC)学生竞赛开发的,适用于各种自动驾驶竞赛。该堆栈包括感知、规划和控制模块,每个模块对于在复杂环境(如 BFMC 中的微型智能城市)中精确可靠地理解场景都至关重要。鉴于模型车的计算能力有限以及低延迟实时应用的必要性,该堆栈采用 C++ 实现,使用 YOLO Version 5 s 进行环境感知,并利用最先进的机器人操作系统 (ROS) 进行进程间通信。我们相信,这篇文章和随附的开源软件将成为未来参加自动驾驶学生竞赛团队的宝贵资源。我们的工作可作为新手团队的基础工具和经验丰富的参赛者的参考资料。代码和数据可在 GitHub 上公开获取。
{"title":"Competing with autonomous model vehicles: a software stack for driving in smart city environments","authors":"Julius Bächle, Jakob Häringer, Noah Köhler, Kadir-Kaan Özer, Markus Enzweiler, Reiner Marchthaler","doi":"10.1007/s43684-024-00074-w","DOIUrl":"10.1007/s43684-024-00074-w","url":null,"abstract":"<div><p>This article introduces an open-source software stack designed for autonomous 1:10 scale model vehicles. Initially developed for the Bosch Future Mobility Challenge (BFMC) student competition, this versatile software stack is applicable to a variety of autonomous driving competitions. The stack comprises perception, planning, and control modules, each essential for precise and reliable scene understanding in complex environments such as a miniature smart city in the context of BFMC. Given the limited computing power of model vehicles and the necessity for low-latency real-time applications, the stack is implemented in C++, employs YOLO Version 5 s for environmental perception, and leverages the state-of-the-art Robot Operating System (ROS) for inter-process communication. We believe that this article and the accompanying open-source software will be a valuable resource for future teams participating in autonomous driving student competitions. Our work can serve as a foundational tool for novice teams and a reference for more experienced participants. The code and data are publicly available on GitHub.</p></div>","PeriodicalId":71187,"journal":{"name":"自主智能系统(英文)","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43684-024-00074-w.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142411651","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-09DOI: 10.1007/s43684-024-00073-x
Quanxi Zhan, Yanmin Zhou, Junrui Zhang, Chenyang Sun, Runjie Shen, Bin He
Accurate velocity measurement of unmanned aerial vehicles (UAVs) is essential in various applications. Traditional vision-based methods rely heavily on visual features, which are often inadequate in low-light or feature-sparse environments. This study presents a novel approach to measure the axial velocity of UAVs using motion blur images captured by a UAV-mounted monocular camera. We introduce a motion blur model that synthesizes imaging from neighboring frames to enhance motion blur visibility. The synthesized blur frames are transformed into spectrograms using the Fast Fourier Transform (FFT) technique. We then apply a binarization process and the Radon transform to extract light-dark stripe spacing, which represents the motion blur length. This length is used to establish a model correlating motion blur with axial velocity, allowing precise velocity calculation. Field tests in a hydropower station penstock demonstrated an average velocity error of 0.048 m/s compared to ultra-wideband (UWB) measurements. The root-mean-square error was 0.025, with an average computational time of 42.3 ms and CPU load of 17%. These results confirm the stability and accuracy of our velocity estimation algorithm in challenging environments.
{"title":"A novel method for measuring center-axis velocity of unmanned aerial vehicles through synthetic motion blur images","authors":"Quanxi Zhan, Yanmin Zhou, Junrui Zhang, Chenyang Sun, Runjie Shen, Bin He","doi":"10.1007/s43684-024-00073-x","DOIUrl":"10.1007/s43684-024-00073-x","url":null,"abstract":"<div><p>Accurate velocity measurement of unmanned aerial vehicles (UAVs) is essential in various applications. Traditional vision-based methods rely heavily on visual features, which are often inadequate in low-light or feature-sparse environments. This study presents a novel approach to measure the axial velocity of UAVs using motion blur images captured by a UAV-mounted monocular camera. We introduce a motion blur model that synthesizes imaging from neighboring frames to enhance motion blur visibility. The synthesized blur frames are transformed into spectrograms using the Fast Fourier Transform (FFT) technique. We then apply a binarization process and the Radon transform to extract light-dark stripe spacing, which represents the motion blur length. This length is used to establish a model correlating motion blur with axial velocity, allowing precise velocity calculation. Field tests in a hydropower station penstock demonstrated an average velocity error of 0.048 m/s compared to ultra-wideband (UWB) measurements. The root-mean-square error was 0.025, with an average computational time of 42.3 ms and CPU load of 17%. These results confirm the stability and accuracy of our velocity estimation algorithm in challenging environments.</p></div>","PeriodicalId":71187,"journal":{"name":"自主智能系统(英文)","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43684-024-00073-x.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141666199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}