首页 > 最新文献

IEEE Transactions on Intelligent Vehicles最新文献

英文 中文
Embodied Intelligence-Based Perception, Decision-Making, and Control for Autonomous Operations of Rail Transportation 基于具身智能的轨道交通自主运营感知、决策与控制
IF 14.3 1区 工程技术 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-12-13 DOI: 10.1109/TIV.2024.3517335
Min Zhou;Hairong Dong;Haifeng Song;Nan Zheng;Wen-Hua Chen;Hongwei Wang
This perspective proposes a framework for autonomous operations of rail transportation based on embodied intelligence to enhance environmental adaptation and autonomous decision-making capabilities. Three key technologies are outlined to enable perception, decision-making, and control in rail transportation, i.e., embodied perception for active environmental understanding, embodied execution, and embodied learning and evolution. It also explores the main challenges of implementing embodied intelligence-based autonomous operations. The proposed framework offers new insights and directions for practitioners in the rail transportation industry.
该观点提出了基于具身智能的轨道交通自主运行框架,以增强环境适应能力和自主决策能力。提出了实现轨道交通感知、决策和控制的三个关键技术,即主动环境理解的具身感知、具身执行和具身学习与进化。它还探讨了实现基于实体智能的自主操作的主要挑战。提出的框架为轨道交通行业的从业者提供了新的见解和方向。
{"title":"Embodied Intelligence-Based Perception, Decision-Making, and Control for Autonomous Operations of Rail Transportation","authors":"Min Zhou;Hairong Dong;Haifeng Song;Nan Zheng;Wen-Hua Chen;Hongwei Wang","doi":"10.1109/TIV.2024.3517335","DOIUrl":"https://doi.org/10.1109/TIV.2024.3517335","url":null,"abstract":"This perspective proposes a framework for autonomous operations of rail transportation based on embodied intelligence to enhance environmental adaptation and autonomous decision-making capabilities. Three key technologies are outlined to enable perception, decision-making, and control in rail transportation, i.e., embodied perception for active environmental understanding, embodied execution, and embodied learning and evolution. It also explores the main challenges of implementing embodied intelligence-based autonomous operations. The proposed framework offers new insights and directions for practitioners in the rail transportation industry.","PeriodicalId":36532,"journal":{"name":"IEEE Transactions on Intelligent Vehicles","volume":"10 12","pages":"5061-5065"},"PeriodicalIF":14.3,"publicationDate":"2024-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145772050","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Innovating Waterway Route Planning as a Service for Marine Traffic Applications 创新航道规划为海上交通应用服务
IF 14.3 1区 工程技术 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-12-12 DOI: 10.1109/TIV.2024.3516361
Zhe Xiao;Xiaocai Zhang;Xiuju Fu;Liye Zhang;Haiyan Xu;Ryan Wen Liu;Chee Seng Chong;Zheng Qin
Waterway pattern mining and route planning are essential system services in maritime applications to support safety, efficiency, and sustainability goals. In this paper, first, we propose a novel waterway pattern mining method. It allows a compact footprint design and is featured with manifold granularities, waypoint and directional tagging, forming a knowledge base with rich traffic features. Second, relying on the extracted waterway patterns as “map context”, we enhance the Theta* path planning algorithm by taking the extracted traffic properties into consideration. An integrated solution combining both the waterway patterns and the enhanced Theta* algorithm has been applied to several domain use cases: a) waterway pattern based trajectory reconstruction; b) passage plan generation for various types of vessels; and c) vessel movement estimation and forecasting. These use cases proved the usefulness and practicality of the proposed solution, its feasibility for other port waters, and potential use on a global scale.
航道模式挖掘和航线规划是海事应用中必不可少的系统服务,以支持安全、效率和可持续性目标。本文首先提出了一种新的航道模式挖掘方法。它允许紧凑的占地设计,并具有多粒度、路点和方向标记的特点,形成了一个具有丰富交通特征的知识库。其次,以提取的水路模式作为“地图上下文”,考虑提取的交通属性,增强Theta*路径规划算法。结合航道模式和增强Theta*算法的集成解决方案已应用于几个领域的用例:a)基于航道模式的轨迹重建;B)各型船舶的航道规划生成;c)船舶运动估计与预测。这些用例证明了所提出的解决方案的有用性和实用性,它在其他港口水域的可行性,以及在全球范围内的潜在使用。
{"title":"Innovating Waterway Route Planning as a Service for Marine Traffic Applications","authors":"Zhe Xiao;Xiaocai Zhang;Xiuju Fu;Liye Zhang;Haiyan Xu;Ryan Wen Liu;Chee Seng Chong;Zheng Qin","doi":"10.1109/TIV.2024.3516361","DOIUrl":"https://doi.org/10.1109/TIV.2024.3516361","url":null,"abstract":"Waterway pattern mining and route planning are essential system services in maritime applications to support safety, efficiency, and sustainability goals. In this paper, first, we propose a novel waterway pattern mining method. It allows a compact footprint design and is featured with manifold granularities, waypoint and directional tagging, forming a knowledge base with rich traffic features. Second, relying on the extracted waterway patterns as “map context”, we enhance the Theta<sup>*</sup> path planning algorithm by taking the extracted traffic properties into consideration. An integrated solution combining both the waterway patterns and the enhanced Theta<sup>*</sup> algorithm has been applied to several domain use cases: a) waterway pattern based trajectory reconstruction; b) passage plan generation for various types of vessels; and c) vessel movement estimation and forecasting. These use cases proved the usefulness and practicality of the proposed solution, its feasibility for other port waters, and potential use on a global scale.","PeriodicalId":36532,"journal":{"name":"IEEE Transactions on Intelligent Vehicles","volume":"10 12","pages":"5151-5161"},"PeriodicalIF":14.3,"publicationDate":"2024-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145772073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LC-Dual: Coupling Predictive Information and Redundant Strategies for Autonomous Vehicle Lane Change Trajectory Planning LC-Dual:自动驾驶汽车变道轨迹规划的耦合预测信息和冗余策略
IF 14.3 1区 工程技术 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-12-11 DOI: 10.1109/TIV.2024.3515219
Kui Xia;Haoran Jiang;Zipeng Man;Yangsheng Jiang;Zhihong Yao
Autonomous vehicles (AVs) face considerable challenges when interacting with their environment, especially during lane changes. Enhancing decision-making and planning processes with prediction regarding the intentions and trajectories of surrounding vehicles can significantly improve lane change performance. However, prediction accuracy is limited by both the technology used and environmental variability. Utilizing low-confidence predictive data can adversely impact the safety and comfort of AV operations. This paper proposes a novel dual-model framework for lane change decision-making and planning of AVs (LC-Dual), which dialectically uses predictive data and provides redundant safety measures in parallel. In Model I, the optimal end-state trajectory for lane changes is planned using predictive information from the upper layer. In Model II, a redundant lane change trajectory is quickly generated based on a spatio-temporal safety corridor constructed from real perception data. Ultimately, the selection of the lane change model and trajectory is determined by rule-based decision-making, factoring in prediction confidence and computational efficiency. Simulation experiments demonstrate that the LC-Dual framework yields more adaptive trajectories in scenarios with accurate predictions and effective switches between lane change models in cases of inaccurate predictions. The LC-Dual framework markedly improves safety and efficiency in lane change operations, thereby facilitating broader AV adoption.
自动驾驶汽车(AVs)在与环境交互时面临着相当大的挑战,尤其是在变道时。通过预测周围车辆的意图和轨迹来增强决策和规划过程可以显著提高变道性能。然而,预测的准确性受到所使用的技术和环境变化的限制。使用低置信度的预测数据可能会对自动驾驶操作的安全性和舒适性产生不利影响。本文提出了一种新的自动驾驶汽车变道决策和规划双模型框架(LC-Dual),该框架辩证地使用预测数据并并行提供冗余安全措施。在模型1中,使用上层的预测信息来规划变道的最优最终状态轨迹。在模型II中,基于真实感知数据构建的时空安全走廊,快速生成冗余变道轨迹。最后,在考虑预测置信度和计算效率的情况下,通过基于规则的决策来确定变道模型和轨迹的选择。仿真实验表明,LC-Dual框架在预测准确的情况下产生更自适应的轨迹,在预测不准确的情况下在变道模型之间有效切换。LC-Dual框架显著提高了变道操作的安全性和效率,从而促进了自动驾驶汽车的广泛采用。
{"title":"LC-Dual: Coupling Predictive Information and Redundant Strategies for Autonomous Vehicle Lane Change Trajectory Planning","authors":"Kui Xia;Haoran Jiang;Zipeng Man;Yangsheng Jiang;Zhihong Yao","doi":"10.1109/TIV.2024.3515219","DOIUrl":"https://doi.org/10.1109/TIV.2024.3515219","url":null,"abstract":"Autonomous vehicles (AVs) face considerable challenges when interacting with their environment, especially during lane changes. Enhancing decision-making and planning processes with prediction regarding the intentions and trajectories of surrounding vehicles can significantly improve lane change performance. However, prediction accuracy is limited by both the technology used and environmental variability. Utilizing low-confidence predictive data can adversely impact the safety and comfort of AV operations. This paper proposes a novel dual-model framework for lane change decision-making and planning of AVs (LC-Dual), which dialectically uses predictive data and provides redundant safety measures in parallel. In Model I, the optimal end-state trajectory for lane changes is planned using predictive information from the upper layer. In Model II, a redundant lane change trajectory is quickly generated based on a spatio-temporal safety corridor constructed from real perception data. Ultimately, the selection of the lane change model and trajectory is determined by rule-based decision-making, factoring in prediction confidence and computational efficiency. Simulation experiments demonstrate that the LC-Dual framework yields more adaptive trajectories in scenarios with accurate predictions and effective switches between lane change models in cases of inaccurate predictions. The LC-Dual framework markedly improves safety and efficiency in lane change operations, thereby facilitating broader AV adoption.","PeriodicalId":36532,"journal":{"name":"IEEE Transactions on Intelligent Vehicles","volume":"10 12","pages":"5136-5150"},"PeriodicalIF":14.3,"publicationDate":"2024-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145772057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatically Improving Scenario Descriptions Derived From Recorded Traffic 自动改进基于话务量的场景描述
IF 14.3 1区 工程技术 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-12-11 DOI: 10.1109/TIV.2024.3514296
Nicola Kolb;Florian Huber;Alexander Pretschner
In scenario-based testing of autonomous vehicles, scenario types serve as input for generating test cases. Towards a comprehensive catalog of scenario types, expert-defined scenario types are complemented with types derived from real traffic data. Recorded traffic contains instances of different scenario types: By clustering these instances, a scenario type can be associated with each cluster. This involves adding semantics to recorded scenario instances, our clusters' elements, in the form of maneuver-based descriptions of the behavior of scenario participants. Several works offer approaches to derive such instance descriptions. These approaches are often automated. The quality of the generated descriptions matters, given that subsequent steps build upon them. How to derive good descriptions automatically has not been addressed until now. In our work, (1) we approximate a description's quality by measuring the Fréchet distance between the trajectories obtained by simulating a description and the original trajectories in the dataset; (2) we introduce a quality threshold as to what makes a good description and, based on this threshold, we identify inadequate descriptions; (3) we build upon these evaluation results to modify identified inadequate descriptions automatically. In experiments, we extract and describe a set of 843 scenario instances from the inD intersection dataset: the average reconstruction quality in terms of a description's aggregated Fréchet distance of 0.67 and a ratio of 54% (459) of adequate descriptions improves to 0.34, and 97% (819). The average aggregated distance is reduced by half, i.e., the reconstruction quality is doubled.
在基于场景的自动驾驶汽车测试中,场景类型作为生成测试用例的输入。为了实现场景类型的全面目录,专家定义的场景类型与来自真实交通数据的类型相辅相成。记录的流量包含不同场景类型的实例:通过将这些实例聚类,可以将一个场景类型与每个集群关联。这包括以基于动作的场景参与者行为描述的形式,向记录的场景实例(我们的集群元素)添加语义。有几部作品提供了推导这种实例描述的方法。这些方法通常是自动化的。生成的描述的质量很重要,因为后续的步骤建立在它们之上。如何自动获得好的描述,目前还没有得到解决。在我们的工作中,(1)我们通过测量通过模拟描述获得的轨迹与数据集中原始轨迹之间的fr切距离来近似描述的质量;(2)我们引入一个质量阈值,以确定什么是好的描述,并在此阈值的基础上识别不充分的描述;(3)在这些评价结果的基础上,对识别出的不足描述进行自动修改。在实验中,我们从inD交叉数据集中提取并描述了一组843个场景实例:描述的聚合fr切距离为0.67,适当描述的比例为54%(459),平均重建质量提高到0.34,97%(819)。平均聚合距离缩短了一半,即重构质量提高了一倍。
{"title":"Automatically Improving Scenario Descriptions Derived From Recorded Traffic","authors":"Nicola Kolb;Florian Huber;Alexander Pretschner","doi":"10.1109/TIV.2024.3514296","DOIUrl":"https://doi.org/10.1109/TIV.2024.3514296","url":null,"abstract":"In scenario-based testing of autonomous vehicles, scenario types serve as input for generating test cases. Towards a comprehensive catalog of scenario types, expert-defined scenario types are complemented with types derived from real traffic data. Recorded traffic contains instances of different scenario types: By clustering these instances, a scenario type can be associated with each cluster. This involves adding semantics to recorded scenario instances, our clusters' elements, in the form of maneuver-based descriptions of the behavior of scenario participants. Several works offer approaches to derive such instance descriptions. These approaches are often automated. The quality of the generated descriptions matters, given that subsequent steps build upon them. How to derive good descriptions automatically has not been addressed until now. In our work, (1) we approximate a description's quality by measuring the Fréchet distance between the trajectories obtained by simulating a description and the original trajectories in the dataset; (2) we introduce a quality threshold as to what makes a <italic>good</i> description and, based on this threshold, we identify inadequate descriptions; (3) we build upon these evaluation results to modify identified <italic>inadequate</i> descriptions automatically. In experiments, we extract and describe a set of 843 scenario instances from the inD intersection dataset: the average reconstruction quality in terms of a description's aggregated Fréchet distance of 0.67 and a ratio of 54% (459) of adequate descriptions improves to 0.34, and 97% (819). The average aggregated distance is reduced by half, i.e., the reconstruction quality is doubled.","PeriodicalId":36532,"journal":{"name":"IEEE Transactions on Intelligent Vehicles","volume":"10 12","pages":"5126-5135"},"PeriodicalIF":14.3,"publicationDate":"2024-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10793072","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145772071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PC-FusionMap: A Novel Point Cloud Generation and Multimodal Fusion Approach for HD Map Construction PC-FusionMap:一种用于高清地图构建的点云生成和多模态融合方法
IF 14.3 1区 工程技术 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-12-09 DOI: 10.1109/TIV.2024.3513401
Yaoming Zhuang;Zhenjie Duan;Li Li;Chengdong Wu;Zhanlin Liu
High-definition (HD) maps are crucial for autonomous driving, supporting decision-making, control, and localization. However, current map generation methods often rely on single-modality images, which lack depth information and environmental context. To overcome these limitations, we propose PC-FusionMap, a novel approach that generates point cloud modality data and fuses multi-modal and temporal features for accurate and efficient online HD map construction. Our method addresses the shortcomings of existing approaches by leveraging an improved depth estimation module and a supervised labeling strategy to generate point cloud data. We also introduce a multi-modal feature fusion architecture (CFMB) and a temporal fusion network (TFC) to effectively integrate multi-modal and temporal information. The CFMB architecture uses a query mechanism and cross-attention to enhance the complementary performance between modalities, simplifying the fusion process and improving accuracy. The TFC Network models dynamic changes in time series data, further enhancing the accuracy and robustness of online HD map construction. Our approach achieves state-of-the-art results on the nuScenes and the Argoverse2 datasets, surpassing baseline models in both accuracy and stability. Additionally, incorporating our method into existing HD map generation models can lead to substantial performance gains.
高清(HD)地图对自动驾驶至关重要,可以支持决策、控制和定位。然而,目前的地图生成方法往往依赖于单模态图像,缺乏深度信息和环境背景。为了克服这些限制,我们提出了PC-FusionMap,这是一种生成点云模态数据并融合多模态和时间特征的新方法,用于准确高效的在线高清地图构建。我们的方法通过利用改进的深度估计模块和监督标记策略来生成点云数据,解决了现有方法的缺点。我们还引入了一种多模态特征融合架构(CFMB)和一种时间融合网络(TFC)来有效地整合多模态和时间信息。CFMB架构利用查询机制和交叉关注增强了模式间的互补性能,简化了融合过程,提高了精度。TFC网络对时间序列数据的动态变化进行建模,进一步提高了在线高清地图构建的准确性和鲁棒性。我们的方法在nuScenes和Argoverse2数据集上实现了最先进的结果,在准确性和稳定性方面都超过了基线模型。此外,将我们的方法整合到现有的高清地图生成模型中可以带来实质性的性能提升。
{"title":"PC-FusionMap: A Novel Point Cloud Generation and Multimodal Fusion Approach for HD Map Construction","authors":"Yaoming Zhuang;Zhenjie Duan;Li Li;Chengdong Wu;Zhanlin Liu","doi":"10.1109/TIV.2024.3513401","DOIUrl":"https://doi.org/10.1109/TIV.2024.3513401","url":null,"abstract":"High-definition (HD) maps are crucial for autonomous driving, supporting decision-making, control, and localization. However, current map generation methods often rely on single-modality images, which lack depth information and environmental context. To overcome these limitations, we propose PC-FusionMap, a novel approach that generates point cloud modality data and fuses multi-modal and temporal features for accurate and efficient online HD map construction. Our method addresses the shortcomings of existing approaches by leveraging an improved depth estimation module and a supervised labeling strategy to generate point cloud data. We also introduce a multi-modal feature fusion architecture (CFMB) and a temporal fusion network (TFC) to effectively integrate multi-modal and temporal information. The CFMB architecture uses a query mechanism and cross-attention to enhance the complementary performance between modalities, simplifying the fusion process and improving accuracy. The TFC Network models dynamic changes in time series data, further enhancing the accuracy and robustness of online HD map construction. Our approach achieves state-of-the-art results on the nuScenes and the Argoverse2 datasets, surpassing baseline models in both accuracy and stability. Additionally, incorporating our method into existing HD map generation models can lead to substantial performance gains.","PeriodicalId":36532,"journal":{"name":"IEEE Transactions on Intelligent Vehicles","volume":"10 12","pages":"5112-5125"},"PeriodicalIF":14.3,"publicationDate":"2024-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145772017","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bootstrapped Neural Models for Predicting Self-Driving Vehicle Collisions With Quantified Confidence: Offline and Online Applications 基于量化置信度的自举神经模型预测自动驾驶汽车碰撞:离线和在线应用
IF 14.3 1区 工程技术 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-12-09 DOI: 10.1109/TIV.2024.3512786
Antonello Cherubini;Gastone Pietro Rosati Papini;Alice Plebe;Mattia Piazza;Mauro Da Lio
Highly automated vehicles are complex systems, and ensuring their safe operation within their Operational Design Domain (ODD) presents significant challenges. Diagnosing failure modes and updating these systems are even more demanding tasks. This paper introduces a method to assist with assessing, diagnosing, and updating these systems by developing a stochastic model that predicts safety outcomes (collision, near-miss, or safe state) with quantified uncertainty in any parametrized scenario. The approach uses bootstrapping aggregation to create an ensemble of predictive models, leveraging fully connected feed-forward neural networks. These networks are designed with a flexible number of trainable parameters and hidden layers, requiring minimal computational resources. The model is trained on a small set of examples obtained through direct simulations that randomly sample the parametric scenario, bypassing the traditional test matrix definition. Once trained, the bootstrapped model serves as an identity card for the system under test, allowing continuous performance evaluation across the parametric scenario. The paper demonstrates applications, including safety assessment, failure mode identification, and developing a safe speed recommendation function. The model's compact size ensures rapid execution, facilitating extensive post-analysis for safety argumentation and diagnosis and real-time online use to extend the system's abilities.
高度自动化车辆是复杂的系统,确保其在运行设计域(ODD)内的安全运行是一项重大挑战。诊断故障模式和更新这些系统是更艰巨的任务。本文介绍了一种方法,以协助评估,诊断和更新这些系统,通过开发一个随机模型,预测安全结果(碰撞,险些,或安全状态)在任何参数化场景的量化不确定性。该方法使用自举聚合来创建预测模型的集合,利用完全连接的前馈神经网络。这些网络被设计成具有灵活数量的可训练参数和隐藏层,所需的计算资源最少。该模型是通过直接模拟获得的小样本集进行训练的,这些样本集随机采样参数场景,绕过传统的测试矩阵定义。一旦经过训练,自举模型就可以作为被测系统的身份证,允许跨参数场景进行连续的性能评估。本文演示了包括安全评估、故障模式识别和开发安全速度推荐功能在内的应用。该模型的紧凑尺寸确保了快速执行,促进了安全论证和诊断的广泛后分析,以及实时在线使用,以扩展系统的能力。
{"title":"Bootstrapped Neural Models for Predicting Self-Driving Vehicle Collisions With Quantified Confidence: Offline and Online Applications","authors":"Antonello Cherubini;Gastone Pietro Rosati Papini;Alice Plebe;Mattia Piazza;Mauro Da Lio","doi":"10.1109/TIV.2024.3512786","DOIUrl":"https://doi.org/10.1109/TIV.2024.3512786","url":null,"abstract":"Highly automated vehicles are complex systems, and ensuring their safe operation within their Operational Design Domain (ODD) presents significant challenges. Diagnosing failure modes and updating these systems are even more demanding tasks. This paper introduces a method to assist with assessing, diagnosing, and updating these systems by developing a stochastic model that predicts safety outcomes (collision, near-miss, or safe state) with quantified uncertainty in any parametrized scenario. The approach uses bootstrapping aggregation to create an ensemble of predictive models, leveraging fully connected feed-forward neural networks. These networks are designed with a flexible number of trainable parameters and hidden layers, requiring minimal computational resources. The model is trained on a small set of examples obtained through direct simulations that randomly sample the parametric scenario, bypassing the traditional test matrix definition. Once trained, the bootstrapped model serves as an identity card for the system under test, allowing continuous performance evaluation across the parametric scenario. The paper demonstrates applications, including safety assessment, failure mode identification, and developing a safe speed recommendation function. The model's compact size ensures rapid execution, facilitating extensive post-analysis for safety argumentation and diagnosis and real-time online use to extend the system's abilities.","PeriodicalId":36532,"journal":{"name":"IEEE Transactions on Intelligent Vehicles","volume":"10 12","pages":"5079-5099"},"PeriodicalIF":14.3,"publicationDate":"2024-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10783023","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145772093","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Divide, Ensemble and Conquer: The Last Mile on Unsupervised Domain Adaptation for Semantic Segmentation 分割、集成和征服:语义分割中无监督域自适应的最后一公里
IF 14.3 1区 工程技术 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-12-09 DOI: 10.1109/TIV.2024.3512995
Tao Lian;Jose L. Gómez;Antonio M. López
The last mile of unsupervised domain adaptation (UDA) for semantic segmentation is the challenge of solving the syn-to-real domain gap. Recent UDA methods have progressed significantly, yet they often rely on strategies customized for synthetic single-source datasets (e.g., GTA5), which limits their generalisation to multi-source datasets. Conversely, synthetic multi-source datasets hold promise for advancing the last mile of UDA but remain underutilized in current research. Thus, we propose DEC, a flexible UDA framework for multi-source datasets. Following a divide-and-conquer strategy, DEC simplifies the task by categorizing semantic classes, training models for each category, and fusing their outputs by an ensemble model trained exclusively on synthetic datasets to obtain the final segmentation mask. DEC can integrate with existing UDA methods, achieving state-of-the-art performance on Cityscapes, BDD100K, and Mapillary Vistas, significantly narrowing the syn-to-real domain gap.
语义分割的无监督域自适应(UDA)的最后一公里是解决语法与实际域差距的挑战。最近的UDA方法取得了重大进展,但它们通常依赖于为合成单源数据集定制的策略(例如,GTA5),这限制了它们对多源数据集的推广。相反,合成多源数据集有望推进UDA的最后一英里,但在当前的研究中仍未得到充分利用。因此,我们提出了DEC,一个灵活的多源数据集UDA框架。采用分而治之的策略,DEC通过对语义类进行分类,为每个类别训练模型,并通过专门在合成数据集上训练的集成模型融合它们的输出来获得最终的分割掩码,从而简化了任务。DEC可以与现有的UDA方法集成,在cityscape、BDD100K和Mapillary远景上实现最先进的性能,显著缩小了语法与真实域的差距。
{"title":"Divide, Ensemble and Conquer: The Last Mile on Unsupervised Domain Adaptation for Semantic Segmentation","authors":"Tao Lian;Jose L. Gómez;Antonio M. López","doi":"10.1109/TIV.2024.3512995","DOIUrl":"https://doi.org/10.1109/TIV.2024.3512995","url":null,"abstract":"The last mile of unsupervised domain adaptation (UDA) for semantic segmentation is the challenge of solving the syn-to-real domain gap. Recent UDA methods have progressed significantly, yet they often rely on strategies customized for synthetic single-source datasets (e.g., GTA5), which limits their generalisation to multi-source datasets. Conversely, synthetic multi-source datasets hold promise for advancing the last mile of UDA but remain underutilized in current research. Thus, we propose DEC, a flexible UDA framework for multi-source datasets. Following a divide-and-conquer strategy, DEC simplifies the task by categorizing semantic classes, training models for each category, and fusing their outputs by an ensemble model trained exclusively on synthetic datasets to obtain the final segmentation mask. DEC can integrate with existing UDA methods, achieving state-of-the-art performance on Cityscapes, BDD100K, and Mapillary Vistas, significantly narrowing the syn-to-real domain gap.","PeriodicalId":36532,"journal":{"name":"IEEE Transactions on Intelligent Vehicles","volume":"10 12","pages":"5100-5111"},"PeriodicalIF":14.3,"publicationDate":"2024-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145772077","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Identifying Research Gaps Through Self-Driving Car Data Analysis 通过自动驾驶汽车数据分析识别研究差距
IF 14.3 1区 工程技术 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-12-04 DOI: 10.1109/TIV.2024.3506936
Mary L. Cummings;Ben Bauchwitz
There are currently around thirty companies testing self-driving cars in San Francisco, CA, effectively creating a living laboratory. Of these companies, only Waymo is engaged in commercial operations, while Zoox conducts routine driverless testing operations in San Francisco. Despite these successes, federal investigations have been opened into both companies for safety concerns, and Cruise is attempting to reinstate its permit after a near-fatal pedestrian crash. An analysis of these three companies’ crash data from required reporting illustrates that many areas of self-driving need improvement. The most significant crash type for Waymo and Zoox are struck-from-behind events, while Cruise struggled most with unexpected actions by others. Computer vision systems are very brittle and likely play an outsized role in crashes. Self-driving cars also struggle to reason under uncertainty, and simulations are not effectively bridging the physical-to-real-world testing gap. This analysis underscores that research is lacking, especially for artificial intelligence involving computer vision and reasoning under uncertainty.
目前,大约有30家公司在加州旧金山测试自动驾驶汽车,实际上创造了一个活生生的实验室。在这些公司中,只有Waymo从事商业运营,而Zoox则在旧金山进行常规的无人驾驶测试。尽管取得了这些成功,但出于安全考虑,联邦政府已经对两家公司展开了调查。在一次几乎致命的行人撞车事故后,克鲁斯正试图恢复其许可证。对这三家公司的事故数据的分析表明,自动驾驶的许多领域都需要改进。对Waymo和Zoox来说,最严重的事故类型是背后的袭击,而克鲁斯最担心的是其他人的意外行为。计算机视觉系统非常脆弱,很可能在坠机事故中发挥巨大作用。自动驾驶汽车也很难在不确定的情况下进行推理,而模拟并不能有效地弥合物理测试与现实测试之间的差距。这一分析强调了研究的缺乏,特别是涉及计算机视觉和不确定性推理的人工智能。
{"title":"Identifying Research Gaps Through Self-Driving Car Data Analysis","authors":"Mary L. Cummings;Ben Bauchwitz","doi":"10.1109/TIV.2024.3506936","DOIUrl":"https://doi.org/10.1109/TIV.2024.3506936","url":null,"abstract":"There are currently around thirty companies testing self-driving cars in San Francisco, CA, effectively creating a living laboratory. Of these companies, only Waymo is engaged in commercial operations, while Zoox conducts routine driverless testing operations in San Francisco. Despite these successes, federal investigations have been opened into both companies for safety concerns, and Cruise is attempting to reinstate its permit after a near-fatal pedestrian crash. An analysis of these three companies’ crash data from required reporting illustrates that many areas of self-driving need improvement. The most significant crash type for Waymo and Zoox are struck-from-behind events, while Cruise struggled most with unexpected actions by others. Computer vision systems are very brittle and likely play an outsized role in crashes. Self-driving cars also struggle to reason under uncertainty, and simulations are not effectively bridging the physical-to-real-world testing gap. This analysis underscores that research is lacking, especially for artificial intelligence involving computer vision and reasoning under uncertainty.","PeriodicalId":36532,"journal":{"name":"IEEE Transactions on Intelligent Vehicles","volume":"10 11","pages":"4903-4912"},"PeriodicalIF":14.3,"publicationDate":"2024-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10778107","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145665808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LiDAR in Connected and Autonomous Vehicles: Perception, Threat Model, and Defense 互联和自动驾驶汽车中的激光雷达:感知、威胁模型和防御
IF 14.3 1区 工程技术 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-12-04 DOI: 10.1109/TIV.2024.3510787
Muhammad Asif Khan;Hamid Menouar;Mohamed Abdallah;Adnan Abu-Dayya
Connected and Autonomous Vehicles (CAVs) are referred to as self-driving vehicles that will become an essential component of future intelligent transportation systems. These CAVs will be equipped with various sensors for perceiving their surroundings and onboard computing capabilities to process sensor data in real-time. Light Detection and Ranging (LiDAR) is one of the essential sensors used for detecting objects and accurate distance estimation. However, LiDAR sensors are susceptible to several types of attacks. Adversaries can exploit LiDAR sensors either physically, by sending signals directly to the sensor, or digitally, by manipulating LiDAR data after gaining access to the in-vehicle network. Over the past few years, there has been significant research on the vulnerabilities, attack models, and security of LiDAR sensors. However, to our knowledge, no comprehensive survey exists that addresses these aspects of autonomous vehicle security. This paper aims to bridge this gap by presenting an overview of LiDAR-based perception, data processing, threat models, and defense mechanisms for LiDAR sensors in CAVs. We believe this paper will serve as a valuable reference for researchers, providing a clear understanding of cyber-physical attacks and defense strategies related to LiDAR sensors in autonomous vehicles and related fields.
联网和自动驾驶车辆(cav)被称为自动驾驶车辆,将成为未来智能交通系统的重要组成部分。这些自动驾驶汽车将配备各种传感器来感知周围环境,并配备车载计算能力来实时处理传感器数据。光探测与测距(LiDAR)是用于物体探测和精确距离估计的重要传感器之一。然而,激光雷达传感器容易受到几种类型的攻击。攻击者可以通过物理方式利用激光雷达传感器,直接向传感器发送信号,也可以通过数字方式利用激光雷达数据,在进入车载网络后操纵激光雷达数据。在过去的几年里,人们对激光雷达传感器的漏洞、攻击模型和安全性进行了大量的研究。然而,据我们所知,目前还没有针对自动驾驶汽车安全这些方面的全面调查。本文旨在通过概述基于激光雷达的感知、数据处理、威胁模型和自动驾驶汽车中激光雷达传感器的防御机制来弥合这一差距。我们相信本文将为研究人员提供有价值的参考,为自动驾驶汽车和相关领域中与LiDAR传感器相关的网络物理攻击和防御策略提供清晰的认识。
{"title":"LiDAR in Connected and Autonomous Vehicles: Perception, Threat Model, and Defense","authors":"Muhammad Asif Khan;Hamid Menouar;Mohamed Abdallah;Adnan Abu-Dayya","doi":"10.1109/TIV.2024.3510787","DOIUrl":"https://doi.org/10.1109/TIV.2024.3510787","url":null,"abstract":"Connected and Autonomous Vehicles (CAVs) are referred to as self-driving vehicles that will become an essential component of future intelligent transportation systems. These CAVs will be equipped with various sensors for perceiving their surroundings and onboard computing capabilities to process sensor data in real-time. Light Detection and Ranging (LiDAR) is one of the essential sensors used for detecting objects and accurate distance estimation. However, LiDAR sensors are susceptible to several types of attacks. Adversaries can exploit LiDAR sensors either physically, by sending signals directly to the sensor, or digitally, by manipulating LiDAR data after gaining access to the in-vehicle network. Over the past few years, there has been significant research on the vulnerabilities, attack models, and security of LiDAR sensors. However, to our knowledge, no comprehensive survey exists that addresses these aspects of autonomous vehicle security. This paper aims to bridge this gap by presenting an overview of LiDAR-based perception, data processing, threat models, and defense mechanisms for LiDAR sensors in CAVs. We believe this paper will serve as a valuable reference for researchers, providing a clear understanding of cyber-physical attacks and defense strategies related to LiDAR sensors in autonomous vehicles and related fields.","PeriodicalId":36532,"journal":{"name":"IEEE Transactions on Intelligent Vehicles","volume":"10 11","pages":"5023-5041"},"PeriodicalIF":14.3,"publicationDate":"2024-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145665722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Synthesizing Realistic Traffic Events From UAV Perspectives: A Mask-Guided Generative Approach Based on Style-Modulated Transformer 无人机视角下的真实交通事件综合:基于样式调制变压器的掩模引导生成方法
IF 14.3 1区 工程技术 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-12-03 DOI: 10.1109/TIV.2024.3510563
Wei Zhou;Nan Zheng;Chen Wang
Unmanned aerial vehicles (UAVs) have emerged as valuable tools in intelligent transportation systems, offering potential for real-time traffic monitoring and emergency response. However, the limited flight endurance of UAVs restricts their ability to collect large amounts of traffic event data crucial for data-driven models. While simulation platforms provide an alternative data source, a significant visual gap persists between synthetic and real images. Given the unique characteristics of UAV perspectives, where objects typically appear smaller, and the challenges faced by existing generative approaches in maintaining structural and semantic consistency, this paper proposes a novel generative approach. Our method integrates semantic mask guidance with a style-modulated Transformer-based GAN architecture to address these issues. Our approach first utilizes the Segment-Anything Model (SAM) and Contrastive Language-Image Pre-training (CLIP) to extract semantic masks from GTA V simulated images. Subsequently, we introduce Spatially-Attentive Denormalization Modules (SADM) within the generator. These modules incorporate semantic mask statistics to enhance image quality and maintain consistency. Furthermore, we develop a perceptual discriminator incorporating a “memory bank” mechanism to more effectively evaluate image realism and stabilize the training process. To further enhance the quality of generated images, we develop a comprehensive training strategy. Our experimental results demonstrate that the proposed method outperforms existing state-of-the-art approaches in both quantitative metrics (i.e., FID and KID) and qualitative visual assessments, thus highlighting its effectiveness and superiority. Overall, our approach offers a robust solution for generating highly realistic traffic event images from UAV perspectives, effectively addressing the scarcity of real-world UAV-recorded traffic event data.
无人驾驶飞行器(uav)已经成为智能交通系统中有价值的工具,为实时交通监控和应急响应提供了潜力。然而,无人机有限的飞行续航力限制了它们收集大量交通事件数据的能力,这些数据对数据驱动模型至关重要。虽然仿真平台提供了另一种数据源,但合成图像和真实图像之间存在显著的视觉差距。考虑到无人机视角的独特特征,其中物体通常看起来更小,以及现有生成方法在保持结构和语义一致性方面面临的挑战,本文提出了一种新的生成方法。我们的方法将语义掩码引导与基于样式调制变压器的GAN架构集成在一起,以解决这些问题。我们的方法首先利用分段-任意模型(SAM)和对比语言-图像预训练(CLIP)从GTA V模拟图像中提取语义掩码。随后,我们在生成器中引入了空间关注反规范化模块(SADM)。这些模块包含语义掩码统计,以提高图像质量并保持一致性。此外,我们开发了一个包含“记忆库”机制的感知鉴别器,以更有效地评估图像真实感并稳定训练过程。为了进一步提高生成图像的质量,我们制定了一套全面的训练策略。我们的实验结果表明,所提出的方法在定量指标(即FID和KID)和定性视觉评估方面都优于现有的最先进方法,从而突出了其有效性和优越性。总的来说,我们的方法为从无人机角度生成高度逼真的交通事件图像提供了一个强大的解决方案,有效地解决了现实世界中无人机记录的交通事件数据的稀缺性。
{"title":"Synthesizing Realistic Traffic Events From UAV Perspectives: A Mask-Guided Generative Approach Based on Style-Modulated Transformer","authors":"Wei Zhou;Nan Zheng;Chen Wang","doi":"10.1109/TIV.2024.3510563","DOIUrl":"https://doi.org/10.1109/TIV.2024.3510563","url":null,"abstract":"Unmanned aerial vehicles (UAVs) have emerged as valuable tools in intelligent transportation systems, offering potential for real-time traffic monitoring and emergency response. However, the limited flight endurance of UAVs restricts their ability to collect large amounts of traffic event data crucial for data-driven models. While simulation platforms provide an alternative data source, a significant visual gap persists between synthetic and real images. Given the unique characteristics of UAV perspectives, where objects typically appear smaller, and the challenges faced by existing generative approaches in maintaining structural and semantic consistency, this paper proposes a novel generative approach. Our method integrates semantic mask guidance with a style-modulated Transformer-based GAN architecture to address these issues. Our approach first utilizes the Segment-Anything Model (SAM) and Contrastive Language-Image Pre-training (CLIP) to extract semantic masks from GTA V simulated images. Subsequently, we introduce Spatially-Attentive Denormalization Modules (SADM) within the generator. These modules incorporate semantic mask statistics to enhance image quality and maintain consistency. Furthermore, we develop a perceptual discriminator incorporating a “memory bank” mechanism to more effectively evaluate image realism and stabilize the training process. To further enhance the quality of generated images, we develop a comprehensive training strategy. Our experimental results demonstrate that the proposed method outperforms existing state-of-the-art approaches in both quantitative metrics (i.e., FID and KID) and qualitative visual assessments, thus highlighting its effectiveness and superiority. Overall, our approach offers a robust solution for generating highly realistic traffic event images from UAV perspectives, effectively addressing the scarcity of real-world UAV-recorded traffic event data.","PeriodicalId":36532,"journal":{"name":"IEEE Transactions on Intelligent Vehicles","volume":"10 11","pages":"5008-5022"},"PeriodicalIF":14.3,"publicationDate":"2024-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145665815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Intelligent Vehicles
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1