首页 > 最新文献

2022 International Conference on Robotics and Automation (ICRA)最新文献

英文 中文
Deep Reinforcement Learning for Next-Best-View Planning in Agricultural Applications 农业应用中次优视图规划的深度强化学习
Pub Date : 2022-05-23 DOI: 10.1109/icra46639.2022.9811800
Xiangyu Zeng, Tobias Zaenker, Maren Bennewitz
Automated agricultural applications, i.e., fruit picking require spatial information about crops and, especially, their fruits. In this paper, we present a novel deep reinforcement learning (DRL) approach to determine the next best view for automatic exploration of 3D environments with a robotic arm equipped with an RGB-D camera. We process the obtained images into an octree with labeled regions of interest (ROIs), i.e., fruits. We use this octree to generate 3D observation maps that serve as encoded input to the DRL network. We hereby do not only rely on known information about the environment, but explicitly also represent information about the unknown space to force exploration. Our network takes as input the encoded 3D observation map and the temporal sequence of camera view pose changes, and outputs the most promising camera movement direction. Our experimental results show an improved ROI targeted exploration performance resulting from our learned network in comparison to a state-of-the-art method.
自动化农业应用,如水果采摘,需要有关作物,特别是其果实的空间信息。在本文中,我们提出了一种新的深度强化学习(DRL)方法,以确定配备RGB-D相机的机械臂自动探索3D环境的下一个最佳视图。我们将获得的图像处理成带有标记感兴趣区域(roi)的八叉树,即水果。我们使用这个八叉树来生成3D观测图,作为DRL网络的编码输入。我们在这里不仅依靠已知的环境信息,而且明确地表示未知空间的信息来强制探索。我们的网络以编码后的三维观测图和摄像机视角姿态变化的时间序列作为输入,输出最有希望的摄像机运动方向。我们的实验结果表明,与最先进的方法相比,我们的学习网络提高了ROI目标勘探性能。
{"title":"Deep Reinforcement Learning for Next-Best-View Planning in Agricultural Applications","authors":"Xiangyu Zeng, Tobias Zaenker, Maren Bennewitz","doi":"10.1109/icra46639.2022.9811800","DOIUrl":"https://doi.org/10.1109/icra46639.2022.9811800","url":null,"abstract":"Automated agricultural applications, i.e., fruit picking require spatial information about crops and, especially, their fruits. In this paper, we present a novel deep reinforcement learning (DRL) approach to determine the next best view for automatic exploration of 3D environments with a robotic arm equipped with an RGB-D camera. We process the obtained images into an octree with labeled regions of interest (ROIs), i.e., fruits. We use this octree to generate 3D observation maps that serve as encoded input to the DRL network. We hereby do not only rely on known information about the environment, but explicitly also represent information about the unknown space to force exploration. Our network takes as input the encoded 3D observation map and the temporal sequence of camera view pose changes, and outputs the most promising camera movement direction. Our experimental results show an improved ROI targeted exploration performance resulting from our learned network in comparison to a state-of-the-art method.","PeriodicalId":341244,"journal":{"name":"2022 International Conference on Robotics and Automation (ICRA)","volume":"86 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130504529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
RangeBird: Multi View Panoptic Segmentation of 3D Point Clouds with Neighborhood Attention RangeBird:基于邻域关注的三维点云多视场分割
Pub Date : 2022-05-23 DOI: 10.1109/icra46639.2022.9811998
Fabian Duerr, H. Weigel, J. Beyerer
Panoptic segmentation of point clouds is one of the key challenges of 3D scene understanding, requiring the simultaneous prediction of semantics and object instances. Tasks like autonomous driving strongly depend on these information to get a holistic understanding of their 3D environment. This work presents a novel proposal free framework for lidar-based panoptic segmentation, which exploits three different point cloud representations, leveraging their strengths and compensating their weaknesses. The efficient projection-based range view and bird's eye view are combined and further extended by a point-based network with a novel attention-based neighborhood aggregation for improved semantic features. Cluster-based object recognition in bird's eye view enables an efficient and high-quality instance segmentation. Semantic and instance segmentation are fused and further refined by a novel instance classification for the final panoptic segmentation. The results on two challenging large-scale datasets, nuScenes and SemanticKITTI, show the success of the proposed framework, which outperforms all existing approaches on nuScenes and achieves state-of-the-art results on SemanticKITTI.
点云的全视分割是3D场景理解的关键挑战之一,需要同时预测语义和对象实例。像自动驾驶这样的任务强烈依赖于这些信息来全面了解他们的3D环境。这项工作提出了一种新的基于激光雷达的全光分割框架,该框架利用三种不同的点云表示,利用它们的优点并弥补它们的缺点。将高效的基于投影的距离视图和鸟瞰视图结合起来,并通过基于点的网络进行扩展,并采用新颖的基于注意力的邻域聚合来改进语义特征。鸟瞰图中基于聚类的目标识别实现了高效、高质量的实例分割。将语义分割和实例分割融合,并通过一种新的实例分类进一步细化,最终实现全视分割。在nuScenes和SemanticKITTI两个具有挑战性的大规模数据集上的结果表明,所提出的框架是成功的,它优于所有现有的nuScenes方法,并在SemanticKITTI上取得了最先进的结果。
{"title":"RangeBird: Multi View Panoptic Segmentation of 3D Point Clouds with Neighborhood Attention","authors":"Fabian Duerr, H. Weigel, J. Beyerer","doi":"10.1109/icra46639.2022.9811998","DOIUrl":"https://doi.org/10.1109/icra46639.2022.9811998","url":null,"abstract":"Panoptic segmentation of point clouds is one of the key challenges of 3D scene understanding, requiring the simultaneous prediction of semantics and object instances. Tasks like autonomous driving strongly depend on these information to get a holistic understanding of their 3D environment. This work presents a novel proposal free framework for lidar-based panoptic segmentation, which exploits three different point cloud representations, leveraging their strengths and compensating their weaknesses. The efficient projection-based range view and bird's eye view are combined and further extended by a point-based network with a novel attention-based neighborhood aggregation for improved semantic features. Cluster-based object recognition in bird's eye view enables an efficient and high-quality instance segmentation. Semantic and instance segmentation are fused and further refined by a novel instance classification for the final panoptic segmentation. The results on two challenging large-scale datasets, nuScenes and SemanticKITTI, show the success of the proposed framework, which outperforms all existing approaches on nuScenes and achieves state-of-the-art results on SemanticKITTI.","PeriodicalId":341244,"journal":{"name":"2022 International Conference on Robotics and Automation (ICRA)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129217141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Kinematic Structure Estimation of Arbitrary Articulated Rigid Objects for Event Cameras 事件相机中任意关节刚体的运动结构估计
Pub Date : 2022-05-23 DOI: 10.1109/icra46639.2022.9812430
Urbano Miguel Nunes, Y. Demiris
We propose a novel method that estimates the Kinematic Structure (KS) of arbitrary articulated rigid objects from event-based data. Event cameras are emerging sensors that asynchronously report brightness changes with a time resolution of microseconds, making them suitable candidates for motion-related perception. By assuming that an articulated rigid object is composed of body parts whose shape can be approximately described by a Gaussian distribution, we jointly segment the different parts by combining an adapted Bayesian inference approach and incremental event-based motion estimation. The respective KS is then generated based on the segmented parts and their respective biharmonic distance, which is estimated by building an affinity matrix of points sampled from the estimated Gaussian distributions. The method outperforms frame-based methods in sequences obtained by simulating events from video sequences and achieves a solid performance on new high-speed motions sequences, which frame-based KS estimation methods can not handle.
提出了一种基于事件数据估计任意关节刚体运动结构的新方法。事件相机是一种新兴的传感器,它以微秒的时间分辨率异步报告亮度变化,使其成为运动相关感知的合适人选。假设一个铰接的刚性物体由形状可以用高斯分布近似描述的身体部分组成,我们结合自适应贝叶斯推理方法和基于增量事件的运动估计来共同分割不同的部分。然后根据被分割的部分及其各自的双谐波距离生成各自的KS,该距离通过从估计的高斯分布中采样的点建立亲和矩阵来估计。该方法在通过模拟视频序列获得的序列上优于基于帧的方法,并且在新的高速运动序列上取得了基于帧的KS估计方法无法处理的良好性能。
{"title":"Kinematic Structure Estimation of Arbitrary Articulated Rigid Objects for Event Cameras","authors":"Urbano Miguel Nunes, Y. Demiris","doi":"10.1109/icra46639.2022.9812430","DOIUrl":"https://doi.org/10.1109/icra46639.2022.9812430","url":null,"abstract":"We propose a novel method that estimates the Kinematic Structure (KS) of arbitrary articulated rigid objects from event-based data. Event cameras are emerging sensors that asynchronously report brightness changes with a time resolution of microseconds, making them suitable candidates for motion-related perception. By assuming that an articulated rigid object is composed of body parts whose shape can be approximately described by a Gaussian distribution, we jointly segment the different parts by combining an adapted Bayesian inference approach and incremental event-based motion estimation. The respective KS is then generated based on the segmented parts and their respective biharmonic distance, which is estimated by building an affinity matrix of points sampled from the estimated Gaussian distributions. The method outperforms frame-based methods in sequences obtained by simulating events from video sequences and achieves a solid performance on new high-speed motions sequences, which frame-based KS estimation methods can not handle.","PeriodicalId":341244,"journal":{"name":"2022 International Conference on Robotics and Automation (ICRA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130674387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Continuum Robot Surface of Woven, McKibben Muscles Embedded in and Giving Shape to Rooms 一个连续的机器人表面编织,麦基本肌肉嵌入和形状的房间
Pub Date : 2022-05-23 DOI: 10.1109/icra46639.2022.9811987
G. Tan, Harrison Hidalgo, H. Kao, I. Walker, K. Green
Robots are typically designed as occupants of rooms, adapting to, and navigating within them. “Robot surfaces,” an emerging robot typology, are not occupants of but integral with rooms, physically shaping rooms to support human activity. We report on an advancement of robot surfaces formed by weaving McKibben Pneumatic Air Muscles that, when actuated, morph a 2D planar surface to generate 3D geometries including a “spherical cap.” Following our foundational study at different scales with different materials, we developed a full-scale prototype that offers an intimate and private space for people meeting in open plan environments. We report on our research, focusing on a design case, and validate the full-scale prototype as compared to our Non-Uniform Rational B-Splines (NURBS) model for three useful configurations. Our quantitative and qualitative results suggest that our robot surface can support human activity as envisioned. This research contributes foundational understanding of an emerging category of robotics from which our team and peers can build.
机器人通常被设计成房间的居住者,适应并在其中导航。“机器人表面”是一种新兴的机器人类型学,它不是房间的居住者,而是与房间融为一体,通过物理方式塑造房间以支持人类活动。我们报告了通过编织McKibben气动空气肌肉形成的机器人表面的进步,当驱动时,可以将二维平面变形为三维几何形状,包括“球形帽”。在我们对不同材料的不同尺度进行基础研究之后,我们开发了一个全尺寸原型,为人们在开放式环境中会面提供了一个亲密和私人的空间。我们报告了我们的研究,专注于一个设计案例,并验证了与我们的非均匀有理b样条(NURBS)模型相比的全尺寸原型,以获得三种有用的配置。我们的定量和定性结果表明,我们的机器人表面可以支持人类活动的设想。这项研究为我们的团队和同行可以建立的新兴机器人类别提供了基础理解。
{"title":"A Continuum Robot Surface of Woven, McKibben Muscles Embedded in and Giving Shape to Rooms","authors":"G. Tan, Harrison Hidalgo, H. Kao, I. Walker, K. Green","doi":"10.1109/icra46639.2022.9811987","DOIUrl":"https://doi.org/10.1109/icra46639.2022.9811987","url":null,"abstract":"Robots are typically designed as occupants of rooms, adapting to, and navigating within them. “Robot surfaces,” an emerging robot typology, are not occupants of but integral with rooms, physically shaping rooms to support human activity. We report on an advancement of robot surfaces formed by weaving McKibben Pneumatic Air Muscles that, when actuated, morph a 2D planar surface to generate 3D geometries including a “spherical cap.” Following our foundational study at different scales with different materials, we developed a full-scale prototype that offers an intimate and private space for people meeting in open plan environments. We report on our research, focusing on a design case, and validate the full-scale prototype as compared to our Non-Uniform Rational B-Splines (NURBS) model for three useful configurations. Our quantitative and qualitative results suggest that our robot surface can support human activity as envisioned. This research contributes foundational understanding of an emerging category of robotics from which our team and peers can build.","PeriodicalId":341244,"journal":{"name":"2022 International Conference on Robotics and Automation (ICRA)","volume":"31 8","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114085593","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Looking for Trouble: Informative Planning for Safe Trajectories with Occlusions 寻找麻烦:信息规划与闭塞的安全轨迹
Pub Date : 2022-05-23 DOI: 10.1109/icra46639.2022.9811994
Barry Gilhuly, Armin Sadeghi, P. Yadmellat, K. Rezaee, Stephen L. Smith
Planning a safe trajectory for an ego vehicle through an environment with occluded regions is a challenging task. Existing methods use some combination of metrics to evaluate a trajectory, either taking a worst case view or allowing for some probabilistic estimate, to eliminate or minimize the risk of collision respectively. Typically, these approaches assume occluded regions of the environment are unsafe and must be avoided, resulting in overly conservative trajectories-particularly when there are no hidden risks present. We propose a local trajectory planning algorithm which generates safe trajectories that maximize observations on un-certain regions. In particular, we seek to gain information on occluded areas that are most likely to pose a risk to the ego vehicle on its future path. Calculating the information gain is a computationally complex problem; our method approximates the maximum information gain and results in vehicle motion that remains safe but is less conservative than state-of-the-art approaches. We evaluate the performance of the proposed method within the CARLA simulator in different scenarios.
为自动驾驶汽车规划安全的行驶轨迹是一项具有挑战性的任务。现有的方法使用一些度量的组合来评估轨迹,要么采取最坏情况的观点,要么允许一些概率估计,分别消除或最小化碰撞的风险。通常,这些方法假设环境中被遮挡的区域是不安全的,必须避开,导致轨迹过于保守——特别是在没有隐藏风险的情况下。我们提出了一种局部轨迹规划算法,该算法生成安全轨迹,使不确定区域的观测值最大化。特别是,我们寻求获得关于最有可能对自我车辆未来路径构成风险的闭塞区域的信息。计算信息增益是一个计算复杂的问题;我们的方法近似于最大信息增益,结果在车辆运动中保持安全,但比最先进的方法更保守。我们在不同的场景下在CARLA模拟器中评估了所提出的方法的性能。
{"title":"Looking for Trouble: Informative Planning for Safe Trajectories with Occlusions","authors":"Barry Gilhuly, Armin Sadeghi, P. Yadmellat, K. Rezaee, Stephen L. Smith","doi":"10.1109/icra46639.2022.9811994","DOIUrl":"https://doi.org/10.1109/icra46639.2022.9811994","url":null,"abstract":"Planning a safe trajectory for an ego vehicle through an environment with occluded regions is a challenging task. Existing methods use some combination of metrics to evaluate a trajectory, either taking a worst case view or allowing for some probabilistic estimate, to eliminate or minimize the risk of collision respectively. Typically, these approaches assume occluded regions of the environment are unsafe and must be avoided, resulting in overly conservative trajectories-particularly when there are no hidden risks present. We propose a local trajectory planning algorithm which generates safe trajectories that maximize observations on un-certain regions. In particular, we seek to gain information on occluded areas that are most likely to pose a risk to the ego vehicle on its future path. Calculating the information gain is a computationally complex problem; our method approximates the maximum information gain and results in vehicle motion that remains safe but is less conservative than state-of-the-art approaches. We evaluate the performance of the proposed method within the CARLA simulator in different scenarios.","PeriodicalId":341244,"journal":{"name":"2022 International Conference on Robotics and Automation (ICRA)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114615387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Perception Engine Using a Multi-Sensor Head to Enable High-level Humanoid Robot Behaviors 使用多传感器头的感知引擎实现高级人形机器人行为
Pub Date : 2022-05-23 DOI: 10.1109/icra46639.2022.9812178
Bhavyansh Mishra, Duncan Calvert, Brendon Ortolano, M. Asselmeier, Luke Fina, Stephen McCrory, H. Sevil, Robert J. Griffin
For achieving significant levels of autonomy, legged robot behaviors require perceptual awareness of both the terrain for traversal, as well as structures and objects in their surroundings for planning, obstacle avoidance, and high-level decision making. In this work, we present a perception engine for legged robots that extracts the necessary information for developing semantic, contextual, and metric awareness of their surroundings. Our custom sensor configuration consists of (1) an active depth sensor, (2) two monocular cameras looking sideways, (3) a passive stereo sensor observing the terrain, (4) a forward facing active depth camera, and (5) a rotating 3D LIDAR with a large vertical field-of-view (FOV). The mutual overlap in the sensors' FOVs allows us to redundantly detect and track objects of both dynamic and static types. We fuse class masks generated by a semantic segmentation model with LIDAR and depth data to accurately identify and track individual instances of dynamically moving objects. In parallel, active depth and passive stereo streams of the terrain are also fused to map the terrain using the on-board GPU. We evaluate the engine using two different humanoid behaviors, (1) look-and-step and (2) track-and-follow, on the Boston Dynamics Atlas.
为了实现显著的自主性,有腿机器人的行为需要对穿越的地形以及周围的结构和物体进行感知,以进行规划、避障和高级决策。在这项工作中,我们提出了一种用于有腿机器人的感知引擎,该引擎可以提取必要的信息,以发展对周围环境的语义、上下文和度量意识。我们定制的传感器配置包括(1)一个主动深度传感器,(2)两个侧视的单目摄像头,(3)一个观察地形的被动立体传感器,(4)一个面向前方的主动深度摄像头,以及(5)一个具有大垂直视场(FOV)的旋转3D激光雷达。传感器fov的相互重叠使我们能够冗余地检测和跟踪动态和静态类型的对象。我们将语义分割模型生成的类掩码与激光雷达和深度数据融合在一起,以准确识别和跟踪动态移动物体的单个实例。同时,地形的主动深度和被动立体流也被融合到使用车载GPU绘制地形。我们在波士顿动力地图集上使用两种不同的类人行为(1)观察-步和(2)跟踪-跟随)来评估引擎。
{"title":"Perception Engine Using a Multi-Sensor Head to Enable High-level Humanoid Robot Behaviors","authors":"Bhavyansh Mishra, Duncan Calvert, Brendon Ortolano, M. Asselmeier, Luke Fina, Stephen McCrory, H. Sevil, Robert J. Griffin","doi":"10.1109/icra46639.2022.9812178","DOIUrl":"https://doi.org/10.1109/icra46639.2022.9812178","url":null,"abstract":"For achieving significant levels of autonomy, legged robot behaviors require perceptual awareness of both the terrain for traversal, as well as structures and objects in their surroundings for planning, obstacle avoidance, and high-level decision making. In this work, we present a perception engine for legged robots that extracts the necessary information for developing semantic, contextual, and metric awareness of their surroundings. Our custom sensor configuration consists of (1) an active depth sensor, (2) two monocular cameras looking sideways, (3) a passive stereo sensor observing the terrain, (4) a forward facing active depth camera, and (5) a rotating 3D LIDAR with a large vertical field-of-view (FOV). The mutual overlap in the sensors' FOVs allows us to redundantly detect and track objects of both dynamic and static types. We fuse class masks generated by a semantic segmentation model with LIDAR and depth data to accurately identify and track individual instances of dynamically moving objects. In parallel, active depth and passive stereo streams of the terrain are also fused to map the terrain using the on-board GPU. We evaluate the engine using two different humanoid behaviors, (1) look-and-step and (2) track-and-follow, on the Boston Dynamics Atlas.","PeriodicalId":341244,"journal":{"name":"2022 International Conference on Robotics and Automation (ICRA)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116251121","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Tightly-coupled GNSS-aided Visual-Inertial Localization 紧密耦合gnss辅助视觉惯性定位
Pub Date : 2022-05-23 DOI: 10.1109/icra46639.2022.9811362
W. Lee, Patrick Geneva, Yulin Yang, G. Huang
A navigation system which can output drift-free global trajectory estimation with local consistency holds great potential for autonomous vehicles and mobile devices. We propose a tightly-coupled GNSS-aided visual-inertial navigation system (GAINS) which is able to leverage the complementary sensing modality from a visual-inertial sensing pair, which provides high-frequency local information, and a Global Navigation Satellite System (GNSS) receiver with low-frequency global observations. Specifically, the raw GNSS measurements (including pseudorange, carrier phase changes, and Doppler frequency shift) are carefully leveraged and tightly fused within a visual-inertial framework. The proposed GAINS can accurately model the raw measurement uncertainties by canceling the atmospheric effects (e.g., ionospheric and tropospheric delays) which requires no prior model information. A robust state initialization procedure is presented to facilitate the fusion of global GNSS information with local visual-inertial odometry, and the spatiotemporal calibration between IMU-GNSS are also optimized in the estimator. The proposed GAINS is evaluated on extensive Monte-Carlo simulations on a trajectory generated from a large-scale urban driving dataset with specific verification for each component (i.e., online calibration and system initialization). GAINS also demonstrates competitive performance against existing state-of-the-art methods on a publicly available dataset with ground truth.
能够输出具有局部一致性的无漂移全局轨迹估计的导航系统在自动驾驶汽车和移动设备中具有很大的潜力。我们提出了一种紧密耦合的GNSS辅助视觉惯性导航系统(gain),该系统能够利用视觉惯性传感对的互补传感方式,提供高频局部信息,并利用全球导航卫星系统(GNSS)接收器提供低频全球观测。具体来说,原始GNSS测量(包括伪距、载波相位变化和多普勒频移)被仔细利用,并在视觉惯性框架内紧密融合。所提出的增益增益可以通过消除不需要先验模型信息的大气影响(例如电离层和对流层延迟)来准确地模拟原始测量不确定性。提出了一种鲁棒状态初始化方法,实现了全局GNSS信息与局部视觉惯性里程计的融合,并对IMU-GNSS间的时空标定进行了优化。所提出的增益在大规模城市驾驶数据集生成的轨迹上进行了广泛的蒙特卡罗模拟,并对每个组件进行了特定的验证(即在线校准和系统初始化)。gain还在公开可用的数据集上展示了与现有最先进方法的竞争性能。
{"title":"Tightly-coupled GNSS-aided Visual-Inertial Localization","authors":"W. Lee, Patrick Geneva, Yulin Yang, G. Huang","doi":"10.1109/icra46639.2022.9811362","DOIUrl":"https://doi.org/10.1109/icra46639.2022.9811362","url":null,"abstract":"A navigation system which can output drift-free global trajectory estimation with local consistency holds great potential for autonomous vehicles and mobile devices. We propose a tightly-coupled GNSS-aided visual-inertial navigation system (GAINS) which is able to leverage the complementary sensing modality from a visual-inertial sensing pair, which provides high-frequency local information, and a Global Navigation Satellite System (GNSS) receiver with low-frequency global observations. Specifically, the raw GNSS measurements (including pseudorange, carrier phase changes, and Doppler frequency shift) are carefully leveraged and tightly fused within a visual-inertial framework. The proposed GAINS can accurately model the raw measurement uncertainties by canceling the atmospheric effects (e.g., ionospheric and tropospheric delays) which requires no prior model information. A robust state initialization procedure is presented to facilitate the fusion of global GNSS information with local visual-inertial odometry, and the spatiotemporal calibration between IMU-GNSS are also optimized in the estimator. The proposed GAINS is evaluated on extensive Monte-Carlo simulations on a trajectory generated from a large-scale urban driving dataset with specific verification for each component (i.e., online calibration and system initialization). GAINS also demonstrates competitive performance against existing state-of-the-art methods on a publicly available dataset with ground truth.","PeriodicalId":341244,"journal":{"name":"2022 International Conference on Robotics and Automation (ICRA)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116322557","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Underwater Dock Detection through Convolutional Neural Networks Trained with Artificial Image Generation 基于人工图像生成训练的卷积神经网络水下船坞检测
Pub Date : 2022-05-23 DOI: 10.1109/icra46639.2022.9812143
Jalil Chavez-Galaviz, N. Mahmoudian
Autonomous Underwater Vehicles (AUVs) are a vital element for ocean exploration in various applications; however, energy sustainability still limits long-term operations. An option to overcome this problem is using underwater docking for power and data transfer. To robustly guide an AUV into a docking station, we propose an underwater vision algorithm for short-distance detection. In this paper, we present a Convolutional Neural Network architecture to accurately estimate the dock position during the terminal homing stage of the docking. Additionally, to alleviate the lack of available underwater datasets, two methods are proposed to generate synthetic datasets, one using a CycleGAN network, and another using Artistic Style transfer network. Both methods are used to train the same CNN architecture to compare the results. Finally, implementation details of the CNN are presented under the backseat architecture and ROS framework, running on an IVER3 AUV.
在各种应用中,自主水下航行器(auv)是海洋探测的重要组成部分。然而,能源可持续性仍然限制了长期运营。解决这个问题的一种方法是使用水下对接进行电力和数据传输。为了鲁棒地引导AUV进入坞站,我们提出了一种用于短距离检测的水下视觉算法。在本文中,我们提出了一种卷积神经网络架构,以准确估计码头在码头归航阶段的位置。此外,为了缓解可用水下数据集的不足,提出了两种生成合成数据集的方法,一种使用CycleGAN网络,另一种使用艺术风格转移网络。两种方法都用于训练相同的CNN架构,以比较结果。最后,介绍了在后座架构和ROS框架下,在IVER3 AUV上运行的CNN的实现细节。
{"title":"Underwater Dock Detection through Convolutional Neural Networks Trained with Artificial Image Generation","authors":"Jalil Chavez-Galaviz, N. Mahmoudian","doi":"10.1109/icra46639.2022.9812143","DOIUrl":"https://doi.org/10.1109/icra46639.2022.9812143","url":null,"abstract":"Autonomous Underwater Vehicles (AUVs) are a vital element for ocean exploration in various applications; however, energy sustainability still limits long-term operations. An option to overcome this problem is using underwater docking for power and data transfer. To robustly guide an AUV into a docking station, we propose an underwater vision algorithm for short-distance detection. In this paper, we present a Convolutional Neural Network architecture to accurately estimate the dock position during the terminal homing stage of the docking. Additionally, to alleviate the lack of available underwater datasets, two methods are proposed to generate synthetic datasets, one using a CycleGAN network, and another using Artistic Style transfer network. Both methods are used to train the same CNN architecture to compare the results. Finally, implementation details of the CNN are presented under the backseat architecture and ROS framework, running on an IVER3 AUV.","PeriodicalId":341244,"journal":{"name":"2022 International Conference on Robotics and Automation (ICRA)","volume":"17 2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113964910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Novel Limbs-Free Variable Structure Wheelchair based on Face-Computer Interface (FCI) with Shared Control 一种基于人脸-计算机界面共享控制的新型无肢变结构轮椅
Pub Date : 2022-05-23 DOI: 10.1109/icra46639.2022.9811571
Bo Zhu, Daohui Zhang, Yaqi Chu, Xingang Zhao
In order to meet the mobility and physical activity needs of people with impaired limbs function, a novel limbs-free variable structure wheelchair system controled by face-computer interface (FCI) was developed in this study. FCI used facial electromyography (fEMG) as a human intention recognition method from 6 facial movements, and the accuracy of intent recognition reached 97.6% under a series of offline optimization including channel optimization based on the Hilbert transform to obtain the envelope of fEMG, features optimization, and channel-independent model optimization. A collection of finite state machines (FSM) was used to control the movement and structural changes of the wheelchair. A shared control strategy called “ Keep Action after Take Over (KAaTO) “ that can reduce user fatigue while increasing safety was used in long-distance movement control of wheelchair. To test the performance of the system, in the braking distance test experiment, the result of 0.429m under KAaTO was better than the EMG-based discrete command control and speech command control method. Finally, an outdoor long-distance control pilot experiment proved the superior performance of the developed system.
为了满足肢体功能障碍人群的移动和体育活动需求,本研究开发了一种基于人脸-计算机接口(FCI)控制的无肢体可变结构轮椅系统。FCI将面部肌电图(facial electromyography, fEMG)作为人类6个面部动作的意图识别方法,在基于Hilbert变换获取fEMG包络线的通道优化、特征优化、与通道无关的模型优化等一系列离线优化下,意图识别准确率达到97.6%。利用有限状态机(FSM)集合控制轮椅的运动和结构变化。在轮椅的远距离运动控制中,采用了既能减少使用者疲劳又能提高安全性的“接管后继续行动”(KAaTO)共享控制策略。为了测试系统的性能,在KAaTO下的制动距离测试实验中,0.429m的结果优于基于肌电图的离散命令控制和语音命令控制方法。最后,通过室外远程控制先导试验,验证了所开发系统的优越性能。
{"title":"A Novel Limbs-Free Variable Structure Wheelchair based on Face-Computer Interface (FCI) with Shared Control","authors":"Bo Zhu, Daohui Zhang, Yaqi Chu, Xingang Zhao","doi":"10.1109/icra46639.2022.9811571","DOIUrl":"https://doi.org/10.1109/icra46639.2022.9811571","url":null,"abstract":"In order to meet the mobility and physical activity needs of people with impaired limbs function, a novel limbs-free variable structure wheelchair system controled by face-computer interface (FCI) was developed in this study. FCI used facial electromyography (fEMG) as a human intention recognition method from 6 facial movements, and the accuracy of intent recognition reached 97.6% under a series of offline optimization including channel optimization based on the Hilbert transform to obtain the envelope of fEMG, features optimization, and channel-independent model optimization. A collection of finite state machines (FSM) was used to control the movement and structural changes of the wheelchair. A shared control strategy called “ Keep Action after Take Over (KAaTO) “ that can reduce user fatigue while increasing safety was used in long-distance movement control of wheelchair. To test the performance of the system, in the braking distance test experiment, the result of 0.429m under KAaTO was better than the EMG-based discrete command control and speech command control method. Finally, an outdoor long-distance control pilot experiment proved the superior performance of the developed system.","PeriodicalId":341244,"journal":{"name":"2022 International Conference on Robotics and Automation (ICRA)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114796405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An MPC Framework For Planning Safe & Trustworthy Robot Motions 安全可靠机器人运动规划的MPC框架
Pub Date : 2022-05-23 DOI: 10.1109/icra46639.2022.9812160
Moritz Eckhoff, R. J. Kirschner, Elena Kern, Saeed Abdolshah, S. Haddadin
Strategies for safe human-robot interaction (HRI), such as the well-established Safe Motion Unit, provide a velocity scaling for biomechanically safe robot motion. In addition, psychologically-based safety approaches are required for trustworthy HRI. Such schemes can be very conservative and robot motion complying with such safety approaches should be time efficient within the robot motion planning. In this study, we improve the efficiency of a previously introduced approach for psychologically-based safety in HRI via a Model Predictive Control robot motion planner that simultaneously adjusts Cartesian path and speed to minimise the distance to the target pose as fast as possible. A subordinate real-time motion generator ensures human physical safety by integrating the Safe Motion Unit. Our motion planner is validated by two experiments. The simultaneous adjustment of path and velocity accomplishes highly time efficient robot motion, while considering the human physical and psychological safety. Compared to direct path velocity scaling approaches our planner enables 28 % faster motion execution.
安全人机交互(HRI)策略,如完善的安全运动单元,为生物力学安全机器人运动提供了速度尺度。此外,可靠的HRI需要基于心理学的安全方法。这种方案可能是非常保守的,并且在机器人运动规划中,符合这种安全方法的机器人运动应该是时间有效的。在本研究中,我们通过模型预测控制机器人运动规划器提高了先前引入的基于心理的HRI安全方法的效率,该计划器同时调整笛卡尔路径和速度,以尽可能快地减少到目标姿势的距离。一个下属的实时运动发生器通过集成安全运动单元来保证人体的人身安全。通过两个实验验证了该运动规划器的有效性。在兼顾人的身心安全的前提下,通过路径和速度的同步调整,实现了高效的机器人运动。与直接路径速度缩放方法相比,我们的规划器使运动执行速度提高了28%。
{"title":"An MPC Framework For Planning Safe & Trustworthy Robot Motions","authors":"Moritz Eckhoff, R. J. Kirschner, Elena Kern, Saeed Abdolshah, S. Haddadin","doi":"10.1109/icra46639.2022.9812160","DOIUrl":"https://doi.org/10.1109/icra46639.2022.9812160","url":null,"abstract":"Strategies for safe human-robot interaction (HRI), such as the well-established Safe Motion Unit, provide a velocity scaling for biomechanically safe robot motion. In addition, psychologically-based safety approaches are required for trustworthy HRI. Such schemes can be very conservative and robot motion complying with such safety approaches should be time efficient within the robot motion planning. In this study, we improve the efficiency of a previously introduced approach for psychologically-based safety in HRI via a Model Predictive Control robot motion planner that simultaneously adjusts Cartesian path and speed to minimise the distance to the target pose as fast as possible. A subordinate real-time motion generator ensures human physical safety by integrating the Safe Motion Unit. Our motion planner is validated by two experiments. The simultaneous adjustment of path and velocity accomplishes highly time efficient robot motion, while considering the human physical and psychological safety. Compared to direct path velocity scaling approaches our planner enables 28 % faster motion execution.","PeriodicalId":341244,"journal":{"name":"2022 International Conference on Robotics and Automation (ICRA)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125579286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2022 International Conference on Robotics and Automation (ICRA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1