首页 > 最新文献

2022 Sixth IEEE International Conference on Robotic Computing (IRC)最新文献

英文 中文
Autonomous Multirotor Landing on Landing Pads and Lava Flows 在着陆垫和熔岩流上的自主多旋翼着陆
Pub Date : 2022-11-11 DOI: 10.1109/IRC55401.2022.00081
Joshua Springer
Landing is a challenging part of autonomous drone flight and a great research opportunity. This PhD proposes to improve on fiducial autonomous landing algorithms by making them more flexible. Further, it leverages its location, Iceland, to develop a method for landing on lava flows in cooperation with analog Mars exploration missions taking place in Iceland now – and potentially for future Mars landings.
着陆是自主无人机飞行的一个具有挑战性的部分,也是一个很好的研究机会。本博士提出改进基准自主着陆算法,使其更加灵活。此外,它还利用冰岛的地理位置,与目前正在冰岛进行的模拟火星探测任务合作,开发一种登陆熔岩流的方法,并有可能在未来登陆火星。
{"title":"Autonomous Multirotor Landing on Landing Pads and Lava Flows","authors":"Joshua Springer","doi":"10.1109/IRC55401.2022.00081","DOIUrl":"https://doi.org/10.1109/IRC55401.2022.00081","url":null,"abstract":"Landing is a challenging part of autonomous drone flight and a great research opportunity. This PhD proposes to improve on fiducial autonomous landing algorithms by making them more flexible. Further, it leverages its location, Iceland, to develop a method for landing on lava flows in cooperation with analog Mars exploration missions taking place in Iceland now – and potentially for future Mars landings.","PeriodicalId":282759,"journal":{"name":"2022 Sixth IEEE International Conference on Robotic Computing (IRC)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129754561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Towards advanced robotic manipulation 迈向先进的机器人操作
Pub Date : 2022-09-19 DOI: 10.1109/IRC55401.2022.00058
Francisco Roldan Sanchez, Stephen Redmond, Kevin McGuinness, Noel E. O'Connor
Robotic manipulation and control has increased in importance in recent years. However, state of the art techniques still have limitations when required to operate in real world applications. This paper explores Hindsight Experience Replay both in simulated and real environments, highlighting its weaknesses and proposing reinforcement-learning based alternatives based on reward and goal shaping. Additionally, several research questions are identified along with potential research directions that could be explored to tackle those questions.
近年来,机器人的操纵和控制变得越来越重要。然而,当需要在现实世界的应用程序中操作时,最先进的技术仍然存在局限性。本文探讨了模拟和真实环境中的后见之明经验回放,突出了其弱点,并提出了基于奖励和目标塑造的强化学习替代方案。此外,还确定了几个研究问题以及可以探索解决这些问题的潜在研究方向。
{"title":"Towards advanced robotic manipulation","authors":"Francisco Roldan Sanchez, Stephen Redmond, Kevin McGuinness, Noel E. O'Connor","doi":"10.1109/IRC55401.2022.00058","DOIUrl":"https://doi.org/10.1109/IRC55401.2022.00058","url":null,"abstract":"Robotic manipulation and control has increased in importance in recent years. However, state of the art techniques still have limitations when required to operate in real world applications. This paper explores Hindsight Experience Replay both in simulated and real environments, highlighting its weaknesses and proposing reinforcement-learning based alternatives based on reward and goal shaping. Additionally, several research questions are identified along with potential research directions that could be explored to tackle those questions.","PeriodicalId":282759,"journal":{"name":"2022 Sixth IEEE International Conference on Robotic Computing (IRC)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126571177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-Calibrating Anomaly and Change Detection for Autonomous Inspection Robots 自主检测机器人的自校准异常与变化检测
Pub Date : 2022-08-26 DOI: 10.1109/IRC55401.2022.00042
Sahar Salimpour, J. P. Queralta, Tomi Westerlund
Automatic detection of visual anomalies and changes in the environment has been a topic of recurrent attention in the fields of machine learning and computer vision over the past decades. A visual anomaly or change detection algorithm identifies regions of an image that differ from a reference image or dataset. The majority of existing approaches focus on anomaly or fault detection in a specific class of images or environments, while general-purpose visual anomaly detection algorithms are more scarce in the literature. In this paper, we propose a comprehensive deep learning framework for detecting anomalies and changes in a priori unknown environments after a reference dataset is gathered, and without need for retraining the model. We use the SuperPoint and SuperGlue feature extraction and matching methods to detect anomalies based on reference images taken from a similar location and with partial overlapping of the field of view. We also introduce a self-calibrating method for the proposed model in order to address the problem of sensitivity to feature matching thresholds and environmental conditions. To evaluate the proposed framework, we have used a ground robot system for the purpose of reference and query data collection. We show that high accuracy can be obtained using the proposed method. We also show that the calibration process enhances changes and foreign object detection performance.
在过去的几十年里,视觉异常和环境变化的自动检测一直是机器学习和计算机视觉领域反复关注的话题。视觉异常或变化检测算法识别图像中与参考图像或数据集不同的区域。现有的大多数方法都集中在特定类别的图像或环境中的异常或故障检测上,而通用的视觉异常检测算法在文献中更为稀缺。在本文中,我们提出了一个全面的深度学习框架,用于在收集参考数据集后检测先验未知环境中的异常和变化,而无需重新训练模型。我们使用SuperPoint和SuperGlue特征提取和匹配方法,基于从相似位置拍摄的参考图像,并在视场部分重叠的情况下检测异常。我们还为所提出的模型引入了一种自校准方法,以解决对特征匹配阈值和环境条件的敏感性问题。为了评估提出的框架,我们使用了一个地面机器人系统来参考和查询数据收集。结果表明,该方法具有较高的精度。我们还表明,校准过程提高了变化和异物检测性能。
{"title":"Self-Calibrating Anomaly and Change Detection for Autonomous Inspection Robots","authors":"Sahar Salimpour, J. P. Queralta, Tomi Westerlund","doi":"10.1109/IRC55401.2022.00042","DOIUrl":"https://doi.org/10.1109/IRC55401.2022.00042","url":null,"abstract":"Automatic detection of visual anomalies and changes in the environment has been a topic of recurrent attention in the fields of machine learning and computer vision over the past decades. A visual anomaly or change detection algorithm identifies regions of an image that differ from a reference image or dataset. The majority of existing approaches focus on anomaly or fault detection in a specific class of images or environments, while general-purpose visual anomaly detection algorithms are more scarce in the literature. In this paper, we propose a comprehensive deep learning framework for detecting anomalies and changes in a priori unknown environments after a reference dataset is gathered, and without need for retraining the model. We use the SuperPoint and SuperGlue feature extraction and matching methods to detect anomalies based on reference images taken from a similar location and with partial overlapping of the field of view. We also introduce a self-calibrating method for the proposed model in order to address the problem of sensitivity to feature matching thresholds and environmental conditions. To evaluate the proposed framework, we have used a ground robot system for the purpose of reference and query data collection. We show that high accuracy can be obtained using the proposed method. We also show that the calibration process enhances changes and foreign object detection performance.","PeriodicalId":282759,"journal":{"name":"2022 Sixth IEEE International Conference on Robotic Computing (IRC)","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125676379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Autonomous Drone Landing with Fiducial Markers and a Gimbal-Mounted Camera for Active Tracking 具有基准标记和用于主动跟踪的云台摄像机的自主无人机着陆
Pub Date : 2022-06-09 DOI: 10.1109/IRC55401.2022.00047
Joshua Springer, M. Kyas
Precision landing is a remaining challenge in autonomous drone flight. Fiducial markers provide a computationally cheap way for a drone to locate a landing pad and autonomously execute precision landings. However, most work in this field depends on either rigidly-mounted or downward-facing cameras which restrict the drone’s ability to detect the marker. We present a method of autonomous landing that uses a gimbal-mounted camera to quickly search for the landing pad by simply spinning in place while tilting the camera up and down, and to continually aim the camera at the landing pad during approach and landing. This method demonstrates successful search, tracking, and landing with 4 of 5 tested fiducial systems on a physical drone with no human intervention. Per fiducial system, we present the distributions of the distances from the drone to the center of the landing pad after each successful landing. We also show representative examples of flight trajectories, marker tracking performance, and control outputs for each channel during the landing. Finally, we discuss qualitative strengths and weaknesses underlying each system.
在自主无人机飞行中,精确着陆仍然是一个挑战。基准标记为无人机定位着陆平台和自动执行精确着陆提供了一种计算成本低廉的方法。然而,该领域的大多数工作都依赖于刚性安装或朝下的摄像头,这限制了无人机检测标记的能力。我们提出了一种自主着陆的方法,该方法使用安装在万向节上的相机,通过简单地在原地旋转同时上下倾斜相机来快速搜索着陆垫,并在接近和着陆期间持续将相机对准着陆垫。该方法演示了在没有人为干预的情况下,在物理无人机上使用5个经过测试的基准系统中的4个成功搜索,跟踪和着陆。根据基准系统,我们给出了每次成功着陆后无人机到着陆平台中心的距离分布。我们还展示了飞行轨迹、标记跟踪性能和着陆过程中每个通道的控制输出的代表性示例。最后,我们讨论了每个系统的定性优点和缺点。
{"title":"Autonomous Drone Landing with Fiducial Markers and a Gimbal-Mounted Camera for Active Tracking","authors":"Joshua Springer, M. Kyas","doi":"10.1109/IRC55401.2022.00047","DOIUrl":"https://doi.org/10.1109/IRC55401.2022.00047","url":null,"abstract":"Precision landing is a remaining challenge in autonomous drone flight. Fiducial markers provide a computationally cheap way for a drone to locate a landing pad and autonomously execute precision landings. However, most work in this field depends on either rigidly-mounted or downward-facing cameras which restrict the drone’s ability to detect the marker. We present a method of autonomous landing that uses a gimbal-mounted camera to quickly search for the landing pad by simply spinning in place while tilting the camera up and down, and to continually aim the camera at the landing pad during approach and landing. This method demonstrates successful search, tracking, and landing with 4 of 5 tested fiducial systems on a physical drone with no human intervention. Per fiducial system, we present the distributions of the distances from the drone to the center of the landing pad after each successful landing. We also show representative examples of flight trajectories, marker tracking performance, and control outputs for each channel during the landing. Finally, we discuss qualitative strengths and weaknesses underlying each system.","PeriodicalId":282759,"journal":{"name":"2022 Sixth IEEE International Conference on Robotic Computing (IRC)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126690945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Gaze-based Object Detection in the Wild 野外基于注视的物体检测
Pub Date : 2022-03-29 DOI: 10.1109/IRC55401.2022.00017
Daniel Weber, Wolfgang Fuhl, A. Zell, Enkelejda Kasneci
I human-robot collaboration, one challenging task is to teach a robot new yet unknown objects enabling it to interact with them. Thereby, gaze can contain valuable information. We investigate if it is possible to detect objects (object or no object) merely from gaze data and determine their bounding box parameters. For this purpose, we explore different sizes of temporal windows, which serve as a basis for the computation of heatmaps, i.e., the spatial distribution of the gaze data. Additionally, we analyze different grid sizes of these heatmaps, and demonstrate the functionality in a proof of concept using different machine learning techniques. Our method is characterized by its speed and resource efficiency compared to conventional object detectors. In order to generate the required data, we conducted a study with five subjects who could move freely and thus, turn towards arbitrary objects. This way, we chose a scenario for our data collection that is as realistic as possible. Since the subjects move while facing objects, the heatmaps also contain gaze data trajectories, complicating the detection and parameter regression. We make our data set publicly available to the research community for download.
在人机协作中,一项具有挑战性的任务是教会机器人新的未知对象,使其能够与之交互。因此,凝视可以包含有价值的信息。我们研究是否有可能仅仅从凝视数据中检测物体(物体或无物体)并确定它们的边界框参数。为此,我们探索了不同大小的时间窗口,作为计算热图的基础,即凝视数据的空间分布。此外,我们分析了这些热图的不同网格大小,并使用不同的机器学习技术在概念验证中演示了功能。与传统的目标检测器相比,我们的方法具有速度快、资源效率高的特点。为了生成所需的数据,我们对五名可以自由移动的受试者进行了研究,因此,他们可以转向任意物体。通过这种方式,我们为数据收集选择了一个尽可能真实的场景。由于受试者在面对物体时移动,热图还包含凝视数据轨迹,使检测和参数回归变得复杂。我们将我们的数据集公开提供给研究社区下载。
{"title":"Gaze-based Object Detection in the Wild","authors":"Daniel Weber, Wolfgang Fuhl, A. Zell, Enkelejda Kasneci","doi":"10.1109/IRC55401.2022.00017","DOIUrl":"https://doi.org/10.1109/IRC55401.2022.00017","url":null,"abstract":"I human-robot collaboration, one challenging task is to teach a robot new yet unknown objects enabling it to interact with them. Thereby, gaze can contain valuable information. We investigate if it is possible to detect objects (object or no object) merely from gaze data and determine their bounding box parameters. For this purpose, we explore different sizes of temporal windows, which serve as a basis for the computation of heatmaps, i.e., the spatial distribution of the gaze data. Additionally, we analyze different grid sizes of these heatmaps, and demonstrate the functionality in a proof of concept using different machine learning techniques. Our method is characterized by its speed and resource efficiency compared to conventional object detectors. In order to generate the required data, we conducted a study with five subjects who could move freely and thus, turn towards arbitrary objects. This way, we chose a scenario for our data collection that is as realistic as possible. Since the subjects move while facing objects, the heatmaps also contain gaze data trajectories, complicating the detection and parameter regression. We make our data set publicly available to the research community for download.","PeriodicalId":282759,"journal":{"name":"2022 Sixth IEEE International Conference on Robotic Computing (IRC)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128001186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Evaluation of Orientation Ambiguity and Detection Rate in April Tag and WhyCode 四月标签和WhyCode中方向模糊度及检出率的评价
Pub Date : 2022-03-18 DOI: 10.1109/IRC55401.2022.00054
Joshua Springer, M. Kyas
Fiducial systems provide a computationally cheap way for mobile robots to estimate the pose of objects, or their own pose, using just a monocular camera. However, the orientation component of the pose of fiducial markers is unreliable, which can have destructive effects in autonomous drone landing on landing pads marked with fiducial markers. This paper evaluates the April Tag and WhyCode fiducial systems in terms of orientation ambiguity and detection rate on embedded hardware. We test 2 April Tag variants – 1 default and 1 custom – and 3 Whycode variants – 1 default and 2 custom. We determine that they are suitable for autonomous drone landing applications in terms of detection rate, but may generate erroneous control signals as a result of orientation ambiguity in the pose estimates.
基准系统为移动机器人提供了一种计算成本低廉的方法来估计物体的姿势,或者它们自己的姿势,只需要一个单目摄像机。然而,基准标记姿态的方向分量是不可靠的,这可能会对自主无人机在有基准标记的着陆平台上着陆产生破坏性影响。本文对四月标签和WhyCode基准系统在嵌入式硬件上的方向模糊度和检测率进行了评估。我们测试了2个四月标签变体(1个默认和1个自定义)和3个Whycode变体(1个默认和2个自定义)。我们确定它们在检测率方面适用于自主无人机着陆应用,但可能由于姿态估计中的方向模糊而产生错误的控制信号。
{"title":"Evaluation of Orientation Ambiguity and Detection Rate in April Tag and WhyCode","authors":"Joshua Springer, M. Kyas","doi":"10.1109/IRC55401.2022.00054","DOIUrl":"https://doi.org/10.1109/IRC55401.2022.00054","url":null,"abstract":"Fiducial systems provide a computationally cheap way for mobile robots to estimate the pose of objects, or their own pose, using just a monocular camera. However, the orientation component of the pose of fiducial markers is unreliable, which can have destructive effects in autonomous drone landing on landing pads marked with fiducial markers. This paper evaluates the April Tag and WhyCode fiducial systems in terms of orientation ambiguity and detection rate on embedded hardware. We test 2 April Tag variants – 1 default and 1 custom – and 3 Whycode variants – 1 default and 2 custom. We determine that they are suitable for autonomous drone landing applications in terms of detection rate, but may generate erroneous control signals as a result of orientation ambiguity in the pose estimates.","PeriodicalId":282759,"journal":{"name":"2022 Sixth IEEE International Conference on Robotic Computing (IRC)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125234512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
GA+DDPG+HER: Genetic Algorithm-Based Function Optimizer in Deep Reinforcement Learning for Robotic Manipulation Tasks GA+DDPG+HER:基于遗传算法的机器人操作任务深度强化学习函数优化器
Pub Date : 2022-02-28 DOI: 10.1109/IRC55401.2022.00022
Adarsh Sehgal, Nicholas Ward, Hung M. La, C. Papachristos, S. Louis
Agents can base decisions made using reinforcement learning (RL) on a reward function. The selection of values for the learning algorithm parameters can, nevertheless, have a substantial impact on the overall learning process. In order to discover values for the learning parameters that are close to optimal, we extended our previously proposed genetic algorithm-based Deep Deterministic Policy Gradient and Hindsight Experience Replay approach (referred to as GA+DDPG+HER) in this study. On the robotic manipulation tasks of FetchReach, FetchSlide, FetchPush, FetchPick&Place, and DoorOpening, we applied the GA+DDPG+HER methodology. Our technique GA+DDPG+HER was also used in the AuboReach environment with a few adjustments. Our experimental analysis demonstrates that our method produces performance that is noticeably better and occurs faster than the original algorithm. We also offer proof that GA+DDPG+HER beat the current approaches. The final results support our assertion and offer sufficient proof that automating the parameter tuning procedure is crucial and does cut down learning time by as much as 57%.
智能体可以在奖励函数的基础上使用强化学习(RL)做出决策。然而,学习算法参数值的选择会对整个学习过程产生重大影响。为了发现接近最优的学习参数值,我们在本研究中扩展了先前提出的基于遗传算法的深度确定性策略梯度和后见之明经验重播方法(称为GA+DDPG+HER)。在机器人操作任务FetchReach、FetchSlide、FetchPush、FetchPick&Place和DoorOpening中,我们应用了GA+DDPG+HER方法。我们的技术GA+DDPG+HER也在AuboReach环境中进行了一些调整。我们的实验分析表明,我们的方法产生的性能明显优于原始算法,并且运行速度更快。我们还提供了GA+DDPG+HER优于当前方法的证据。最终的结果支持了我们的断言,并提供了充分的证据,证明自动化参数调整过程是至关重要的,并且确实减少了多达57%的学习时间。
{"title":"GA+DDPG+HER: Genetic Algorithm-Based Function Optimizer in Deep Reinforcement Learning for Robotic Manipulation Tasks","authors":"Adarsh Sehgal, Nicholas Ward, Hung M. La, C. Papachristos, S. Louis","doi":"10.1109/IRC55401.2022.00022","DOIUrl":"https://doi.org/10.1109/IRC55401.2022.00022","url":null,"abstract":"Agents can base decisions made using reinforcement learning (RL) on a reward function. The selection of values for the learning algorithm parameters can, nevertheless, have a substantial impact on the overall learning process. In order to discover values for the learning parameters that are close to optimal, we extended our previously proposed genetic algorithm-based Deep Deterministic Policy Gradient and Hindsight Experience Replay approach (referred to as GA+DDPG+HER) in this study. On the robotic manipulation tasks of FetchReach, FetchSlide, FetchPush, FetchPick&Place, and DoorOpening, we applied the GA+DDPG+HER methodology. Our technique GA+DDPG+HER was also used in the AuboReach environment with a few adjustments. Our experimental analysis demonstrates that our method produces performance that is noticeably better and occurs faster than the original algorithm. We also offer proof that GA+DDPG+HER beat the current approaches. The final results support our assertion and offer sufficient proof that automating the parameter tuning procedure is crucial and does cut down learning time by as much as 57%.","PeriodicalId":282759,"journal":{"name":"2022 Sixth IEEE International Conference on Robotic Computing (IRC)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115536785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi-View Contrastive Learning from Demonstrations 从示范中进行多视角对比学习
Pub Date : 2022-01-30 DOI: 10.1109/IRC55401.2022.00067
André Rosa de Sousa Porfírio Correia, L. Alexandre
This paper presents a framework for learning visual representations from unlabeled video demonstrations captured from multiple viewpoints. We show that these representations are applicable for imitating robotic tasks. We use contrastive learning to enhance task-relevant information while suppressing irrelevant information in the feature embeddings. We validate the proposed method on the publicly available Multi-View Pouring and a custom Pick and Place data sets and compare it with the TCN and CMC baselines. We evaluate the learned representations using three metrics: viewpoint alignment, stage classification and reinforcement learning. In all cases, the results improve when compared to state-of-the-art approaches.
本文提出了一个从多个视点捕获的未标记视频演示中学习视觉表示的框架。我们证明这些表征适用于模仿机器人任务。我们使用对比学习来增强任务相关信息,同时抑制特征嵌入中的不相关信息。我们在公开可用的多视图浇注和自定义取放数据集上验证了所提出的方法,并将其与TCN和CMC基线进行了比较。我们使用三个度量来评估学习到的表示:视点对齐、阶段分类和强化学习。在所有情况下,与最先进的方法相比,结果都有所改善。
{"title":"Multi-View Contrastive Learning from Demonstrations","authors":"André Rosa de Sousa Porfírio Correia, L. Alexandre","doi":"10.1109/IRC55401.2022.00067","DOIUrl":"https://doi.org/10.1109/IRC55401.2022.00067","url":null,"abstract":"This paper presents a framework for learning visual representations from unlabeled video demonstrations captured from multiple viewpoints. We show that these representations are applicable for imitating robotic tasks. We use contrastive learning to enhance task-relevant information while suppressing irrelevant information in the feature embeddings. We validate the proposed method on the publicly available Multi-View Pouring and a custom Pick and Place data sets and compare it with the TCN and CMC baselines. We evaluate the learned representations using three metrics: viewpoint alignment, stage classification and reinforcement learning. In all cases, the results improve when compared to state-of-the-art approaches.","PeriodicalId":282759,"journal":{"name":"2022 Sixth IEEE International Conference on Robotic Computing (IRC)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133285026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2022 Sixth IEEE International Conference on Robotic Computing (IRC)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1