首页 > 最新文献

Journal of Intelligent & Robotic Systems最新文献

英文 中文
Manipulation of a Complex Object Using Dual-Arm Robot with Mask R-CNN and Grasping Strategy 使用带掩模 R-CNN 和抓取策略的双臂机器人操纵复杂物体
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-10 DOI: 10.1007/s10846-024-02132-0
Dumrongsak Kijdech, Supachai Vongbunyong

Hot forging is one of the common manufacturing processes for producing brass workpieces. However forging produces flash which is a thin metal part around the desired part formed with an excessive material. Using robots with vision system to manipulate this workpiece has encountered several challenging issues, e.g. the uncertain shape of flash, color, reflection of brass surface, different lighting condition, and the uncertainty surrounding the position and orientation of the workpiece. In this research, Mask region-based convolutional neural network together with image processing is used to resolve these issues. The depth camera can provide images for visual detection. Machine learning Mask region-based convolutional neural network model was trained with color images and the position of the object is determined by the depth image. A dual arm 7 degree of freedom collaborative robot with proposed grasping strategy is used to grasp the workpiece that can be in inappropriate position and pose. Eventually, experiments were conducted to assess the visual detection process and the grasp planning of the robot.

热锻是生产黄铜工件的常见制造工艺之一。然而,锻造过程中会产生闪光,即在所需工件周围形成的薄金属部分,材料过多。使用带有视觉系统的机器人来操纵这种工件会遇到一些具有挑战性的问题,例如闪光的不确定形状、颜色、黄铜表面的反射、不同的照明条件以及工件位置和方向的不确定性。在这项研究中,基于掩膜区域的卷积神经网络与图像处理一起被用来解决这些问题。深度摄像头可为视觉检测提供图像。使用彩色图像训练基于掩膜区域的机器学习卷积神经网络模型,并通过深度图像确定物体的位置。采用建议的抓取策略的双臂 7 自由度协作机器人可抓取位置和姿势不合适的工件。最终,实验对视觉检测过程和机器人的抓取规划进行了评估。
{"title":"Manipulation of a Complex Object Using Dual-Arm Robot with Mask R-CNN and Grasping Strategy","authors":"Dumrongsak Kijdech, Supachai Vongbunyong","doi":"10.1007/s10846-024-02132-0","DOIUrl":"https://doi.org/10.1007/s10846-024-02132-0","url":null,"abstract":"<p>Hot forging is one of the common manufacturing processes for producing brass workpieces. However forging produces flash which is a thin metal part around the desired part formed with an excessive material. Using robots with vision system to manipulate this workpiece has encountered several challenging issues, e.g. the uncertain shape of flash, color, reflection of brass surface, different lighting condition, and the uncertainty surrounding the position and orientation of the workpiece. In this research, Mask region-based convolutional neural network together with image processing is used to resolve these issues. The depth camera can provide images for visual detection. Machine learning Mask region-based convolutional neural network model was trained with color images and the position of the object is determined by the depth image. A dual arm 7 degree of freedom collaborative robot with proposed grasping strategy is used to grasp the workpiece that can be in inappropriate position and pose. Eventually, experiments were conducted to assess the visual detection process and the grasp planning of the robot.</p>","PeriodicalId":54794,"journal":{"name":"Journal of Intelligent & Robotic Systems","volume":"62 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141570668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Teledrive: An Embodied AI Based Telepresence System Teledrive:基于嵌入式人工智能的网真系统
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-09 DOI: 10.1007/s10846-024-02124-0
Snehasis Banerjee, Sayan Paul, Ruddradev Roychoudhury, Abhijan Bhattacharya, Chayan Sarkar, Ashis Sau, Pradip Pramanick, Brojeshwar Bhowmick

This article presents ‘Teledrive’, a telepresence robotic system with embodied AI features that empowers an operator to navigate the telerobot in any unknown remote place with minimal human intervention. We conceive Teledrive in the context of democratizing remote ‘care-giving’ for elderly citizens as well as for isolated patients, affected by contagious diseases. In particular, this paper focuses on the problem of navigating to a rough target area (like ‘bedroom’ or ‘kitchen’) rather than pre-specified point destinations. This ushers in a unique ‘AreaGoal’ based navigation feature, which has not been explored in depth in the contemporary solutions. Further, we describe an edge computing-based software system built on a WebRTC-based communication framework to realize the aforementioned scheme through an easy-to-use speech-based human-robot interaction. Moreover, to enhance the ease of operation for the remote caregiver, we incorporate a ‘person following’ feature, whereby a robot follows a person on the move in its premises as directed by the operator. Moreover, the system presented is loosely coupled with specific robot hardware, unlike the existing solutions. We have evaluated the efficacy of the proposed system through baseline experiments, user study, and real-life deployment.

本文介绍的 "Teledrive "是一种具有人工智能功能的远程呈现机器人系统,它能让操作员在极少人为干预的情况下,在任何未知的偏远地区为远程机器人导航。我们设想 Teledrive 的背景是,为老年公民和受传染病影响的孤立病人提供远程 "护理 "的民主化。本文特别关注的问题是导航到一个大致的目标区域(如 "卧室 "或 "厨房"),而不是预先指定的点目的地。这带来了一种独特的基于 "区域目标 "的导航功能,而当代的解决方案尚未对这一功能进行深入探讨。此外,我们还介绍了一种基于边缘计算的软件系统,该系统建立在基于 WebRTC 的通信框架上,通过简单易用的语音人机交互实现上述方案。此外,为了提高远程护理人员的操作便利性,我们还加入了 "人员跟随 "功能,即机器人在操作员的指示下跟随人员在其场所内移动。此外,与现有的解决方案不同,本系统与特定的机器人硬件是松散耦合的。我们通过基线实验、用户研究和实际部署评估了所提系统的功效。
{"title":"Teledrive: An Embodied AI Based Telepresence System","authors":"Snehasis Banerjee, Sayan Paul, Ruddradev Roychoudhury, Abhijan Bhattacharya, Chayan Sarkar, Ashis Sau, Pradip Pramanick, Brojeshwar Bhowmick","doi":"10.1007/s10846-024-02124-0","DOIUrl":"https://doi.org/10.1007/s10846-024-02124-0","url":null,"abstract":"<p>This article presents ‘Teledrive’, a telepresence robotic system with embodied AI features that empowers an operator to navigate the telerobot in any unknown remote place with minimal human intervention. We conceive Teledrive in the context of democratizing remote ‘care-giving’ for elderly citizens as well as for isolated patients, affected by contagious diseases. In particular, this paper focuses on the problem of navigating to a rough target area (like ‘bedroom’ or ‘kitchen’) rather than pre-specified point destinations. This ushers in a unique ‘AreaGoal’ based navigation feature, which has not been explored in depth in the contemporary solutions. Further, we describe an edge computing-based software system built on a WebRTC-based communication framework to realize the aforementioned scheme through an easy-to-use speech-based human-robot interaction. Moreover, to enhance the ease of operation for the remote caregiver, we incorporate a ‘person following’ feature, whereby a robot follows a person on the move in its premises as directed by the operator. Moreover, the system presented is loosely coupled with specific robot hardware, unlike the existing solutions. We have evaluated the efficacy of the proposed system through baseline experiments, user study, and real-life deployment.</p>","PeriodicalId":54794,"journal":{"name":"Journal of Intelligent & Robotic Systems","volume":"37 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141570669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Model-Based Reinforcement Learning for Predictive Control of Robotic Systems with Dense and Sparse Rewards 基于深度模型的强化学习用于具有密集和稀疏奖励的机器人系统的预测控制
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-09 DOI: 10.1007/s10846-024-02118-y
Luka Antonyshyn, Sidney Givigi

Sparse rewards and sample efficiency are open areas of research in the field of reinforcement learning. These problems are especially important when considering applications of reinforcement learning to robotics and other cyber-physical systems. This is so because in these domains many tasks are goal-based and naturally expressed with binary successes and failures, action spaces are large and continuous, and real interactions with the environment are limited. In this work, we propose Deep Value-and-Predictive-Model Control (DVPMC), a model-based predictive reinforcement learning algorithm for continuous control that uses system identification, value function approximation and sampling-based optimization to select actions. The algorithm is evaluated on a dense reward and a sparse reward task. We show that it can match the performance of a predictive control approach to the dense reward problem, and outperforms model-free and model-based learning algorithms on the sparse reward task on the metrics of sample efficiency and performance. We verify the performance of an agent trained in simulation using DVPMC on a real robot playing the reach-avoid game. Video of the experiment can be found here: https://youtu.be/0Q274kcfn4c.

稀疏奖励和样本效率是强化学习领域的开放研究领域。在考虑将强化学习应用于机器人和其他网络物理系统时,这些问题尤为重要。这是因为在这些领域中,许多任务都是基于目标的,并自然地以二进制的成功和失败来表示,行动空间大且连续,而与环境的实际交互是有限的。在这项工作中,我们提出了深度值与预测模型控制(DVPMC),这是一种基于模型的预测强化学习算法,用于连续控制,它使用系统识别、值函数近似和基于采样的优化来选择行动。该算法在密集奖励和稀疏奖励任务中进行了评估。结果表明,在密集奖励问题上,该算法的性能可以与预测控制方法相媲美;在稀疏奖励任务上,该算法在采样效率和性能指标上优于无模型和基于模型的学习算法。我们在一个玩躲避游戏的真实机器人身上验证了使用 DVPMC 模拟训练的代理的性能。实验视频请点击:https://youtu.be/0Q274kcfn4c。
{"title":"Deep Model-Based Reinforcement Learning for Predictive Control of Robotic Systems with Dense and Sparse Rewards","authors":"Luka Antonyshyn, Sidney Givigi","doi":"10.1007/s10846-024-02118-y","DOIUrl":"https://doi.org/10.1007/s10846-024-02118-y","url":null,"abstract":"<p>Sparse rewards and sample efficiency are open areas of research in the field of reinforcement learning. These problems are especially important when considering applications of reinforcement learning to robotics and other cyber-physical systems. This is so because in these domains many tasks are goal-based and naturally expressed with binary successes and failures, action spaces are large and continuous, and real interactions with the environment are limited. In this work, we propose Deep Value-and-Predictive-Model Control (DVPMC), a model-based predictive reinforcement learning algorithm for continuous control that uses system identification, value function approximation and sampling-based optimization to select actions. The algorithm is evaluated on a dense reward and a sparse reward task. We show that it can match the performance of a predictive control approach to the dense reward problem, and outperforms model-free and model-based learning algorithms on the sparse reward task on the metrics of sample efficiency and performance. We verify the performance of an agent trained in simulation using DVPMC on a real robot playing the reach-avoid game. Video of the experiment can be found here: https://youtu.be/0Q274kcfn4c.</p>","PeriodicalId":54794,"journal":{"name":"Journal of Intelligent & Robotic Systems","volume":"30 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141570673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Encoder Spatio-Temporal Feature Fusion Network for Electric Vehicle Charging Load Prediction 用于电动汽车充电负荷预测的多编码器时空特征融合网络
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-09 DOI: 10.1007/s10846-024-02125-z
Yufan Chen, Mengqin Wang, Yanling Wei, Xueliang Huang, Shan Gao

Electric vehicles (EVs) have been initiated as a preference for decarbonizing road transport. Accurate charging load prediction is essential for the construction of EV charging facilities systematically and for the coordination of EV energy demand with the requisite peak power supply. It is noted that the charging load of EVs exhibits high complexity and randomness due to temporal and spatial uncertainties. Therefore, this paper proposes a SEDformer-based charging road prediction method to capture the spatio-temporal characteristics of charging load data. As a deep learning model, SEDformer comprises multiple encoders and a single decoder. In particular, the proposed model includes a Temporal Encoder Block based on the self-attention mechanism and a Spatial Encoder Block based on the channel attention mechanism with sequence decomposition, followed by an aggregated decoder for information fusion. It is shown that the proposed method outperforms various baseline models on a real-world dataset from Palo Alto, U.S., demonstrating its superiority in addressing spatio-temporal data-driven load forecasting problems.

电动汽车(EV)已成为道路交通去碳化的首选。准确的充电负荷预测对于系统地建设电动汽车充电设施以及协调电动汽车能源需求和必要的高峰电力供应至关重要。人们注意到,由于时间和空间的不确定性,电动汽车的充电负荷表现出高度的复杂性和随机性。因此,本文提出了一种基于 SEDformer 的充电道路预测方法,以捕捉充电负荷数据的时空特征。作为一种深度学习模型,SEDformer 由多个编码器和一个解码器组成。具体而言,所提出的模型包括一个基于自我注意机制的时间编码器块和一个基于序列分解的信道注意机制的空间编码器块,然后是一个用于信息融合的聚合解码器。研究表明,在美国帕洛阿尔托的实际数据集上,所提出的方法优于各种基线模型,证明了它在解决时空数据驱动的负荷预测问题方面的优越性。
{"title":"Multi-Encoder Spatio-Temporal Feature Fusion Network for Electric Vehicle Charging Load Prediction","authors":"Yufan Chen, Mengqin Wang, Yanling Wei, Xueliang Huang, Shan Gao","doi":"10.1007/s10846-024-02125-z","DOIUrl":"https://doi.org/10.1007/s10846-024-02125-z","url":null,"abstract":"<p>Electric vehicles (EVs) have been initiated as a preference for decarbonizing road transport. Accurate charging load prediction is essential for the construction of EV charging facilities systematically and for the coordination of EV energy demand with the requisite peak power supply. It is noted that the charging load of EVs exhibits high complexity and randomness due to temporal and spatial uncertainties. Therefore, this paper proposes a SEDformer-based charging road prediction method to capture the spatio-temporal characteristics of charging load data. As a deep learning model, SEDformer comprises multiple encoders and a single decoder. In particular, the proposed model includes a Temporal Encoder Block based on the self-attention mechanism and a Spatial Encoder Block based on the channel attention mechanism with sequence decomposition, followed by an aggregated decoder for information fusion. It is shown that the proposed method outperforms various baseline models on a real-world dataset from Palo Alto, U.S., demonstrating its superiority in addressing spatio-temporal data-driven load forecasting problems.</p>","PeriodicalId":54794,"journal":{"name":"Journal of Intelligent & Robotic Systems","volume":"13 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141570775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GreatBlue: a 55-Pound Vertical-Takeoff-and-Landing Fixed-Wing sUAS for Science; Systems, Communication, Simulation, Data Processing, Payloads, Package Delivery, and Mission Flight Performance GreatBlue:55 磅垂直起降固定翼科学无人机系统;系统、通信、模拟、数据处理、有效载荷、包裹交付和任务飞行性能
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-09 DOI: 10.1007/s10846-024-02052-z
Calvin Coopmans, Stockton Slack, Nathan Schwemmer, Chase Vance, A. J. Beckwith, Daniel J. Robinson

As small, uncrewed systems (sUAS) grow in popularity and in number, larger and larger drone aircraft will become more common–up to the FAA limit of 55 pound gross take-off weight (GTOW) and beyond. Due to their larger payload capabilities, longer flight time, and better safety systems, autonomous systems that maximize CFR 14 Part 107 flight drone operations regulations will become more common, especially for operations such as imagery or other data collection which scale well with longer flight times and larger flight areas. In this new paper, a unique all-electric 55-pound VTOL transition fixed-wing sUAS specifically engineered for scientific data collection named “GreatBlue” is presented, along with systems, communications, scientific payload, data collection and processing, package delivery payload, ground control station, and mission simulation system. Able to fly for up to 2.5 hours while collecting multispectral remotely-sensed imagery, the unique GreatBlue system is shown, along with a package delivery flight example, flight data from two scientific data collection flights over California almond fields and a Utah Reservoir are shown including flight plan vs. as-flown.

随着小型无人驾驶系统 (sUAS) 的普及和数量的增加,越来越大的无人驾驶飞机将变得越来越常见,甚至达到美国联邦航空局规定的 55 磅起飞总重 (GTOW) 或更高的限制。由于具有更大的有效载荷能力、更长的飞行时间和更好的安全系统,能够最大限度地满足 CFR 14 107 部无人机飞行操作规定的自主系统将变得越来越普遍,特别是在图像或其他数据收集等操作方面,这些操作可以通过更长的飞行时间和更大的飞行区域进行扩展。在这篇新论文中,介绍了一种独特的全电动 55 磅 VTOL 过渡固定翼无人机系统,该系统专门设计用于科学数据收集,命名为 "GreatBlue",同时还介绍了系统、通信、科学有效载荷、数据收集和处理、包裹递送有效载荷、地面控制站和任务模拟系统。独特的 GreatBlue 系统能够在收集多光谱遥感图像的同时飞行长达 2.5 小时,展示了一个包裹投递飞行示例,还展示了在加利福尼亚杏仁田和犹他州水库上空进行的两次科学数据收集飞行的飞行数据,包括飞行计划与实际飞行的对比。
{"title":"GreatBlue: a 55-Pound Vertical-Takeoff-and-Landing Fixed-Wing sUAS for Science; Systems, Communication, Simulation, Data Processing, Payloads, Package Delivery, and Mission Flight Performance","authors":"Calvin Coopmans, Stockton Slack, Nathan Schwemmer, Chase Vance, A. J. Beckwith, Daniel J. Robinson","doi":"10.1007/s10846-024-02052-z","DOIUrl":"https://doi.org/10.1007/s10846-024-02052-z","url":null,"abstract":"<p>As small, uncrewed systems (sUAS) grow in popularity and in number, larger and larger drone aircraft will become more common–up to the FAA limit of 55 pound gross take-off weight (GTOW) and beyond. Due to their larger payload capabilities, longer flight time, and better safety systems, autonomous systems that maximize CFR 14 Part 107 flight drone operations regulations will become more common, especially for operations such as imagery or other data collection which scale well with longer flight times and larger flight areas. In this new paper, a unique all-electric 55-pound VTOL transition fixed-wing sUAS specifically engineered for scientific data collection named “GreatBlue” is presented, along with systems, communications, scientific payload, data collection and processing, package delivery payload, ground control station, and mission simulation system. Able to fly for up to 2.5 hours while collecting multispectral remotely-sensed imagery, the unique GreatBlue system is shown, along with a package delivery flight example, flight data from two scientific data collection flights over California almond fields and a Utah Reservoir are shown including flight plan vs. as-flown.</p>","PeriodicalId":54794,"journal":{"name":"Journal of Intelligent & Robotic Systems","volume":"37 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141570674","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integration of Deep Q-Learning with a Grasp Quality Network for Robot Grasping in Cluttered Environments 将深度 Q-Learning 与抓取质量网络相结合,实现机器人在杂乱环境中的抓取操作
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-09 DOI: 10.1007/s10846-024-02127-x
Chih-Yung Huang, Yu-Hsiang Shao

During the movement of a robotic arm, collisions can easily occur if the arm directly grasps at multiple tightly stacked objects, thereby leading to grasp failures or machine damage. Grasp success can be improved through the rearrangement or movement of objects to clear space for grasping. This paper presents a high-performance deep Q-learning framework that can help robotic arms to learn synchronized push and grasp tasks. In this framework, a grasp quality network is used for precisely identifying stable grasp positions on objects to expedite model convergence and solve the problem of sparse rewards caused during training because of grasp failures. Furthermore, a novel reward function is proposed for effectively evaluating whether a pushing action is effective. The proposed framework achieved grasp success rates of 92% and 89% in simulations and real-world experiments, respectively. Furthermore, only 200 training steps were required to achieve a grasp success rate of 80%, which indicates the suitability of the proposed framework for rapid deployment in industrial settings.

在机械臂的运动过程中,如果机械臂直接抓取多个紧密堆叠的物体,很容易发生碰撞,从而导致抓取失败或机器损坏。通过重新排列或移动物体,为抓取腾出空间,可以提高抓取成功率。本文提出了一种高性能深度 Q-learning 框架,可帮助机械臂学习同步推动和抓取任务。在该框架中,抓取质量网络用于精确识别物体上的稳定抓取位置,以加快模型收敛,并解决训练过程中因抓取失败而导致的奖励稀疏问题。此外,还提出了一种新颖的奖励函数,用于有效评估推动动作是否有效。所提出的框架在模拟和实际实验中的抓取成功率分别达到了 92% 和 89%。此外,只需要 200 个训练步骤就能达到 80% 的抓取成功率,这表明所提出的框架适合在工业环境中快速部署。
{"title":"Integration of Deep Q-Learning with a Grasp Quality Network for Robot Grasping in Cluttered Environments","authors":"Chih-Yung Huang, Yu-Hsiang Shao","doi":"10.1007/s10846-024-02127-x","DOIUrl":"https://doi.org/10.1007/s10846-024-02127-x","url":null,"abstract":"<p>During the movement of a robotic arm, collisions can easily occur if the arm directly grasps at multiple tightly stacked objects, thereby leading to grasp failures or machine damage. Grasp success can be improved through the rearrangement or movement of objects to clear space for grasping. This paper presents a high-performance deep Q-learning framework that can help robotic arms to learn synchronized push and grasp tasks. In this framework, a grasp quality network is used for precisely identifying stable grasp positions on objects to expedite model convergence and solve the problem of sparse rewards caused during training because of grasp failures. Furthermore, a novel reward function is proposed for effectively evaluating whether a pushing action is effective. The proposed framework achieved grasp success rates of 92% and 89% in simulations and real-world experiments, respectively. Furthermore, only 200 training steps were required to achieve a grasp success rate of 80%, which indicates the suitability of the proposed framework for rapid deployment in industrial settings.</p>","PeriodicalId":54794,"journal":{"name":"Journal of Intelligent & Robotic Systems","volume":"29 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141570670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Privacy’s Sky-High Battle: The Use of Unmanned Aircraft Systems for Law Enforcement in the European Union 隐私权的天空之战:欧盟在执法中使用无人驾驶航空器系统的情况
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-09 DOI: 10.1007/s10846-024-02071-w
E. Öykü Kurtpınar

Benefiting from the rapid advancements in Unmanned Aircraft Systems (UAS) technology with enhanced tracking and data collection capabilities, law enforcement authorities re-discovered air as a dimension where state power can be exercised in a more affordable, accessible, and compact way. On the other hand, during law enforcement operations, UAS can collect various types of data that can be personal or sensitive, threatening the right to privacy and data protection of the data subjects. Risks include challenges related to data security, bulk data collection, the diminished transparency and fairness resulting from the inconspicuous nature of UAS, as well as ethical concerns intertwined with privacy and data protection. Upon examination of the legal framework including the General Data Protection Regulation the Law Enforcement Directive, various Aviation rules, and the new proposal for the Artificial Intelligence Act, it becomes apparent that the EU legal framework’s adequacy in safeguarding privacy and data protection against law enforcement use of UAS is context-dependent, varying across use cases. The current framework lacks clarity, leading to arbitrary application and limited protection for data subjects. Enforcement of safeguards is insufficient, and the Aviation Regulations, applicable to law enforcement UAS, require member states' opt-in, which has not occurred as of the authors’ knowledge. The Artificial Intelligence Act addresses UAS operations but focuses on market risks rather than obligations imposed on law enforcement authorities. Consequently, the existing framework is rendered inadequate for medium to high-risk law enforcement operations, leaving individuals vulnerable and insufficiently protected against intrusive UAS surveillance. Rectifying this involves addressing the enforcement gap and making the necessary amendments to relevant regulatory aspects. Additionally, the implementation of specific technical measures and steps to foster effective cooperation among stakeholders in UAS deployment for law enforcement is imperative.

无人驾驶航空器系统(UAS)技术的飞速发展增强了跟踪和数据收集能力,受益于此,执法当局重新发现空中是一个可以以更经济、更便捷、更紧凑的方式行使国家权力的维度。另一方面,在执法行动中,无人机系统可以收集各种类型的数据,这些数据可能是个人数据或敏感数据,威胁到数据主体的隐私权和数据保护。风险包括与数据安全相关的挑战、大量数据收集、无人机系统的不显眼性导致的透明度和公平性降低,以及与隐私和数据保护相关的道德问题。在对法律框架(包括《通用数据保护条例》、《执法指令》、各种航空规则和《人工智能法》的新提案)进行研究后,可以明显看出,欧盟法律框架在针对无人机系统的执法用途保障隐私和数据保护方面的充分性取决于具体情况,在不同的使用情况下也各不相同。目前的框架缺乏明确性,导致任意适用,对数据主体的保护有限。保障措施的执行力度不够,适用于执法无人机系统的《航空条例》要求成员国选择加入,但据作者所知,成员国尚未选择加入。人工智能法》涉及无人机系统的操作,但其重点是市场风险,而不是执法当局的义务。因此,现有框架不足以应对中高风险的执法行动,使个人容易受到伤害,也无法充分保护其免受无人机系统的侵入性监视。要纠正这一问题,就必须解决执法漏洞,并对相关监管方面进行必要的修订。此外,还必须实施具体的技术措施和步骤,以促进利益攸关方在部署无人机系统进行执法时开展有效合作。
{"title":"Privacy’s Sky-High Battle: The Use of Unmanned Aircraft Systems for Law Enforcement in the European Union","authors":"E. Öykü Kurtpınar","doi":"10.1007/s10846-024-02071-w","DOIUrl":"https://doi.org/10.1007/s10846-024-02071-w","url":null,"abstract":"<p>Benefiting from the rapid advancements in Unmanned Aircraft Systems (UAS) technology with enhanced tracking and data collection capabilities, law enforcement authorities re-discovered air as a dimension where state power can be exercised in a more affordable, accessible, and compact way. On the other hand, during law enforcement operations, UAS can collect various types of data that can be personal or sensitive, threatening the right to privacy and data protection of the data subjects. Risks include challenges related to data security, bulk data collection, the diminished transparency and fairness resulting from the inconspicuous nature of UAS, as well as ethical concerns intertwined with privacy and data protection. Upon examination of the legal framework including the General Data Protection Regulation the Law Enforcement Directive, various Aviation rules, and the new proposal for the Artificial Intelligence Act, it becomes apparent that the EU legal framework’s adequacy in safeguarding privacy and data protection against law enforcement use of UAS is context-dependent, varying across use cases. The current framework lacks clarity, leading to arbitrary application and limited protection for data subjects. Enforcement of safeguards is insufficient, and the Aviation Regulations, applicable to law enforcement UAS, require member states' opt-in, which has not occurred as of the authors’ knowledge. The Artificial Intelligence Act addresses UAS operations but focuses on market risks rather than obligations imposed on law enforcement authorities. Consequently, the existing framework is rendered inadequate for medium to high-risk law enforcement operations, leaving individuals vulnerable and insufficiently protected against intrusive UAS surveillance. Rectifying this involves addressing the enforcement gap and making the necessary amendments to relevant regulatory aspects. Additionally, the implementation of specific technical measures and steps to foster effective cooperation among stakeholders in UAS deployment for law enforcement is imperative.</p>","PeriodicalId":54794,"journal":{"name":"Journal of Intelligent & Robotic Systems","volume":"27 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141570671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Depth-Enhanced Deep Learning Approach For Monocular Camera Based 3D Object Detection 基于深度增强深度学习的单目摄像头三维物体检测方法
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-09 DOI: 10.1007/s10846-024-02128-w
Chuyao Wang, Nabil Aouf

Automatic 3D object detection using monocular cameras presents significant challenges in the context of autonomous driving. Precise labeling of 3D object scales requires accurate spatial information, which is difficult to obtain from a single image due to the inherent lack of depth information in monocular images, compared to LiDAR data. In this paper, we propose a novel approach to address this issue by enhancing deep neural networks with depth information for monocular 3D object detection. The proposed method comprises three key components: 1)Feature Enhancement Pyramid Module: We extend the conventional Feature Pyramid Networks (FPN) by introducing a feature enhancement pyramid network. This module fuses feature maps from the original pyramid and captures contextual correlations across multiple scales. To increase the connectivity between low-level and high-level features, additional pathways are incorporated. 2)Auxiliary Dense Depth Estimator: We introduce an auxiliary dense depth estimator that generates dense depth maps to enhance the spatial perception capabilities of the deep network model without adding computational burden. 3)Augmented Center Depth Regression: To aid center depth estimation, we employ additional bounding box vertex depth regression based on geometry. Our experimental results demonstrate the superiority of the proposed technique over existing competitive methods reported in the literature. The approach showcases remarkable performance improvements in monocular 3D object detection, making it a promising solution for autonomous driving applications.

使用单目摄像头自动检测三维物体给自动驾驶带来了巨大挑战。精确标注三维物体的尺度需要准确的空间信息,而与激光雷达数据相比,单目图像本身缺乏深度信息,因此很难从单幅图像中获取空间信息。在本文中,我们提出了一种解决这一问题的新方法,即利用深度信息增强深度神经网络,用于单目三维物体检测。所提出的方法由三个关键部分组成:1)特征增强金字塔模块:我们通过引入特征增强金字塔网络来扩展传统的特征金字塔网络(FPN)。该模块融合了原始金字塔中的特征图,并捕捉跨尺度的上下文相关性。为了增加低层次特征与高层次特征之间的联系,还加入了额外的路径。2)辅助密集深度估计器:我们引入了一个辅助密集深度估计器,它能生成密集深度图,在不增加计算负担的情况下增强深度网络模型的空间感知能力。3)增强中心深度回归:为了帮助中心深度估计,我们采用了基于几何形状的附加边界框顶点深度回归。我们的实验结果表明,所提出的技术优于文献中报道的现有竞争方法。该方法在单目三维物体检测中表现出了显著的性能提升,使其成为自动驾驶应用中一个前景广阔的解决方案。
{"title":"Depth-Enhanced Deep Learning Approach For Monocular Camera Based 3D Object Detection","authors":"Chuyao Wang, Nabil Aouf","doi":"10.1007/s10846-024-02128-w","DOIUrl":"https://doi.org/10.1007/s10846-024-02128-w","url":null,"abstract":"<p>Automatic 3D object detection using monocular cameras presents significant challenges in the context of autonomous driving. Precise labeling of 3D object scales requires accurate spatial information, which is difficult to obtain from a single image due to the inherent lack of depth information in monocular images, compared to LiDAR data. In this paper, we propose a novel approach to address this issue by enhancing deep neural networks with depth information for monocular 3D object detection. The proposed method comprises three key components: 1)Feature Enhancement Pyramid Module: We extend the conventional Feature Pyramid Networks (FPN) by introducing a feature enhancement pyramid network. This module fuses feature maps from the original pyramid and captures contextual correlations across multiple scales. To increase the connectivity between low-level and high-level features, additional pathways are incorporated. 2)Auxiliary Dense Depth Estimator: We introduce an auxiliary dense depth estimator that generates dense depth maps to enhance the spatial perception capabilities of the deep network model without adding computational burden. 3)Augmented Center Depth Regression: To aid center depth estimation, we employ additional bounding box vertex depth regression based on geometry. Our experimental results demonstrate the superiority of the proposed technique over existing competitive methods reported in the literature. The approach showcases remarkable performance improvements in monocular 3D object detection, making it a promising solution for autonomous driving applications.</p>","PeriodicalId":54794,"journal":{"name":"Journal of Intelligent & Robotic Systems","volume":"13 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141570675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
China’s New Pattern of Rule of Law on UAS 中国无人机系统法治新模式
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-09 DOI: 10.1007/s10846-024-02105-3
Luping Zhang

China as the world’s prominent market for Unmanned Aircraft Systems (UAS) has just passed a new regulation on UAS. The new regulation is expected to form a new pattern of rule of law on UAS in China. With the need for harmonisation of laws internationally, this article highlights the three aspects out of China’s new UAS legislation against an international setting.

中国作为世界上重要的无人机系统(UAS)市场,刚刚通过了一项关于无人机系统的新法规。新法规有望形成中国无人机系统法治的新格局。鉴于国际间法律协调的需要,本文从三个方面着重介绍了中国在国际背景下的无人机系统新立法。
{"title":"China’s New Pattern of Rule of Law on UAS","authors":"Luping Zhang","doi":"10.1007/s10846-024-02105-3","DOIUrl":"https://doi.org/10.1007/s10846-024-02105-3","url":null,"abstract":"<p>China as the world’s prominent market for Unmanned Aircraft Systems (UAS) has just passed a new regulation on UAS. The new regulation is expected to form a new pattern of rule of law on UAS in China. With the need for harmonisation of laws internationally, this article highlights the three aspects out of China’s new UAS legislation against an international setting.</p>","PeriodicalId":54794,"journal":{"name":"Journal of Intelligent & Robotic Systems","volume":"43 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141570672","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stereo-RIVO: Stereo-Robust Indirect Visual Odometry Stereo-RIVO: 立体稳固间接目视测距仪
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-09 DOI: 10.1007/s10846-024-02116-0
Erfan Salehi, Ali Aghagolzadeh, Reshad Hosseini

Mobile robots and autonomous systems rely on advanced guidance modules which often incorporate cameras to enable key functionalities. These modules are equipped with visual odometry (VO) and visual simultaneous localization and mapping (VSLAM) algorithms that work by analyzing changes between successive frames captured by cameras. VO/VSLAM-based systems are critical backbones for autonomous vehicles, virtual reality, structure from motion, and other robotic operations. VO/VSLAM systems encounter difficulties when implementing real-time applications in outdoor environments with restricted hardware and software platforms. While many VO systems target achieving high accuracy and speed, they often exhibit high degree of complexity and limited robustness. To overcome these challenges, this paper aims to propose a new VO system called Stereo-RIVO that balances accuracy, speed, and computational cost. Furthermore, this algorithm is based on a new data association module which consists of two primary components: a scene-matching process that achieves exceptional precision without feature extraction and a key-frame detection technique based on a model of scene movement. The performance of this proposed VO system has been tested extensively for all sequences of KITTI and UTIAS datasets for analyzing efficiency for outdoor dynamic and indoor static environments, respectively. The results of these tests indicate that the proposed Stereo-RIVO outperforms other state-of-the-art methods in terms of robustness, accuracy, and speed. Our implementation code of stereo-RIVO is available at: https://github.com/salehierfan/Stereo-RIVO.

移动机器人和自主系统依赖于先进的制导模块,这些模块通常包含摄像头以实现关键功能。这些模块配备了视觉里程测量(VO)和视觉同步定位与映射(VSLAM)算法,通过分析摄像头捕捉的连续帧之间的变化来工作。基于 VO/VSLAM 的系统是自动驾驶汽车、虚拟现实、运动结构和其他机器人操作的重要基础。在硬件和软件平台受限的室外环境中实施实时应用时,VO/VSLAM 系统会遇到困难。虽然许多虚拟机系统以实现高精度和高速度为目标,但它们往往表现出高度的复杂性和有限的鲁棒性。为了克服这些挑战,本文旨在提出一种名为 Stereo-RIVO 的新型虚拟化系统,它能在精度、速度和计算成本之间取得平衡。此外,该算法基于一个新的数据关联模块,该模块由两个主要部分组成:一个是无需特征提取即可实现超高精度的场景匹配过程,另一个是基于场景运动模型的关键帧检测技术。我们对 KITTI 和 UTIAS 数据集的所有序列进行了广泛的测试,以分析所提出的 VO 系统在室外动态环境和室内静态环境下的性能。测试结果表明,所提出的立体-RIVO 在鲁棒性、准确性和速度方面都优于其他最先进的方法。我们的立体-RIVO实现代码可在以下网址获取:https://github.com/salehierfan/Stereo-RIVO。
{"title":"Stereo-RIVO: Stereo-Robust Indirect Visual Odometry","authors":"Erfan Salehi, Ali Aghagolzadeh, Reshad Hosseini","doi":"10.1007/s10846-024-02116-0","DOIUrl":"https://doi.org/10.1007/s10846-024-02116-0","url":null,"abstract":"<p>Mobile robots and autonomous systems rely on advanced guidance modules which often incorporate cameras to enable key functionalities. These modules are equipped with visual odometry (VO) and visual simultaneous localization and mapping (VSLAM) algorithms that work by analyzing changes between successive frames captured by cameras. VO/VSLAM-based systems are critical backbones for autonomous vehicles, virtual reality, structure from motion, and other robotic operations. VO/VSLAM systems encounter difficulties when implementing real-time applications in outdoor environments with restricted hardware and software platforms. While many VO systems target achieving high accuracy and speed, they often exhibit high degree of complexity and limited robustness. To overcome these challenges, this paper aims to propose a new VO system called Stereo-RIVO that balances accuracy, speed, and computational cost. Furthermore, this algorithm is based on a new data association module which consists of two primary components: a scene-matching process that achieves exceptional precision without feature extraction and a key-frame detection technique based on a model of scene movement. The performance of this proposed VO system has been tested extensively for all sequences of KITTI and UTIAS datasets for analyzing efficiency for outdoor dynamic and indoor static environments, respectively. The results of these tests indicate that the proposed Stereo-RIVO outperforms other state-of-the-art methods in terms of robustness, accuracy, and speed. Our implementation code of stereo-RIVO is available at: https://github.com/salehierfan/Stereo-RIVO.</p>","PeriodicalId":54794,"journal":{"name":"Journal of Intelligent & Robotic Systems","volume":"11 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141570676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Intelligent & Robotic Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1