首页 > 最新文献

arXiv - CS - Robotics最新文献

英文 中文
AnySkin: Plug-and-play Skin Sensing for Robotic Touch AnySkin:用于机器人触摸的即插即用皮肤传感技术
Pub Date : 2024-09-12 DOI: arxiv-2409.08276
Raunaq Bhirangi, Venkatesh Pattabiraman, Enes Erciyes, Yifeng Cao, Tess Hellebrekers, Lerrel Pinto
While tactile sensing is widely accepted as an important and useful sensingmodality, its use pales in comparison to other sensory modalities like visionand proprioception. AnySkin addresses the critical challenges that impede theuse of tactile sensing -- versatility, replaceability, and data reusability.Building on the simplistic design of ReSkin, and decoupling the sensingelectronics from the sensing interface, AnySkin simplifies integration makingit as straightforward as putting on a phone case and connecting a charger.Furthermore, AnySkin is the first uncalibrated tactile-sensor withcross-instance generalizability of learned manipulation policies. To summarize,this work makes three key contributions: first, we introduce a streamlinedfabrication process and a design tool for creating an adhesive-free, durableand easily replaceable magnetic tactile sensor; second, we characterize slipdetection and policy learning with the AnySkin sensor; and third, wedemonstrate zero-shot generalization of models trained on one instance ofAnySkin to new instances, and compare it with popular existing tactilesolutions like DIGIT and ReSkin.https://any-skin.github.io/
虽然触觉传感作为一种重要而有用的传感方式已被广泛接受,但与视觉和本体感觉等其他传感方式相比,其应用却显得微不足道。AnySkin解决了阻碍触觉传感应用的关键难题--多功能性、可替换性和数据可重用性。AnySkin以ReSkin的简洁设计为基础,将传感电子元件与传感接口分离,简化了集成过程,使其就像装上手机壳和连接充电器一样简单。总之,这项工作有三个主要贡献:首先,我们介绍了一种简化的制造工艺和设计工具,用于制造一种无胶、耐用且易于更换的磁性触觉传感器;其次,我们描述了使用AnySkin传感器进行滑动检测和策略学习的特点;第三,我们演示了在AnySkin的一个实例上训练的模型对新实例的零次泛化,并将其与DIGIT和ReSkin等现有流行触觉解决方案进行了比较。https://any-skin.github.io/。
{"title":"AnySkin: Plug-and-play Skin Sensing for Robotic Touch","authors":"Raunaq Bhirangi, Venkatesh Pattabiraman, Enes Erciyes, Yifeng Cao, Tess Hellebrekers, Lerrel Pinto","doi":"arxiv-2409.08276","DOIUrl":"https://doi.org/arxiv-2409.08276","url":null,"abstract":"While tactile sensing is widely accepted as an important and useful sensing\u0000modality, its use pales in comparison to other sensory modalities like vision\u0000and proprioception. AnySkin addresses the critical challenges that impede the\u0000use of tactile sensing -- versatility, replaceability, and data reusability.\u0000Building on the simplistic design of ReSkin, and decoupling the sensing\u0000electronics from the sensing interface, AnySkin simplifies integration making\u0000it as straightforward as putting on a phone case and connecting a charger.\u0000Furthermore, AnySkin is the first uncalibrated tactile-sensor with\u0000cross-instance generalizability of learned manipulation policies. To summarize,\u0000this work makes three key contributions: first, we introduce a streamlined\u0000fabrication process and a design tool for creating an adhesive-free, durable\u0000and easily replaceable magnetic tactile sensor; second, we characterize slip\u0000detection and policy learning with the AnySkin sensor; and third, we\u0000demonstrate zero-shot generalization of models trained on one instance of\u0000AnySkin to new instances, and compare it with popular existing tactile\u0000solutions like DIGIT and ReSkin.https://any-skin.github.io/","PeriodicalId":501031,"journal":{"name":"arXiv - CS - Robotics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222087","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MosquitoMiner: A Light Weight Rover for Detecting and Eliminating Mosquito Breeding Sites MosquitoMiner:用于探测和消除蚊子繁殖地的轻型漫游车
Pub Date : 2024-09-12 DOI: arxiv-2409.08078
Md. Adnanul Islam, Md. Faiyaz Abdullah Sayeedi, Jannatul Ferdous Deepti, Shahanur Rahman Bappy, Safrin Sanzida Islam, Fahim Hafiz
In this paper, we present a novel approach to the development and deploymentof an autonomous mosquito breeding place detector rover with the object andobstacle detection capabilities to control mosquitoes. Mosquito-borne diseasescontinue to pose significant health threats globally, with conventional controlmethods proving slow and inefficient. Amidst rising concerns over the rapidspread of these diseases, there is an urgent need for innovative and efficientstrategies to manage mosquito populations and prevent disease transmission. Tomitigate the limitations of manual labor and traditional methods, our roveremploys autonomous control strategies. Leveraging our own custom dataset, therover can autonomously navigate along a pre-defined path, identifying andmitigating potential breeding grounds with precision. It then proceeds toeliminate these breeding grounds by spraying a chemical agent, effectivelyeradicating mosquito habitats. Our project demonstrates the effectiveness thatis absent in traditional ways of controlling and safeguarding public health.The code for this project is available on GitHub at -https://github.com/faiyazabdullah/MosquitoMiner
在本文中,我们提出了一种开发和部署具有物体和障碍物探测能力的自主蚊子滋生地探测漫游车的新方法,以控制蚊子。蚊子传播的疾病继续对全球健康构成重大威胁,而传统的控制方法证明既缓慢又低效。随着人们对这些疾病迅速蔓延的担忧日益加剧,迫切需要创新、高效的策略来管理蚊子种群和预防疾病传播。为了克服人工和传统方法的局限性,我们的研究采用了自主控制策略。利用我们自己定制的数据集,该机器可以沿着预定义的路径自主导航,精确地识别和清除潜在的滋生地。然后,它通过喷洒化学制剂来消灭这些滋生地,从而有效地消灭蚊子的栖息地。我们的项目展示了传统控制和保护公共健康方式所不具备的有效性。该项目的代码可在 GitHub 上获取,网址是:https://github.com/faiyazabdullah/MosquitoMiner。
{"title":"MosquitoMiner: A Light Weight Rover for Detecting and Eliminating Mosquito Breeding Sites","authors":"Md. Adnanul Islam, Md. Faiyaz Abdullah Sayeedi, Jannatul Ferdous Deepti, Shahanur Rahman Bappy, Safrin Sanzida Islam, Fahim Hafiz","doi":"arxiv-2409.08078","DOIUrl":"https://doi.org/arxiv-2409.08078","url":null,"abstract":"In this paper, we present a novel approach to the development and deployment\u0000of an autonomous mosquito breeding place detector rover with the object and\u0000obstacle detection capabilities to control mosquitoes. Mosquito-borne diseases\u0000continue to pose significant health threats globally, with conventional control\u0000methods proving slow and inefficient. Amidst rising concerns over the rapid\u0000spread of these diseases, there is an urgent need for innovative and efficient\u0000strategies to manage mosquito populations and prevent disease transmission. To\u0000mitigate the limitations of manual labor and traditional methods, our rover\u0000employs autonomous control strategies. Leveraging our own custom dataset, the\u0000rover can autonomously navigate along a pre-defined path, identifying and\u0000mitigating potential breeding grounds with precision. It then proceeds to\u0000eliminate these breeding grounds by spraying a chemical agent, effectively\u0000eradicating mosquito habitats. Our project demonstrates the effectiveness that\u0000is absent in traditional ways of controlling and safeguarding public health.\u0000The code for this project is available on GitHub at -\u0000https://github.com/faiyazabdullah/MosquitoMiner","PeriodicalId":501031,"journal":{"name":"arXiv - CS - Robotics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222094","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Collaborating for Success: Optimizing System Efficiency and Resilience Under Agile Industrial Settings 合作共赢:在灵活的工业环境下优化系统效率和复原力
Pub Date : 2024-09-12 DOI: arxiv-2409.08166
Sunny Katyara, Suchita Sharma, Praveen Damacharla, Carlos Garcia Santiago, Francis O'Farrell, Philip Long
Designing an efficient and resilient human-robot collaboration strategy thatnot only upholds the safety and ergonomics of shared workspace but alsoenhances the performance and agility of collaborative setup presentssignificant challenges concerning environment perception and robot control. Inthis research, we introduce a novel approach for collaborative environmentmonitoring and robot motion regulation to address this multifaceted problem.Our study proposes novel computation and division of safety monitoring zones,adhering to ISO 13855 and TS 15066 standards, utilizing 2D lasers information.These zones are not only configured in the standard three-layer arrangement butare also expanded into two adjacent quadrants, thereby enhancing system uptimeand preventing unnecessary deadlocks. Moreover, we also leverage 3D visualinformation to track dynamic human articulations and extended intrusions.Drawing upon the fused sensory data from 2D and 3D perceptual spaces, ourproposed hierarchical controller stably regulates robot velocity, validatedusing Lasalle in-variance principle. Empirical evaluations demonstrate that ourapproach significantly reduces task execution time and system response delay,resulting in improved efficiency and resilience within collaborative settings.
设计一种高效、有弹性的人机协作策略,不仅要维护共享工作空间的安全性和人体工程学,还要提高协作设置的性能和灵活性,这对环境感知和机器人控制提出了重大挑战。我们的研究根据 ISO 13855 和 TS 15066 标准,利用二维激光信息,提出了新颖的安全监控区域计算和划分方法。此外,我们还利用三维视觉信息来跟踪人类的动态动作和扩展入侵。利用来自二维和三维感知空间的融合感知数据,我们提出的分层控制器可以稳定地调节机器人的速度,并通过拉萨尔内方差原理进行了验证。经验评估表明,我们的方法大大缩短了任务执行时间和系统响应延迟,从而提高了协作环境下的效率和适应能力。
{"title":"Collaborating for Success: Optimizing System Efficiency and Resilience Under Agile Industrial Settings","authors":"Sunny Katyara, Suchita Sharma, Praveen Damacharla, Carlos Garcia Santiago, Francis O'Farrell, Philip Long","doi":"arxiv-2409.08166","DOIUrl":"https://doi.org/arxiv-2409.08166","url":null,"abstract":"Designing an efficient and resilient human-robot collaboration strategy that\u0000not only upholds the safety and ergonomics of shared workspace but also\u0000enhances the performance and agility of collaborative setup presents\u0000significant challenges concerning environment perception and robot control. In\u0000this research, we introduce a novel approach for collaborative environment\u0000monitoring and robot motion regulation to address this multifaceted problem.\u0000Our study proposes novel computation and division of safety monitoring zones,\u0000adhering to ISO 13855 and TS 15066 standards, utilizing 2D lasers information.\u0000These zones are not only configured in the standard three-layer arrangement but\u0000are also expanded into two adjacent quadrants, thereby enhancing system uptime\u0000and preventing unnecessary deadlocks. Moreover, we also leverage 3D visual\u0000information to track dynamic human articulations and extended intrusions.\u0000Drawing upon the fused sensory data from 2D and 3D perceptual spaces, our\u0000proposed hierarchical controller stably regulates robot velocity, validated\u0000using Lasalle in-variance principle. Empirical evaluations demonstrate that our\u0000approach significantly reduces task execution time and system response delay,\u0000resulting in improved efficiency and resilience within collaborative settings.","PeriodicalId":501031,"journal":{"name":"arXiv - CS - Robotics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222093","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Design of Informative Take-Over Requests for Semi-Autonomous Cyber-Physical Systems: Combining Spoken Language and Visual Icons in a Drone-Controller Setting 为半自主网络物理系统设计知情接管请求:在无人机控制器环境中结合口语和视觉图标
Pub Date : 2024-09-12 DOI: arxiv-2409.08253
Ashwini Gundappa, Emilia Ellsiepen, Lukas Schmitz, Frederik Wiehr, Vera Demberg
The question of how cyber-physical systems should interact with humanpartners that can take over control or exert oversight is becoming morepressing, as these systems are deployed for an ever larger range of tasks.Drawing on the literatures on handing over control during semi-autonomousdriving and human-robot interaction, we propose a design of a take-over requestthat combines an abstract pre-alert with an informative TOR: Relevant sensorinformation is highlighted on the controller's display, while a spoken messageverbalizes the reason for the TOR. We conduct our study in the context of asemi-autonomous drone control scenario as our testbed. The goal of our onlinestudy is to assess in more detail what form a language-based TOR should take.Specifically, we compare a full sentence condition to shorter fragments, andtest whether the visual highlighting should be done synchronously orasynchronously with the speech. Participants showed a higher accuracy inchoosing the correct solution with our bi-modal TOR and felt that they werebetter able to recognize the critical situation. Using only fragments in thespoken message rather than full sentences did not lead to improved accuracy orfaster reactions. Also, synchronizing the visual highlighting with the spokenmessage did not result in better accuracy and response times were evenincreased in this condition.
借鉴半自主驾驶和人机交互过程中控制权移交的相关文献,我们提出了一种接管请求的设计方案,它将抽象的预先警报与信息性的职权范围相结合:在控制器的显示屏上突出显示相关的传感器信息,同时用口语信息来平衡职权范围的原因。我们的研究以半自动无人机控制场景为测试平台。具体来说,我们比较了完整句子和较短的片段,并测试了视觉高亮应该与语音同步还是同步进行。结果显示,使用我们的双模态 TOR,受试者选择正确答案的准确率更高,而且他们认为自己能够更好地识别关键情况。只使用口语信息中的片段而不是完整句子并不能提高准确率或加快反应速度。此外,将视觉突出显示与口语信息同步也没有提高准确性,在这种情况下,反应时间甚至会增加。
{"title":"The Design of Informative Take-Over Requests for Semi-Autonomous Cyber-Physical Systems: Combining Spoken Language and Visual Icons in a Drone-Controller Setting","authors":"Ashwini Gundappa, Emilia Ellsiepen, Lukas Schmitz, Frederik Wiehr, Vera Demberg","doi":"arxiv-2409.08253","DOIUrl":"https://doi.org/arxiv-2409.08253","url":null,"abstract":"The question of how cyber-physical systems should interact with human\u0000partners that can take over control or exert oversight is becoming more\u0000pressing, as these systems are deployed for an ever larger range of tasks.\u0000Drawing on the literatures on handing over control during semi-autonomous\u0000driving and human-robot interaction, we propose a design of a take-over request\u0000that combines an abstract pre-alert with an informative TOR: Relevant sensor\u0000information is highlighted on the controller's display, while a spoken message\u0000verbalizes the reason for the TOR. We conduct our study in the context of a\u0000semi-autonomous drone control scenario as our testbed. The goal of our online\u0000study is to assess in more detail what form a language-based TOR should take.\u0000Specifically, we compare a full sentence condition to shorter fragments, and\u0000test whether the visual highlighting should be done synchronously or\u0000asynchronously with the speech. Participants showed a higher accuracy in\u0000choosing the correct solution with our bi-modal TOR and felt that they were\u0000better able to recognize the critical situation. Using only fragments in the\u0000spoken message rather than full sentences did not lead to improved accuracy or\u0000faster reactions. Also, synchronizing the visual highlighting with the spoken\u0000message did not result in better accuracy and response times were even\u0000increased in this condition.","PeriodicalId":501031,"journal":{"name":"arXiv - CS - Robotics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222121","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Q-value Regularized Decision ConvFormer for Offline Reinforcement Learning 用于离线强化学习的 Q 值正则化决策 ConvFormer
Pub Date : 2024-09-12 DOI: arxiv-2409.08062
Teng Yan, Zhendong Ruan, Yaobang Cai, Yu Han, Wenxian Li, Yang Zhang
As a data-driven paradigm, offline reinforcement learning (Offline RL) hasbeen formulated as sequence modeling, where the Decision Transformer (DT) hasdemonstrated exceptional capabilities. Unlike previous reinforcement learningmethods that fit value functions or compute policy gradients, DT adjusts theautoregressive model based on the expected returns, past states, and actions,using a causally masked Transformer to output the optimal action. However, dueto the inconsistency between the sampled returns within a single trajectory andthe optimal returns across multiple trajectories, it is challenging to set anexpected return to output the optimal action and stitch together suboptimaltrajectories. Decision ConvFormer (DC) is easier to understand in the contextof modeling RL trajectories within a Markov Decision Process compared to DT. Wepropose the Q-value Regularized Decision ConvFormer (QDC), which combines theunderstanding of RL trajectories by DC and incorporates a term that maximizesaction values using dynamic programming methods during training. This ensuresthat the expected returns of the sampled actions are consistent with theoptimal returns. QDC achieves excellent performance on the D4RL benchmark,outperforming or approaching the optimal level in all tested environments. Itparticularly demonstrates outstanding competitiveness in trajectory stitchingcapability.
作为一种数据驱动范例,离线强化学习(Offline RL)被表述为序列建模,其中的决策转换器(DT)已经展现出非凡的能力。与以往拟合值函数或计算策略梯度的强化学习方法不同,DT 根据预期收益、过去状态和行动调整自回归模型,并使用因果掩蔽变换器输出最优行动。然而,由于单条轨迹中的采样收益与多条轨迹中的最优收益不一致,要设置一个预期收益来输出最优行动并将次优轨迹拼接在一起是很有挑战性的。与 DT 相比,Decision ConvFormer(DC)在马尔可夫决策过程中的 RL 轨迹建模中更容易理解。我们提出了 Q 值正则化决策 ConvFormer(QDC),它结合了 DC 对 RL 轨迹的理解,并在训练过程中加入了使用动态编程方法最大化行动值的术语。这确保了采样行动的预期收益与最优收益一致。QDC 在 D4RL 基准测试中表现出色,在所有测试环境中都超过或接近最优水平。它在轨迹拼接能力方面尤其表现出了出色的竞争力。
{"title":"Q-value Regularized Decision ConvFormer for Offline Reinforcement Learning","authors":"Teng Yan, Zhendong Ruan, Yaobang Cai, Yu Han, Wenxian Li, Yang Zhang","doi":"arxiv-2409.08062","DOIUrl":"https://doi.org/arxiv-2409.08062","url":null,"abstract":"As a data-driven paradigm, offline reinforcement learning (Offline RL) has\u0000been formulated as sequence modeling, where the Decision Transformer (DT) has\u0000demonstrated exceptional capabilities. Unlike previous reinforcement learning\u0000methods that fit value functions or compute policy gradients, DT adjusts the\u0000autoregressive model based on the expected returns, past states, and actions,\u0000using a causally masked Transformer to output the optimal action. However, due\u0000to the inconsistency between the sampled returns within a single trajectory and\u0000the optimal returns across multiple trajectories, it is challenging to set an\u0000expected return to output the optimal action and stitch together suboptimal\u0000trajectories. Decision ConvFormer (DC) is easier to understand in the context\u0000of modeling RL trajectories within a Markov Decision Process compared to DT. We\u0000propose the Q-value Regularized Decision ConvFormer (QDC), which combines the\u0000understanding of RL trajectories by DC and incorporates a term that maximizes\u0000action values using dynamic programming methods during training. This ensures\u0000that the expected returns of the sampled actions are consistent with the\u0000optimal returns. QDC achieves excellent performance on the D4RL benchmark,\u0000outperforming or approaching the optimal level in all tested environments. It\u0000particularly demonstrates outstanding competitiveness in trajectory stitching\u0000capability.","PeriodicalId":501031,"journal":{"name":"arXiv - CS - Robotics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222124","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Skateboarding for Humanoid Robots through Massively Parallel Reinforcement Learning 通过大规模并行强化学习为仿人机器人学习滑板运动
Pub Date : 2024-09-12 DOI: arxiv-2409.07846
William Thibault, Vidyasagar Rajendran, William Melek, Katja Mombaur
Learning-based methods have proven useful at generating complex motions forrobots, including humanoids. Reinforcement learning (RL) has been used to learnlocomotion policies, some of which leverage a periodic reward formulation. Thiswork extends the periodic reward formulation of locomotion to skateboarding forthe REEM-C robot. Brax/MJX is used to implement the RL problem to achieve fasttraining. Initial results in simulation are presented with hardware experimentsin progress.
事实证明,基于学习的方法有助于为机器人(包括人形机器人)生成复杂的运动。强化学习(RL)已被用于学习运动策略,其中一些策略利用了周期性奖励公式。本研究将运动的周期性奖励公式扩展到 REEM-C 机器人的滑板运动。Brax/MJX 用于实现 RL 问题,以实现快速训练。本文介绍了仿真的初步结果,硬件实验正在进行中。
{"title":"Learning Skateboarding for Humanoid Robots through Massively Parallel Reinforcement Learning","authors":"William Thibault, Vidyasagar Rajendran, William Melek, Katja Mombaur","doi":"arxiv-2409.07846","DOIUrl":"https://doi.org/arxiv-2409.07846","url":null,"abstract":"Learning-based methods have proven useful at generating complex motions for\u0000robots, including humanoids. Reinforcement learning (RL) has been used to learn\u0000locomotion policies, some of which leverage a periodic reward formulation. This\u0000work extends the periodic reward formulation of locomotion to skateboarding for\u0000the REEM-C robot. Brax/MJX is used to implement the RL problem to achieve fast\u0000training. Initial results in simulation are presented with hardware experiments\u0000in progress.","PeriodicalId":501031,"journal":{"name":"arXiv - CS - Robotics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222098","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Touch2Touch: Cross-Modal Tactile Generation for Object Manipulation Touch2Touch:用于物体操作的跨模式触觉生成
Pub Date : 2024-09-12 DOI: arxiv-2409.08269
Samanta Rodriguez, Yiming Dou, Miquel Oller, Andrew Owens, Nima Fazeli
Today's touch sensors come in many shapes and sizes. This has made itchallenging to develop general-purpose touch processing methods since modelsare generally tied to one specific sensor design. We address this problem byperforming cross-modal prediction between touch sensors: given the tactilesignal from one sensor, we use a generative model to estimate how the samephysical contact would be perceived by another sensor. This allows us to applysensor-specific methods to the generated signal. We implement this idea bytraining a diffusion model to translate between the popular GelSlim and SoftBubble sensors. As a downstream task, we perform in-hand object pose estimationusing GelSlim sensors while using an algorithm that operates only on SoftBubble signals. The dataset, the code, and additional details can be found athttps://www.mmintlab.com/research/touch2touch/.
当今的触摸传感器有多种形状和尺寸。这给开发通用触摸处理方法带来了挑战,因为模型一般都与一种特定的传感器设计有关。我们通过在触摸传感器之间进行跨模态预测来解决这个问题:给定一个传感器的触觉信号,我们使用生成模型来估计另一个传感器将如何感知相同的物理接触。这样,我们就能对生成的信号应用特定于传感器的方法。我们通过训练扩散模型来实现这一想法,从而在流行的 GelSlim 和 SoftBubble 传感器之间进行转换。作为下游任务,我们使用 GelSlim 传感器执行手持物体姿态估计,同时使用一种仅在 SoftBubble 信号上运行的算法。有关数据集、代码和其他详细信息,请访问https://www.mmintlab.com/research/touch2touch/。
{"title":"Touch2Touch: Cross-Modal Tactile Generation for Object Manipulation","authors":"Samanta Rodriguez, Yiming Dou, Miquel Oller, Andrew Owens, Nima Fazeli","doi":"arxiv-2409.08269","DOIUrl":"https://doi.org/arxiv-2409.08269","url":null,"abstract":"Today's touch sensors come in many shapes and sizes. This has made it\u0000challenging to develop general-purpose touch processing methods since models\u0000are generally tied to one specific sensor design. We address this problem by\u0000performing cross-modal prediction between touch sensors: given the tactile\u0000signal from one sensor, we use a generative model to estimate how the same\u0000physical contact would be perceived by another sensor. This allows us to apply\u0000sensor-specific methods to the generated signal. We implement this idea by\u0000training a diffusion model to translate between the popular GelSlim and Soft\u0000Bubble sensors. As a downstream task, we perform in-hand object pose estimation\u0000using GelSlim sensors while using an algorithm that operates only on Soft\u0000Bubble signals. The dataset, the code, and additional details can be found at\u0000https://www.mmintlab.com/research/touch2touch/.","PeriodicalId":501031,"journal":{"name":"arXiv - CS - Robotics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Characterization and Design of A Hollow Cylindrical Ultrasonic Motor 中空圆柱形超声波电机的特性分析与设计
Pub Date : 2024-09-12 DOI: arxiv-2409.07690
Zhanyue Zhao, Yang Wang, Charles Bales, Daniel Ruiz-Cadalso, Howard Zheng, Cosme Furlong-Vazquez, Gregory Fischer
Piezoelectric ultrasonic motors perform the advantages of compact design,faster reaction time, and simpler setup compared to other motion units such aspneumatic and hydraulic motors, especially its non-ferromagnetic property makesit a perfect match in MRI-compatible robotics systems compared to traditionalDC motors. Hollow shaft motors address the advantages of being lightweight andcomparable to solid shafts of the same diameter, low rotational inertia, hightolerance to rotational imbalance due to low weight, and tolerance to hightemperature due to low specific mass. This article presents a prototype of ahollow cylindrical ultrasonic motor (HCM) to perform direct drive, eliminatemechanical non-linearity, and reduce the size and complexity of the actuator orend effector assembly. Two equivalent HCMs are presented in this work, andunder 50g prepressure on the rotor, it performed 383.3333rpm rotation speed and57.3504mNm torque output when applying 282$V_{pp}$ driving voltage.
与气动和液压电机等其他运动单元相比,压电超声波电机具有设计紧凑、反应速度快、安装简单等优点,特别是其无铁磁性,与传统直流电机相比,可完美地用于核磁共振兼容机器人系统。空心轴电机具有重量轻、可与相同直径的实心轴相媲美、转动惯量小、因重量轻而对转动不平衡的耐受性高、因比质量小而耐高温等优点。本文介绍了一种中空圆柱形超声波电机(HCM)的原型,该电机可实现直接驱动,消除机械非线性,并减小致动器和效应器组件的尺寸和复杂性。在转子上施加 50g 预压力的情况下,当施加 282V_{pp}$ 的驱动电压时,它能实现 383.3333rpm 的转速和 57.3504mNm 的扭矩输出。
{"title":"Characterization and Design of A Hollow Cylindrical Ultrasonic Motor","authors":"Zhanyue Zhao, Yang Wang, Charles Bales, Daniel Ruiz-Cadalso, Howard Zheng, Cosme Furlong-Vazquez, Gregory Fischer","doi":"arxiv-2409.07690","DOIUrl":"https://doi.org/arxiv-2409.07690","url":null,"abstract":"Piezoelectric ultrasonic motors perform the advantages of compact design,\u0000faster reaction time, and simpler setup compared to other motion units such as\u0000pneumatic and hydraulic motors, especially its non-ferromagnetic property makes\u0000it a perfect match in MRI-compatible robotics systems compared to traditional\u0000DC motors. Hollow shaft motors address the advantages of being lightweight and\u0000comparable to solid shafts of the same diameter, low rotational inertia, high\u0000tolerance to rotational imbalance due to low weight, and tolerance to high\u0000temperature due to low specific mass. This article presents a prototype of a\u0000hollow cylindrical ultrasonic motor (HCM) to perform direct drive, eliminate\u0000mechanical non-linearity, and reduce the size and complexity of the actuator or\u0000end effector assembly. Two equivalent HCMs are presented in this work, and\u0000under 50g prepressure on the rotor, it performed 383.3333rpm rotation speed and\u000057.3504mNm torque output when applying 282$V_{pp}$ driving voltage.","PeriodicalId":501031,"journal":{"name":"arXiv - CS - Robotics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A three-dimensional force estimation method for the cable-driven soft robot based on monocular images 基于单目图像的缆索驱动软体机器人三维力估算方法
Pub Date : 2024-09-12 DOI: arxiv-2409.08033
Xiaohan Zhu, Ran Bu, Zhen Li, Fan Xu, Hesheng Wang
Soft manipulators are known for their superiority in coping withhigh-safety-demanding interaction tasks, e.g., robot-assisted surgeries,elderly caring, etc. Yet the challenges residing in real-time contact feedbackhave hindered further applications in precise manipulation. This paper proposesan end-to-end network to estimate the 3D contact force of the soft robot, withthe aim of enhancing its capabilities in interactive tasks. The presentedmethod features directly utilizing monocular images fused with multidimensionalactuation information as the network inputs. This approach simplifies thepreprocessing of raw data compared to related studies that utilize 3D shapeinformation for network inputs, consequently reducing configurationreconstruction errors. The unified feature representation module is devised toelevate low-dimensional features from the system's actuation signals to thesame level as image features, facilitating smoother integration of multimodalinformation. The proposed method has been experimentally validated in the softrobot testbed, achieving satisfying accuracy in 3D force estimation (with amean relative error of 0.84% compared to the best-reported result of 2.2% inthe related works).
众所周知,软机械手在应对安全性要求较高的交互任务(如机器人辅助手术、老年人护理等)方面具有优势。然而,实时接触反馈所带来的挑战阻碍了其在精确操控领域的进一步应用。本文提出了一种端到端网络来估算软体机器人的三维接触力,旨在增强其在交互任务中的能力。该方法的特点是直接利用单目图像融合多维动作信息作为网络输入。与利用三维形状信息作为网络输入的相关研究相比,这种方法简化了原始数据的预处理,从而减少了配置重建误差。统一特征表示模块的设计目的是将系统执行信号中的低维特征提升到与图像特征相同的水平,从而促进多模态信息的平滑整合。所提出的方法已在软机器人测试平台上进行了实验验证,在三维力估算方面达到了令人满意的精度(平均相对误差为 0.84%,而相关著作中报告的最佳结果为 2.2%)。
{"title":"A three-dimensional force estimation method for the cable-driven soft robot based on monocular images","authors":"Xiaohan Zhu, Ran Bu, Zhen Li, Fan Xu, Hesheng Wang","doi":"arxiv-2409.08033","DOIUrl":"https://doi.org/arxiv-2409.08033","url":null,"abstract":"Soft manipulators are known for their superiority in coping with\u0000high-safety-demanding interaction tasks, e.g., robot-assisted surgeries,\u0000elderly caring, etc. Yet the challenges residing in real-time contact feedback\u0000have hindered further applications in precise manipulation. This paper proposes\u0000an end-to-end network to estimate the 3D contact force of the soft robot, with\u0000the aim of enhancing its capabilities in interactive tasks. The presented\u0000method features directly utilizing monocular images fused with multidimensional\u0000actuation information as the network inputs. This approach simplifies the\u0000preprocessing of raw data compared to related studies that utilize 3D shape\u0000information for network inputs, consequently reducing configuration\u0000reconstruction errors. The unified feature representation module is devised to\u0000elevate low-dimensional features from the system's actuation signals to the\u0000same level as image features, facilitating smoother integration of multimodal\u0000information. The proposed method has been experimentally validated in the soft\u0000robot testbed, achieving satisfying accuracy in 3D force estimation (with a\u0000mean relative error of 0.84% compared to the best-reported result of 2.2% in\u0000the related works).","PeriodicalId":501031,"journal":{"name":"arXiv - CS - Robotics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LED: Light Enhanced Depth Estimation at Night 发光二极管夜间光增强深度估计
Pub Date : 2024-09-12 DOI: arxiv-2409.08031
Simon de Moreau, Yasser Almehio, Andrei Bursuc, Hafid El-Idrissi, Bogdan Stanciulescu, Fabien Moutarde
Nighttime camera-based depth estimation is a highly challenging task,especially for autonomous driving applications, where accurate depth perceptionis essential for ensuring safe navigation. We aim to improve the reliability ofperception systems at night time, where models trained on daytime data oftenfail in the absence of precise but costly LiDAR sensors. In this work, weintroduce Light Enhanced Depth (LED), a novel cost-effective approach thatsignificantly improves depth estimation in low-light environments by harnessinga pattern projected by high definition headlights available in modern vehicles.LED leads to significant performance boosts across multiple depth-estimationarchitectures (encoder-decoder, Adabins, DepthFormer) both on synthetic andreal datasets. Furthermore, increased performances beyond illuminated areasreveal a holistic enhancement in scene understanding. Finally, we release theNighttime Synthetic Drive Dataset, a new synthetic and photo-realisticnighttime dataset, which comprises 49,990 comprehensively annotated images.
基于摄像头的夜间深度估计是一项极具挑战性的任务,尤其是在自动驾驶应用中,准确的深度感知对于确保导航安全至关重要。我们的目标是提高夜间感知系统的可靠性,因为在缺乏精确但昂贵的激光雷达传感器的情况下,根据白天数据训练的模型往往会失败。在这项工作中,我们引入了光增强深度(Light Enhanced Depth,LED),这是一种新颖的高性价比方法,通过利用现代汽车高清前大灯投射的图案,显著提高了低光环境下的深度估计能力。此外,照明区域之外的性能提升也显示出场景理解能力的全面增强。最后,我们发布了夜间合成驱动数据集(Nighttime Synthetic Drive Dataset),这是一个全新的合成和照片逼真夜间数据集,包含 49,990 张全面注释的图像。
{"title":"LED: Light Enhanced Depth Estimation at Night","authors":"Simon de Moreau, Yasser Almehio, Andrei Bursuc, Hafid El-Idrissi, Bogdan Stanciulescu, Fabien Moutarde","doi":"arxiv-2409.08031","DOIUrl":"https://doi.org/arxiv-2409.08031","url":null,"abstract":"Nighttime camera-based depth estimation is a highly challenging task,\u0000especially for autonomous driving applications, where accurate depth perception\u0000is essential for ensuring safe navigation. We aim to improve the reliability of\u0000perception systems at night time, where models trained on daytime data often\u0000fail in the absence of precise but costly LiDAR sensors. In this work, we\u0000introduce Light Enhanced Depth (LED), a novel cost-effective approach that\u0000significantly improves depth estimation in low-light environments by harnessing\u0000a pattern projected by high definition headlights available in modern vehicles.\u0000LED leads to significant performance boosts across multiple depth-estimation\u0000architectures (encoder-decoder, Adabins, DepthFormer) both on synthetic and\u0000real datasets. Furthermore, increased performances beyond illuminated areas\u0000reveal a holistic enhancement in scene understanding. Finally, we release the\u0000Nighttime Synthetic Drive Dataset, a new synthetic and photo-realistic\u0000nighttime dataset, which comprises 49,990 comprehensively annotated images.","PeriodicalId":501031,"journal":{"name":"arXiv - CS - Robotics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142222126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - CS - Robotics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1