{"title":"面向基于手势的货物搬运无人机合作","authors":"Marvin Brenner, Peter Stutz","doi":"10.1142/s1793351x23620015","DOIUrl":null,"url":null,"abstract":"This work provides the fundament for a gesture-based interaction system between cargo-handling unmanned aerial vehicles (UAVs) and ground personnel. It enables novice operators to visually communicate commands with higher abstractions through a minimum number of necessary gestures. The interaction concept intends to transfer two goal-directed control techniques to a cargo-handling use case: Selecting objects via deictic pointing communicates intention and a single proxy manipulation gesture controls the UAV’s flight. A visual processing pipeline built around an RGB-D sensor is presented and its subordinate components like lightweight object detectors and human pose estimation methods are benchmarked on the UAV-Human dataset. The results provide an overview of suitable methods for 3D gesture-based human drone interaction. A first unoptimized model ensemble runs with 7[Formula: see text]Hz on a Jetson Orin AGX Developer Kit.","PeriodicalId":43471,"journal":{"name":"International Journal of Semantic Computing","volume":"5 1","pages":"0"},"PeriodicalIF":0.3000,"publicationDate":"2023-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Towards gesture-based cooperation with cargo handling unmanned aerial vehicles\",\"authors\":\"Marvin Brenner, Peter Stutz\",\"doi\":\"10.1142/s1793351x23620015\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This work provides the fundament for a gesture-based interaction system between cargo-handling unmanned aerial vehicles (UAVs) and ground personnel. It enables novice operators to visually communicate commands with higher abstractions through a minimum number of necessary gestures. The interaction concept intends to transfer two goal-directed control techniques to a cargo-handling use case: Selecting objects via deictic pointing communicates intention and a single proxy manipulation gesture controls the UAV’s flight. A visual processing pipeline built around an RGB-D sensor is presented and its subordinate components like lightweight object detectors and human pose estimation methods are benchmarked on the UAV-Human dataset. The results provide an overview of suitable methods for 3D gesture-based human drone interaction. A first unoptimized model ensemble runs with 7[Formula: see text]Hz on a Jetson Orin AGX Developer Kit.\",\"PeriodicalId\":43471,\"journal\":{\"name\":\"International Journal of Semantic Computing\",\"volume\":\"5 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.3000,\"publicationDate\":\"2023-08-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Journal of Semantic Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1142/s1793351x23620015\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Semantic Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1142/s1793351x23620015","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
摘要
该工作为货物搬运无人机与地面人员之间基于手势的交互系统提供了基础。它使新手操作员能够通过最少数量的必要手势,以更高的抽象直观地传达命令。交互概念旨在将两种目标导向控制技术转移到货物处理用例中:通过指示指向选择对象来传达意图,而单个代理操作手势控制无人机的飞行。提出了一种围绕RGB-D传感器构建的视觉处理管道,并在无人机-人类数据集上对其下属组件如轻型目标检测器和人体姿态估计方法进行了基准测试。研究结果为基于3D手势的人类无人机交互提供了合适的方法概述。第一个未优化的模型集合在Jetson Orin AGX Developer Kit上以7 Hz运行。
Towards gesture-based cooperation with cargo handling unmanned aerial vehicles
This work provides the fundament for a gesture-based interaction system between cargo-handling unmanned aerial vehicles (UAVs) and ground personnel. It enables novice operators to visually communicate commands with higher abstractions through a minimum number of necessary gestures. The interaction concept intends to transfer two goal-directed control techniques to a cargo-handling use case: Selecting objects via deictic pointing communicates intention and a single proxy manipulation gesture controls the UAV’s flight. A visual processing pipeline built around an RGB-D sensor is presented and its subordinate components like lightweight object detectors and human pose estimation methods are benchmarked on the UAV-Human dataset. The results provide an overview of suitable methods for 3D gesture-based human drone interaction. A first unoptimized model ensemble runs with 7[Formula: see text]Hz on a Jetson Orin AGX Developer Kit.