首页 > 最新文献

ACM Transactions on Human-Robot Interaction最新文献

英文 中文
Brain-Behavior Relationships of Trust in Shared Space Human-Robot Collaboration 共享空间人机协作中信任的脑-行为关系
Q2 Computer Science Pub Date : 2023-11-10 DOI: 10.1145/3632149
Sarah K. Hopko, Yinsu Zhang, Aakash Yadav, Prabhakar R. Pagilla, Ranjana K. Mehta
Trust in human-robot collaboration is an essential consideration that relates to operator performance, utilization, and experience. While trust’s importance is understood, the state-of-the-art methods to study trust in automation, like surveys, drastically limit the types of insights that can be made. Improvements in measuring techniques can provide a granular understanding of influencers like robot reliability and their subsequent impact on human behavior and experience. This investigation quantifies the brain-behavior relationships associated with trust manipulation in shared space human-robot collaboration (HRC) to advance the scope of metrics to study trust. Thirty-eight participants, balanced by sex, were recruited to perform an assembly task with a collaborative robot under reliable and unreliable robot conditions. Brain imaging, psychological and behavioral eye-tracking, quantitative and qualitative performance, and subjective experiences were monitored. Results from this investigation identify specific information processing and cognitive strategies that result in identified trust-related behaviors, that were found to be sex-specific. The use of covert measurements of trust can reveal insights that humans cannot consciously report, thus shedding light on processes systematically overlooked by subjective measures. Our findings connect a trust influencer (robot reliability) to upstream cognition and downstream human behavior and are enabled by the utilization of granular metrics.
人机协作中的信任关系到操作者的性能、利用率和经验,是一个重要的考虑因素。虽然信任的重要性是众所周知的,但研究自动化中的信任的最先进的方法,如调查,极大地限制了可以获得的见解的类型。测量技术的改进可以让我们更细致地了解机器人可靠性等影响因素及其对人类行为和经验的后续影响。本研究量化了共享空间人机协作(HRC)中与信任操纵相关的脑-行为关系,以扩大研究信任的指标范围。38名参与者,按性别平衡,被招募在可靠和不可靠的机器人条件下与协作机器人一起执行组装任务。脑成像、心理和行为眼动追踪、定量和定性表现以及主观体验均被监测。这项调查的结果确定了导致确定的信任相关行为的特定信息处理和认知策略,这些行为被发现是性别特异性的。使用隐蔽的信任测量可以揭示人类无法有意识地报告的见解,从而揭示被主观测量系统地忽视的过程。我们的研究结果将信任影响者(机器人可靠性)与上游认知和下游人类行为联系起来,并通过使用粒度指标实现。
{"title":"Brain-Behavior Relationships of Trust in Shared Space Human-Robot Collaboration","authors":"Sarah K. Hopko, Yinsu Zhang, Aakash Yadav, Prabhakar R. Pagilla, Ranjana K. Mehta","doi":"10.1145/3632149","DOIUrl":"https://doi.org/10.1145/3632149","url":null,"abstract":"Trust in human-robot collaboration is an essential consideration that relates to operator performance, utilization, and experience. While trust’s importance is understood, the state-of-the-art methods to study trust in automation, like surveys, drastically limit the types of insights that can be made. Improvements in measuring techniques can provide a granular understanding of influencers like robot reliability and their subsequent impact on human behavior and experience. This investigation quantifies the brain-behavior relationships associated with trust manipulation in shared space human-robot collaboration (HRC) to advance the scope of metrics to study trust. Thirty-eight participants, balanced by sex, were recruited to perform an assembly task with a collaborative robot under reliable and unreliable robot conditions. Brain imaging, psychological and behavioral eye-tracking, quantitative and qualitative performance, and subjective experiences were monitored. Results from this investigation identify specific information processing and cognitive strategies that result in identified trust-related behaviors, that were found to be sex-specific. The use of covert measurements of trust can reveal insights that humans cannot consciously report, thus shedding light on processes systematically overlooked by subjective measures. Our findings connect a trust influencer (robot reliability) to upstream cognition and downstream human behavior and are enabled by the utilization of granular metrics.","PeriodicalId":36515,"journal":{"name":"ACM Transactions on Human-Robot Interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-11-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135138234","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Which Voice for which Robot? Designing Robot Voices that Indicate Robot Size 哪种声音适合哪个机器人?设计指示机器人大小的机器人声音
Q2 Computer Science Pub Date : 2023-11-08 DOI: 10.1145/3632124
Kerstin Fischer, Oliver Niebuhr
Many social robots will have the capacity to interact via speech in the future, and thus they will have to have a voice. However, so far it is unclear how we can create voices that fit their robotic speakers. In this paper, we explore how robot voices can be designed to fit the size of the respective robot. We therefore investigate the acoustic correlates of human voices and body size. In Study I, we analyzed 163 speech samples in connection with their speakers’ body size and body height. Our results show that specific acoustic parameters are significantly associated with body height, and to a lesser degree to body weight, but that different features are relevant for female and male voices. In Study II, we tested then for female and male voices to what extent the acoustic features identified can be used to create voices that are reliably associated with the size of robots. The results show that the acoustic features identified provide reliable clues to whether a large or a small robot is speaking.
未来,许多社交机器人将具备通过语音进行互动的能力,因此它们必须有声音。然而,到目前为止,我们还不清楚如何创造适合机器人扬声器的声音。在本文中,我们探讨了如何设计机器人声音以适应各自机器人的大小。因此,我们研究了人类声音和体型的声学相关性。在研究一中,我们分析了163个语音样本与说话人的体型和身高的关系。我们的研究结果表明,特定的声学参数与身高显著相关,与体重的关系较小,但女性和男性的声音有不同的特征。在研究II中,我们测试了女性和男性的声音,在多大程度上识别出的声学特征可以用来创建与机器人大小可靠相关的声音。结果表明,识别出的声学特征为判断机器人说话是大还是小提供了可靠的线索。
{"title":"Which Voice for which Robot? Designing Robot Voices that Indicate Robot Size","authors":"Kerstin Fischer, Oliver Niebuhr","doi":"10.1145/3632124","DOIUrl":"https://doi.org/10.1145/3632124","url":null,"abstract":"Many social robots will have the capacity to interact via speech in the future, and thus they will have to have a voice. However, so far it is unclear how we can create voices that fit their robotic speakers. In this paper, we explore how robot voices can be designed to fit the size of the respective robot. We therefore investigate the acoustic correlates of human voices and body size. In Study I, we analyzed 163 speech samples in connection with their speakers’ body size and body height. Our results show that specific acoustic parameters are significantly associated with body height, and to a lesser degree to body weight, but that different features are relevant for female and male voices. In Study II, we tested then for female and male voices to what extent the acoustic features identified can be used to create voices that are reliably associated with the size of robots. The results show that the acoustic features identified provide reliable clues to whether a large or a small robot is speaking.","PeriodicalId":36515,"journal":{"name":"ACM Transactions on Human-Robot Interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135391105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Assistance in Teleoperation of Redundant Robots through Predictive Joint Maneuvering 基于预测关节机动的冗余机器人遥操作辅助
Q2 Computer Science Pub Date : 2023-11-03 DOI: 10.1145/3630265
Connor Brooks, Wyatt Rees, Daniel Szafir
In teleoperation of redundant robotic manipulators, translating an operator’s end effector motion command to joint space can be a tool for maintaining feasible and precise robot motion. Through optimizing redundancy resolution, the control system can ensure the end effector maintains maneuverability by avoiding joint limits and kinematic singularities. In autonomous motion planning, this optimization can be done over an entire trajectory to improve performance over local optimization. However, teleoperation involves a human-in-the-loop who determines the trajectory to be executed through a dynamic sequence of motion commands. We present two systems, PrediKCT and PrediKCS, for utilizing a predictive model of operator commands in order to accomplish this redundancy resolution in a manner that considers future expected motion during teleoperation. Using a probabilistic model of operator commands allows optimization over an expected trajectory of future motion rather than consideration of local motion alone. Evaluation through a user study demonstrates improved control outcomes from this predictive redundancy resolution over minimum joint velocity solutions and inverse kinematics-based motion controllers.
在冗余机器人遥操作中,将操作者的末端执行器运动指令转换到关节空间是保持机器人运动可行和精确的一种工具。控制系统通过优化冗余分辨率,避免了关节极限和运动奇异性,保证了末端执行器保持可操作性。在自主运动规划中,这种优化可以在整个轨迹上进行,以提高局部优化的性能。然而,远程操作涉及到一个人在环谁决定轨迹要执行通过一个动态序列的运动命令。我们提出了两个系统,predickct和PrediKCS,用于利用操作员命令的预测模型,以便以考虑远程操作过程中未来预期运动的方式完成这种冗余解决方案。使用操作员命令的概率模型允许对未来运动的预期轨迹进行优化,而不是单独考虑局部运动。通过用户研究的评估表明,与最小关节速度解和基于逆运动学的运动控制器相比,这种预测冗余分辨率改善了控制效果。
{"title":"Assistance in Teleoperation of Redundant Robots through Predictive Joint Maneuvering","authors":"Connor Brooks, Wyatt Rees, Daniel Szafir","doi":"10.1145/3630265","DOIUrl":"https://doi.org/10.1145/3630265","url":null,"abstract":"In teleoperation of redundant robotic manipulators, translating an operator’s end effector motion command to joint space can be a tool for maintaining feasible and precise robot motion. Through optimizing redundancy resolution, the control system can ensure the end effector maintains maneuverability by avoiding joint limits and kinematic singularities. In autonomous motion planning, this optimization can be done over an entire trajectory to improve performance over local optimization. However, teleoperation involves a human-in-the-loop who determines the trajectory to be executed through a dynamic sequence of motion commands. We present two systems, PrediKCT and PrediKCS, for utilizing a predictive model of operator commands in order to accomplish this redundancy resolution in a manner that considers future expected motion during teleoperation. Using a probabilistic model of operator commands allows optimization over an expected trajectory of future motion rather than consideration of local motion alone. Evaluation through a user study demonstrates improved control outcomes from this predictive redundancy resolution over minimum joint velocity solutions and inverse kinematics-based motion controllers.","PeriodicalId":36515,"journal":{"name":"ACM Transactions on Human-Robot Interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135818720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robots’ “Woohoo” and “Argh” can Enhance Users’ Emotional and Social Perceptions: An Exploratory Study on Non-Lexical Vocalizations and Non-Linguistic Sounds 机器人的“Woohoo”和“Argh”可以增强用户的情感感知和社会感知——非词汇发声和非语言发声的探索性研究
Q2 Computer Science Pub Date : 2023-10-17 DOI: 10.1145/3626185
Xiaozhen Liu, Jiayuan Dong, Myounghoon Jeon
As robots have become more pervasive in our everyday life, social aspects of robots have attracted researchers’ attention. Because emotions play a crucial role in social interactions, research has been conducted on conveying emotions via speech. Our study sought to investigate the synchronization of multimodal interaction in human-robot interaction (HRI). We conducted a within-subjects exploratory study with 40 participants to investigate the effects of non-speech sounds (natural voice, synthesized voice, musical sound, and no sound) and basic emotions (anger, fear, happiness, sadness, and surprise) on user perception with emotional body gestures of an anthropomorphic robot (Pepper). While listening to a fairytale with the participant, a humanoid robot responded to the story with a recorded emotional non-speech sounds and gestures. Participants showed significantly higher emotion recognition accuracy from the natural voice than from other sounds. The confusion matrix showed that happiness and sadness had the highest emotion recognition accuracy, which is in line with previous research. The natural voice also induced higher trust, naturalness, and preference, compared to other sounds. Interestingly, the musical sound mostly showed lower perception ratings, even compared to the no sound. Results are discussed with design guidelines for emotional cues from social robots and future research directions.
随着机器人在我们的日常生活中变得越来越普遍,机器人的社交方面引起了研究人员的注意。由于情绪在社会交往中起着至关重要的作用,人们对通过言语传递情绪进行了研究。本研究旨在探讨人机交互(HRI)中多模态交互的同步性。我们对40名参与者进行了一项受试者内探索性研究,以调查非言语声音(自然声音、合成声音、音乐声和无声)和基本情绪(愤怒、恐惧、快乐、悲伤和惊讶)对拟人机器人(Pepper)情感肢体动作对用户感知的影响。在与参与者一起听童话故事的同时,一个人形机器人用记录下来的非言语的情感声音和手势来回应故事。参与者对自然声音的情绪识别准确率明显高于其他声音。混淆矩阵显示,快乐和悲伤的情绪识别准确率最高,这与前人的研究结果一致。与其他声音相比,自然的声音也会引起更高的信任、自然和偏好。有趣的是,即使与没有声音的声音相比,音乐声音也大多表现出较低的感知评分。讨论了社交机器人情感线索的设计准则和未来的研究方向。
{"title":"Robots’ “Woohoo” and “Argh” can Enhance Users’ Emotional and Social Perceptions: An Exploratory Study on Non-Lexical Vocalizations and Non-Linguistic Sounds","authors":"Xiaozhen Liu, Jiayuan Dong, Myounghoon Jeon","doi":"10.1145/3626185","DOIUrl":"https://doi.org/10.1145/3626185","url":null,"abstract":"As robots have become more pervasive in our everyday life, social aspects of robots have attracted researchers’ attention. Because emotions play a crucial role in social interactions, research has been conducted on conveying emotions via speech. Our study sought to investigate the synchronization of multimodal interaction in human-robot interaction (HRI). We conducted a within-subjects exploratory study with 40 participants to investigate the effects of non-speech sounds (natural voice, synthesized voice, musical sound, and no sound) and basic emotions (anger, fear, happiness, sadness, and surprise) on user perception with emotional body gestures of an anthropomorphic robot (Pepper). While listening to a fairytale with the participant, a humanoid robot responded to the story with a recorded emotional non-speech sounds and gestures. Participants showed significantly higher emotion recognition accuracy from the natural voice than from other sounds. The confusion matrix showed that happiness and sadness had the highest emotion recognition accuracy, which is in line with previous research. The natural voice also induced higher trust, naturalness, and preference, compared to other sounds. Interestingly, the musical sound mostly showed lower perception ratings, even compared to the no sound. Results are discussed with design guidelines for emotional cues from social robots and future research directions.","PeriodicalId":36515,"journal":{"name":"ACM Transactions on Human-Robot Interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136034445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Experimental Assessment of Human-Robot Teaming for Multi-Step Remote Manipulation with Expert Operators 专家操作下多步远程操作人机组队的实验评估
Q2 Computer Science Pub Date : 2023-10-17 DOI: 10.1145/3618258
Claudia Pérez-D'Arpino, Rebecca P. Khurshid, Julie A. Shah
Remote robot manipulation with human control enables applications where safety and environmental constraints are adverse to humans (e.g. underwater, space robotics and disaster response) or the complexity of the task demands human-level cognition and dexterity (e.g. robotic surgery and manufacturing). These systems typically use direct teleoperation at the motion level, and are usually limited to low-DOF arms and 2D perception. Improving dexterity and situational awareness demands new interaction and planning workflows. We explore the use of human-robot teaming through teleautonomy with assisted planning for remote control of a dual-arm dexterous robot for multi-step manipulation, and conduct a within-subjects experimental assessment (n=12 expert users) to compare it with direct teleoperation with an imitation controller with 2D and 3D perception, as well as teleoperation through a teleautonomy interface. The proposed assisted planning approach achieves task times comparable with direct teleoperation while improving other objective and subjective metrics, including re-grasps, collisions, and TLX workload. Assisted planning in the teleautonomy interface achieves faster task execution, and removes a significant interaction with the operator’s expertise level, resulting in a performance equalizer across users. Our study protocol, metrics and models for statistical analysis might also serve as a general benchmarking framework in teleoperation domains. Accompanying video and reference R code: https://people.csail.mit.edu/cdarpino/THRIteleop/
具有人类控制的远程机器人操作使安全和环境限制对人类不利的应用(例如水下,空间机器人和灾难响应)或任务的复杂性需要人类水平的认知和灵活性(例如机器人手术和制造)。这些系统通常在运动水平上使用直接远程操作,并且通常仅限于低自由度臂和2D感知。提高灵活性和态势感知需要新的交互和规划工作流程。我们探索了通过远程自主辅助规划的人机合作,对双臂灵巧机器人进行多步操作的远程控制,并进行了受试者内部实验评估(n=12名专家用户),将其与具有2D和3D感知的模拟控制器的直接远程操作以及通过远程自主界面的远程操作进行了比较。所提出的辅助规划方法实现了与直接遥操作相当的任务时间,同时改善了其他客观和主观指标,包括重新抓取、碰撞和TLX工作量。远程自治界面中的辅助规划实现了更快的任务执行,并消除了与操作员专业水平的重要交互,从而实现了跨用户的性能均衡器。我们的研究方案,统计分析的指标和模型也可以作为远程操作领域的一般基准框架。附带视频和参考R代码:https://people.csail.mit.edu/cdarpino/THRIteleop/
{"title":"Experimental Assessment of Human-Robot Teaming for Multi-Step Remote Manipulation with Expert Operators","authors":"Claudia Pérez-D'Arpino, Rebecca P. Khurshid, Julie A. Shah","doi":"10.1145/3618258","DOIUrl":"https://doi.org/10.1145/3618258","url":null,"abstract":"Remote robot manipulation with human control enables applications where safety and environmental constraints are adverse to humans (e.g. underwater, space robotics and disaster response) or the complexity of the task demands human-level cognition and dexterity (e.g. robotic surgery and manufacturing). These systems typically use direct teleoperation at the motion level, and are usually limited to low-DOF arms and 2D perception. Improving dexterity and situational awareness demands new interaction and planning workflows. We explore the use of human-robot teaming through teleautonomy with assisted planning for remote control of a dual-arm dexterous robot for multi-step manipulation, and conduct a within-subjects experimental assessment (n=12 expert users) to compare it with direct teleoperation with an imitation controller with 2D and 3D perception, as well as teleoperation through a teleautonomy interface. The proposed assisted planning approach achieves task times comparable with direct teleoperation while improving other objective and subjective metrics, including re-grasps, collisions, and TLX workload. Assisted planning in the teleautonomy interface achieves faster task execution, and removes a significant interaction with the operator’s expertise level, resulting in a performance equalizer across users. Our study protocol, metrics and models for statistical analysis might also serve as a general benchmarking framework in teleoperation domains. Accompanying video and reference R code: https://people.csail.mit.edu/cdarpino/THRIteleop/","PeriodicalId":36515,"journal":{"name":"ACM Transactions on Human-Robot Interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135944946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
IMPRINT: Interactional Dynamics-aware Motion Prediction in Teams using Multimodal Context 印记:使用多模态上下文的团队中的交互式动态感知运动预测
Q2 Computer Science Pub Date : 2023-10-16 DOI: 10.1145/3626954
Mohammad Samin Yasar, Md Mofijul Islam, Tariq Iqbal
Robots are moving from working in isolation to working with humans as a part of human-robot teams. In such situations, they are expected to work with multiple humans and need to understand and predict the team members’ actions. To address this challenge, in this work, we introduce IMPRINT, a multi-agent motion prediction framework that models the interactional dynamics and incorporates the multimodal context (e.g., data from RGB and depth sensors and skeleton joint positions) to accurately predict the motion of all the agents in a team. In IMPRINT, we propose an Interaction module that can extract the intra-agent and inter-agent dynamics before fusing them to obtain the interactional dynamics. Furthermore, we propose a Multimodal Context module that incorporates multimodal context information to improve multi-agent motion prediction. We evaluated IMPRINT by comparing its performance on human-human and human-robot team scenarios against state-of-the-art methods. The results suggest that IMPRINT outperformed all other methods over all evaluated temporal horizons. Additionally, we provide an interpretation of how IMPRINT incorporates the multimodal context information from all the modalities during multi-agent motion prediction. The superior performance of IMPRINT provides a promising direction to integrate motion prediction with robot perception and enable safe and effective human-robot collaboration.
机器人正在从孤立工作转向作为人-机器人团队的一部分与人类合作。在这种情况下,他们需要与多人一起工作,并且需要理解和预测团队成员的行为。为了应对这一挑战,在这项工作中,我们引入了IMPRINT,这是一个多智能体运动预测框架,它对交互动力学进行建模,并结合多模态上下文(例如,来自RGB和深度传感器以及骨骼关节位置的数据)来准确预测团队中所有智能体的运动。在IMPRINT中,我们提出了一个交互模块,该模块可以提取agent内和agent间的动态,然后将它们融合以获得交互动态。此外,我们提出了一个包含多模态上下文信息的多模态上下文模块,以改进多智能体运动预测。我们通过比较人与人和人机团队场景与最先进方法的表现来评估IMPRINT。结果表明,在所有评估的时间范围内,IMPRINT优于所有其他方法。此外,我们还解释了IMPRINT如何在多智能体运动预测过程中整合来自所有模态的多模态上下文信息。IMPRINT的优越性能为将运动预测与机器人感知相结合,实现安全有效的人机协作提供了一个有前景的方向。
{"title":"IMPRINT: Interactional Dynamics-aware Motion Prediction in Teams using Multimodal Context","authors":"Mohammad Samin Yasar, Md Mofijul Islam, Tariq Iqbal","doi":"10.1145/3626954","DOIUrl":"https://doi.org/10.1145/3626954","url":null,"abstract":"Robots are moving from working in isolation to working with humans as a part of human-robot teams. In such situations, they are expected to work with multiple humans and need to understand and predict the team members’ actions. To address this challenge, in this work, we introduce IMPRINT, a multi-agent motion prediction framework that models the interactional dynamics and incorporates the multimodal context (e.g., data from RGB and depth sensors and skeleton joint positions) to accurately predict the motion of all the agents in a team. In IMPRINT, we propose an Interaction module that can extract the intra-agent and inter-agent dynamics before fusing them to obtain the interactional dynamics. Furthermore, we propose a Multimodal Context module that incorporates multimodal context information to improve multi-agent motion prediction. We evaluated IMPRINT by comparing its performance on human-human and human-robot team scenarios against state-of-the-art methods. The results suggest that IMPRINT outperformed all other methods over all evaluated temporal horizons. Additionally, we provide an interpretation of how IMPRINT incorporates the multimodal context information from all the modalities during multi-agent motion prediction. The superior performance of IMPRINT provides a promising direction to integrate motion prediction with robot perception and enable safe and effective human-robot collaboration.","PeriodicalId":36515,"journal":{"name":"ACM Transactions on Human-Robot Interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136078655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Face2Gesture: Translating Facial Expressions Into Robot Movements Through Shared Latent Space Neural Networks Face2Gesture:通过共享潜在空间神经网络将面部表情转化为机器人动作
Q2 Computer Science Pub Date : 2023-10-04 DOI: 10.1145/3623386
Michael Suguitan, Nick DePalma, Guy Hoffman, Jessica Hodgins
In this work, we present a method for personalizing human-robot interaction by using emotive facial expressions to generate affective robot movements. Movement is an important medium for robots to communicate affective states, but the expertise and time required to craft new robot movements promotes a reliance on fixed preprogrammed behaviors. Enabling robots to respond to multimodal user input with newly generated movements could stave off staleness of interaction and convey a deeper degree of affective understanding than current retrieval-based methods. We use autoencoder neural networks to compress robot movement data and facial expression images into a shared latent embedding space. Then, we use a reconstruction loss to generate movements from these embeddings and triplet loss to align the embeddings by emotion classes rather than data modality. To subjectively evaluate our method, we conducted a user survey and found that generated happy and sad movements could be matched to their source face images. However, angry movements were most often mismatched to sad images. This multimodal data-driven generative method can expand an interactive agent’s behavior library and could be adopted for other multimodal affective applications.
在这项工作中,我们提出了一种通过使用情感面部表情来产生情感机器人动作的个性化人机交互方法。运动是机器人交流情感状态的重要媒介,但制作新的机器人运动所需的专业知识和时间促进了对固定预编程行为的依赖。使机器人能够用新生成的动作响应多模态用户输入,可以避免交互的陈旧,并传达比当前基于检索的方法更深程度的情感理解。我们使用自编码器神经网络将机器人运动数据和面部表情图像压缩到一个共享的潜在嵌入空间中。然后,我们使用重建损失来从这些嵌入中生成运动,并使用三重损失来根据情感类别而不是数据模式对齐嵌入。为了主观地评价我们的方法,我们对用户进行了调查,发现生成的快乐和悲伤的动作可以与他们的源面部图像相匹配。然而,愤怒的动作通常与悲伤的图像不匹配。这种多模态数据驱动生成方法可以扩展交互式智能体的行为库,并可用于其他多模态情感应用。
{"title":"Face2Gesture: Translating Facial Expressions Into Robot Movements Through Shared Latent Space Neural Networks","authors":"Michael Suguitan, Nick DePalma, Guy Hoffman, Jessica Hodgins","doi":"10.1145/3623386","DOIUrl":"https://doi.org/10.1145/3623386","url":null,"abstract":"In this work, we present a method for personalizing human-robot interaction by using emotive facial expressions to generate affective robot movements. Movement is an important medium for robots to communicate affective states, but the expertise and time required to craft new robot movements promotes a reliance on fixed preprogrammed behaviors. Enabling robots to respond to multimodal user input with newly generated movements could stave off staleness of interaction and convey a deeper degree of affective understanding than current retrieval-based methods. We use autoencoder neural networks to compress robot movement data and facial expression images into a shared latent embedding space. Then, we use a reconstruction loss to generate movements from these embeddings and triplet loss to align the embeddings by emotion classes rather than data modality. To subjectively evaluate our method, we conducted a user survey and found that generated happy and sad movements could be matched to their source face images. However, angry movements were most often mismatched to sad images. This multimodal data-driven generative method can expand an interactive agent’s behavior library and could be adopted for other multimodal affective applications.","PeriodicalId":36515,"journal":{"name":"ACM Transactions on Human-Robot Interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135596832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
“Do this instead” – Robots that Adequately Respond to Corrected Instructions “做这个”——对纠正指令做出充分反应的机器人
Q2 Computer Science Pub Date : 2023-09-22 DOI: 10.1145/3623385
Christopher Thierauf, Ravenna Thielstrom, Bradley Oosterveld, Will Becker, Matthias Scheutz
Natural language instructions are effective at tasking autonomous robots and for teaching them new knowledge quickly. Yet, human instructors are not perfect and are likely to make mistakes at times, and will correct themselves when they notice errors in their own instructions. In this paper, we introduce a complete system for robot behaviors to handle such corrections, during both task instruction and action execution. We then demonstrate its operation in an integrated cognitive robotic architecture through spoken language in two tasks: a navigation and retrieval task and a meal assembly task. Verbal corrections occur before, during, and after verbally taught sequences of tasks, demonstrating that the proposed methods enable fast corrections not only of the semantics generated from the instructions, but also of overt robot behavior in a manner shown to be reasonable when compared to human behavior and expectations.
自然语言指令在给自主机器人分配任务和快速教授新知识方面是有效的。然而,人类教师并不完美,有时可能会犯错误,当他们注意到自己的指示中的错误时,他们会纠正自己。在本文中,我们引入了一个完整的机器人行为系统,在任务指令和动作执行过程中处理这种纠正。然后,我们通过口语在两个任务中演示其在集成认知机器人架构中的操作:导航和检索任务以及饭菜组装任务。口头纠正发生在口头教导任务序列之前、期间和之后,这表明所提出的方法不仅能够快速纠正指令生成的语义,而且能够以与人类行为和期望相比合理的方式纠正明显的机器人行为。
{"title":"“Do this instead” – Robots that Adequately Respond to Corrected Instructions","authors":"Christopher Thierauf, Ravenna Thielstrom, Bradley Oosterveld, Will Becker, Matthias Scheutz","doi":"10.1145/3623385","DOIUrl":"https://doi.org/10.1145/3623385","url":null,"abstract":"Natural language instructions are effective at tasking autonomous robots and for teaching them new knowledge quickly. Yet, human instructors are not perfect and are likely to make mistakes at times, and will correct themselves when they notice errors in their own instructions. In this paper, we introduce a complete system for robot behaviors to handle such corrections, during both task instruction and action execution. We then demonstrate its operation in an integrated cognitive robotic architecture through spoken language in two tasks: a navigation and retrieval task and a meal assembly task. Verbal corrections occur before, during, and after verbally taught sequences of tasks, demonstrating that the proposed methods enable fast corrections not only of the semantics generated from the instructions, but also of overt robot behavior in a manner shown to be reasonable when compared to human behavior and expectations.","PeriodicalId":36515,"journal":{"name":"ACM Transactions on Human-Robot Interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136060335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unified Learning from Demonstrations, Corrections, and Preferences during Physical Human-Robot Interaction 在物理人机交互过程中,从演示、修正和偏好中统一学习
Q2 Computer Science Pub Date : 2023-09-22 DOI: 10.1145/3623384
Shaunak A. Mehta, Dylan P. Losey
Humans can leverage physical interaction to teach robot arms. This physical interaction takes multiple forms depending on the task, the user, and what the robot has learned so far. State-of-the-art approaches focus on learning from a single modality, or combine some interaction types. Some methods do so by assuming that the robot has prior information about the features of the task and the reward structure. By contrast, in this paper we introduce an algorithmic formalism that unites learning from demonstrations, corrections, and preferences. Our approach makes no assumptions about the tasks the human wants to teach the robot; instead, we learn a reward model from scratch by comparing the human’s input to nearby alternatives, i.e., trajectories close to the human’s feedback. We first derive a loss function that trains an ensemble of reward models to match the human’s demonstrations, corrections, and preferences. The type and order of feedback is up to the human teacher: we enable the robot to collect this feedback passively or actively. We then apply constrained optimization to convert our learned reward into a desired robot trajectory. Through simulations and a user study we demonstrate that our proposed approach more accurately learns manipulation tasks from physical human interaction than existing baselines, particularly when the robot is faced with new or unexpected objectives. Videos of our user study are available at: https://youtu.be/FSUJsTYvEKU
人类可以利用物理互动来教机器人手臂。这种物理交互根据任务、用户和机器人迄今所学的知识采取多种形式。最先进的方法侧重于从单一模式学习,或者结合一些交互类型。一些方法通过假设机器人具有关于任务特征和奖励结构的先验信息来做到这一点。相比之下,在本文中,我们引入了一种算法形式主义,它将从演示、修正和偏好中学习结合起来。我们的方法没有对人类想教机器人的任务做任何假设;相反,我们通过将人类的输入与附近的替代方案(即接近人类反馈的轨迹)进行比较,从头开始学习奖励模型。我们首先推导了一个损失函数,该函数训练了一个奖励模型集合,以匹配人类的演示、纠正和偏好。反馈的类型和顺序取决于人类老师:我们使机器人能够被动或主动地收集这些反馈。然后,我们应用约束优化将我们学习到的奖励转换为期望的机器人轨迹。通过模拟和用户研究,我们证明了我们提出的方法比现有的基线更准确地从物理人机交互中学习操作任务,特别是当机器人面临新的或意想不到的目标时。我们的用户研究视频可以在https://youtu.be/FSUJsTYvEKU上找到
{"title":"Unified Learning from Demonstrations, Corrections, and Preferences during Physical Human-Robot Interaction","authors":"Shaunak A. Mehta, Dylan P. Losey","doi":"10.1145/3623384","DOIUrl":"https://doi.org/10.1145/3623384","url":null,"abstract":"Humans can leverage physical interaction to teach robot arms. This physical interaction takes multiple forms depending on the task, the user, and what the robot has learned so far. State-of-the-art approaches focus on learning from a single modality, or combine some interaction types. Some methods do so by assuming that the robot has prior information about the features of the task and the reward structure. By contrast, in this paper we introduce an algorithmic formalism that unites learning from demonstrations, corrections, and preferences. Our approach makes no assumptions about the tasks the human wants to teach the robot; instead, we learn a reward model from scratch by comparing the human’s input to nearby alternatives, i.e., trajectories close to the human’s feedback. We first derive a loss function that trains an ensemble of reward models to match the human’s demonstrations, corrections, and preferences. The type and order of feedback is up to the human teacher: we enable the robot to collect this feedback passively or actively. We then apply constrained optimization to convert our learned reward into a desired robot trajectory. Through simulations and a user study we demonstrate that our proposed approach more accurately learns manipulation tasks from physical human interaction than existing baselines, particularly when the robot is faced with new or unexpected objectives. Videos of our user study are available at: https://youtu.be/FSUJsTYvEKU","PeriodicalId":36515,"journal":{"name":"ACM Transactions on Human-Robot Interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136062298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
UHTP: A User-Aware Hierarchical Task Planning Framework for Communication-Free, Mutually-Adaptive Human-Robot Collaboration http:面向无通信、相互适应的人机协作的用户感知分层任务规划框架
Q2 Computer Science Pub Date : 2023-09-22 DOI: 10.1145/3623387
Kartik Ramachandruni, Cassandra Kent, Sonia Chernova
Collaborative human-robot task execution approaches require mutual adaptation, allowing both the human and robot partners to take active roles in action selection and role assignment to achieve a single shared goal. Prior works have utilized a leader-follower paradigm in which either agent must follow the actions specified by the other agent. We introduce the User-aware Hierarchical Task Planning (UHTP) framework, a communication-free human-robot collaborative approach for adaptive execution of multi-step tasks that moves beyond the leader-follower paradigm. Specifically, our approach enables the robot to observe the human, perform actions that support the human’s decisions, and actively select actions that maximize the expected efficiency of the collaborative task. In turn, the human chooses actions based on their observation of the task and the robot, without being dictated by a scheduler or the robot. We evaluate UHTP both in simulation and in a human subjects experiment of a collaborative drill assembly task. Our results show that UHTP achieves more efficient task plans and shorter task completion times than non-adaptive baselines across a wide range of human behaviors, that interacting with a UHTP-controlled robot reduces the human’s cognitive workload, and that humans prefer to work with our adaptive robot over a fixed-policy alternative.
协作式人机任务执行方法需要相互适应,允许人和机器人合作伙伴在行动选择和角色分配中发挥积极作用,以实现单一的共享目标。先前的研究使用了领导者-追随者范式,其中任何一个代理都必须遵循另一个代理指定的动作。我们介绍了用户感知分层任务规划(http)框架,这是一种无需通信的人机协作方法,用于自适应执行多步骤任务,超越了领导者-追随者范式。具体来说,我们的方法使机器人能够观察人类,执行支持人类决策的动作,并主动选择最大限度提高协作任务预期效率的动作。反过来,人类根据他们对任务和机器人的观察来选择行动,而不受调度程序或机器人的支配。我们在模拟和协作演练装配任务的人类受试者实验中评估UHTP。我们的研究结果表明,在广泛的人类行为中,与非自适应基线相比,uhttp实现了更有效的任务计划和更短的任务完成时间,与uhttp控制的机器人交互减少了人类的认知工作量,并且与固定策略替代方案相比,人类更喜欢与我们的自适应机器人一起工作。
{"title":"UHTP: A User-Aware Hierarchical Task Planning Framework for Communication-Free, Mutually-Adaptive Human-Robot Collaboration","authors":"Kartik Ramachandruni, Cassandra Kent, Sonia Chernova","doi":"10.1145/3623387","DOIUrl":"https://doi.org/10.1145/3623387","url":null,"abstract":"Collaborative human-robot task execution approaches require mutual adaptation, allowing both the human and robot partners to take active roles in action selection and role assignment to achieve a single shared goal. Prior works have utilized a leader-follower paradigm in which either agent must follow the actions specified by the other agent. We introduce the User-aware Hierarchical Task Planning (UHTP) framework, a communication-free human-robot collaborative approach for adaptive execution of multi-step tasks that moves beyond the leader-follower paradigm. Specifically, our approach enables the robot to observe the human, perform actions that support the human’s decisions, and actively select actions that maximize the expected efficiency of the collaborative task. In turn, the human chooses actions based on their observation of the task and the robot, without being dictated by a scheduler or the robot. We evaluate UHTP both in simulation and in a human subjects experiment of a collaborative drill assembly task. Our results show that UHTP achieves more efficient task plans and shorter task completion times than non-adaptive baselines across a wide range of human behaviors, that interacting with a UHTP-controlled robot reduces the human’s cognitive workload, and that humans prefer to work with our adaptive robot over a fixed-policy alternative.","PeriodicalId":36515,"journal":{"name":"ACM Transactions on Human-Robot Interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136059980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ACM Transactions on Human-Robot Interaction
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1