Pub Date : 2014-07-10DOI: 10.5898/JHRI.3.2.Lazzeri
N. Lazzeri, D. Mazzei, D. De Rossi
Human-Robot Interaction (HRI) studies have recently received increasing attention in various fields, from academic communities to engineering firms and the media. Many researchers have been focusing on the development of tools to evaluate the performance of robotic systems and studying how to extend the range of robot interaction modalities and contexts. Because people are emotionally engaged when interacting with computers and robots, researchers have been focusing attention on the study of affective human-robot interaction. This new field of study requires the integration of various approaches typical of different research backgrounds, such as psychology and engineering, to gain more insight into the human-robot affective interaction. In this paper, we report the development of a multimodal acquisition platform called HIPOP (Human Interaction Pervasive Observation Platform). HIPOP is a modular data-gathering platform based on various hardware and software units that can be easily used to create a custom acquisition setup for HRI studies. The platform uses modules for physiological signals, eye gaze, video and audio acquisition to perform an integrated affective and behavioral analysis. It is also possible to include new hardware devices into the platform. The open-source hardware and software revolution has made many high-quality commercial and open-source products freely available for HRI and HCI research. These devices are currently most often used for data acquisition and robot control, and they can be easily included in HIPOP. Technical tests demonstrated the ability of HIPOP to reliably acquire a large set of data in terms of failure management and data synchronization. The platform was able to automatically recover from errors and faults without affecting the entire system, and the misalignment observed in the acquired data was not significant and did not affect the multimodal analysis. HIPOP was also tested in the context of the FACET (FACE Therapy) project, in which a humanoid robot called FACE (Facial Automaton for Conveying Emotions) was used to convey affective stimuli to children with autism. In the FACET project, psychologists without technical skills were able to use HIPOP to collect the data needed for their experiments without dealing with hardware issues, data integration challenges, or synchronization problems. The FACET case study highlighted the real core feature of the HIPOP platform (i.e., multimodal data integration and fusion). This analytical approach allowed psychologists to study both behavioral and psychophysiological reactions to obtain a more complete view of the subjects' state during interaction with the robot. These results indicate that HIPOP could become an innovative tool for HRI affective studies aimed at inferring a more detailed view of a subject's feelings and behavior during interaction with affective and empathic robots.
人机交互(HRI)研究最近在各个领域受到越来越多的关注,从学术界到工程公司和媒体。许多研究者一直致力于开发评估机器人系统性能的工具,并研究如何扩展机器人交互方式和环境的范围。由于人在与计算机和机器人的互动中具有情感投入,因此情感人机交互的研究一直是研究人员关注的焦点。这个新的研究领域需要整合不同研究背景的各种典型方法,如心理学和工程学,以更深入地了解人与机器人的情感互动。在本文中,我们报告了一种称为HIPOP(人类交互普遍观察平台)的多模式采集平台的开发。HIPOP是一个基于各种硬件和软件单元的模块化数据收集平台,可以很容易地用于创建HRI研究的自定义采集设置。该平台使用生理信号、眼睛注视、视频和音频采集模块来执行综合情感和行为分析。还可以在平台中包含新的硬件设备。开源硬件和软件革命已经为HRI和HCI研究提供了许多高质量的商业和开源产品。这些设备目前最常用于数据采集和机器人控制,它们可以很容易地包含在HIPOP中。技术测试表明,HIPOP能够在故障管理和数据同步方面可靠地获取大量数据。该平台能够在不影响整个系统的情况下自动从错误和故障中恢复,并且在获取的数据中观察到的偏差不显著,也不影响多模态分析。HIPOP还在FACET (FACE Therapy)项目的背景下进行了测试,在这个项目中,一个名为FACE (Facial Automaton for convey Emotions)的人形机器人被用来向自闭症儿童传递情感刺激。在FACET项目中,没有技术技能的心理学家可以使用HIPOP来收集实验所需的数据,而无需处理硬件问题、数据集成挑战或同步问题。FACET案例研究突出了HIPOP平台的真正核心特征(即多模式数据集成和融合)。这种分析方法使心理学家能够研究行为和心理生理反应,从而更全面地了解受试者在与机器人互动时的状态。这些结果表明,HIPOP可以成为HRI情感研究的创新工具,旨在推断出受试者在与情感和移情机器人互动时更详细的感受和行为。
{"title":"Development and testing of a multimodal acquisition platform for human-robot interaction affective studies","authors":"N. Lazzeri, D. Mazzei, D. De Rossi","doi":"10.5898/JHRI.3.2.Lazzeri","DOIUrl":"https://doi.org/10.5898/JHRI.3.2.Lazzeri","url":null,"abstract":"Human-Robot Interaction (HRI) studies have recently received increasing attention in various fields, from academic communities to engineering firms and the media. Many researchers have been focusing on the development of tools to evaluate the performance of robotic systems and studying how to extend the range of robot interaction modalities and contexts. Because people are emotionally engaged when interacting with computers and robots, researchers have been focusing attention on the study of affective human-robot interaction. This new field of study requires the integration of various approaches typical of different research backgrounds, such as psychology and engineering, to gain more insight into the human-robot affective interaction. In this paper, we report the development of a multimodal acquisition platform called HIPOP (Human Interaction Pervasive Observation Platform). HIPOP is a modular data-gathering platform based on various hardware and software units that can be easily used to create a custom acquisition setup for HRI studies. The platform uses modules for physiological signals, eye gaze, video and audio acquisition to perform an integrated affective and behavioral analysis. It is also possible to include new hardware devices into the platform. The open-source hardware and software revolution has made many high-quality commercial and open-source products freely available for HRI and HCI research. These devices are currently most often used for data acquisition and robot control, and they can be easily included in HIPOP. Technical tests demonstrated the ability of HIPOP to reliably acquire a large set of data in terms of failure management and data synchronization. The platform was able to automatically recover from errors and faults without affecting the entire system, and the misalignment observed in the acquired data was not significant and did not affect the multimodal analysis. HIPOP was also tested in the context of the FACET (FACE Therapy) project, in which a humanoid robot called FACE (Facial Automaton for Conveying Emotions) was used to convey affective stimuli to children with autism. In the FACET project, psychologists without technical skills were able to use HIPOP to collect the data needed for their experiments without dealing with hardware issues, data integration challenges, or synchronization problems. The FACET case study highlighted the real core feature of the HIPOP platform (i.e., multimodal data integration and fusion). This analytical approach allowed psychologists to study both behavioral and psychophysiological reactions to obtain a more complete view of the subjects' state during interaction with the robot. These results indicate that HIPOP could become an innovative tool for HRI affective studies aimed at inferring a more detailed view of a subject's feelings and behavior during interaction with affective and empathic robots.","PeriodicalId":92076,"journal":{"name":"Journal of human-robot interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.5898/JHRI.3.2.Lazzeri","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71218001","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-07-10DOI: 10.5898/JHRI.3.2.Morales
Y. Morales, T. Kanda, N. Hagita
This paper presents a computational model for side-by-side walking within human-robot interaction (HRI). In this work we address the importance of future motion utility (motion anticipation) of two walking partners. Previous studies only considered a robot moving alongside a person without collisions and with simple velocity-based predictions. In contrast, our proposed model includes two major considerations. First, it considers the current goal, modeling side-by-side walking as a process of moving toward a goal while maintaining a relative position with the partner. Second, it takes the partner's utility into consideration; it models side-by-side walking as a phenomenon where two agents maximize mutual utilities rather than only considering a single agent utility. The model is constructed based in a set of trajectories from pairs of people recorded in side-by-side walking; then, parameters of the model were calibrated for a mobile robot and tested in an autonomous robot walking side-by-side with participants. Finally, two evaluations were performed. The first evaluation shows that the proposed model considering mutual utilities performs better than a single utility method and a method that keeps distance from the walking partner. In the second evaluation the proposed method was used for a robot deployed in a shopping mall environment where it demonstrated to be effective.
{"title":"Walking together","authors":"Y. Morales, T. Kanda, N. Hagita","doi":"10.5898/JHRI.3.2.Morales","DOIUrl":"https://doi.org/10.5898/JHRI.3.2.Morales","url":null,"abstract":"This paper presents a computational model for side-by-side walking within human-robot interaction (HRI). In this work we address the importance of future motion utility (motion anticipation) of two walking partners. Previous studies only considered a robot moving alongside a person without collisions and with simple velocity-based predictions. In contrast, our proposed model includes two major considerations. First, it considers the current goal, modeling side-by-side walking as a process of moving toward a goal while maintaining a relative position with the partner. Second, it takes the partner's utility into consideration; it models side-by-side walking as a phenomenon where two agents maximize mutual utilities rather than only considering a single agent utility. The model is constructed based in a set of trajectories from pairs of people recorded in side-by-side walking; then, parameters of the model were calibrated for a mobile robot and tested in an autonomous robot walking side-by-side with participants. Finally, two evaluations were performed. The first evaluation shows that the proposed model considering mutual utilities performs better than a single utility method and a method that keeps distance from the walking partner. In the second evaluation the proposed method was used for a robot deployed in a shopping mall environment where it demonstrated to be effective.","PeriodicalId":92076,"journal":{"name":"Journal of human-robot interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.5898/JHRI.3.2.Morales","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71218021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Social roles are a design option for robots that behave in accordance with user expectations. We believe that robots have to exceed stereotypical role behaviors and dynamically provide roles that suit the people's living conditions in order to achieve long-term acceptance. We are introducing a new user-focused design method to develop social role repertoires for adaptive human-robot interaction (HRI). The method consists of five sequential steps: (1) user group and application scenario identification; (2) acquisition of users' mental associations; (3) derivation of role traits; (4) prioritization of these traits; and (5) synthesis of an adaptive social role repertoire. We tested our method with two specific user groups: elder adults living at home and those living in care facilities. The results reveal basic role concepts and specific preference clusters in each user group. The empirically based clusters are suitable for the parameterizations and development of robots with adaptive social roles.
{"title":"Designing adaptive roles for socially assistive robots","authors":"Andreas Huber, L. Lammer, A. Weiss, M. Vincze","doi":"10.5898/JHRI.3.2.Huber","DOIUrl":"https://doi.org/10.5898/JHRI.3.2.Huber","url":null,"abstract":"Social roles are a design option for robots that behave in accordance with user expectations. We believe that robots have to exceed stereotypical role behaviors and dynamically provide roles that suit the people's living conditions in order to achieve long-term acceptance. We are introducing a new user-focused design method to develop social role repertoires for adaptive human-robot interaction (HRI). The method consists of five sequential steps: (1) user group and application scenario identification; (2) acquisition of users' mental associations; (3) derivation of role traits; (4) prioritization of these traits; and (5) synthesis of an adaptive social role repertoire. We tested our method with two specific user groups: elder adults living at home and those living in care facilities. The results reveal basic role concepts and specific preference clusters in each user group. The empirically based clusters are suitable for the parameterizations and development of robots with adaptive social roles.","PeriodicalId":92076,"journal":{"name":"Journal of human-robot interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.5898/JHRI.3.2.Huber","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71217944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A critical construct related to human-robot interaction (HRI) is autonomy, which varies widely across robot platforms. Levels of robot autonomy (LORA), ranging from teleoperation to fully autonomous systems, influence the way in which humans and robots may interact with one another. Thus, there is a need to understand HRI by identifying variables that influence - and are influenced by - robot autonomy. Our overarching goal is to develop a framework for levels of robot autonomy in HRI. To reach this goal, the framework draws links between HRI and human-automation interaction, a field with a long history of studying and understanding human-related variables. The construct of autonomy is reviewed and redefined within the context of HRI. Additionally, the framework proposes a process for determining a robot's autonomy level, by categorizing autonomy along a 10-point taxonomy. The framework is intended to be treated as guidelines to determine autonomy, categorize the LORA along a qualitative taxonomy, and consider which HRI variables (e.g., acceptance, situation awareness, reliability) may be influenced by the LORA.
{"title":"Toward a framework for levels of robot autonomy in human-robot interaction.","authors":"Jenay M Beer, Arthur D Fisk, Wendy A Rogers","doi":"10.5898/JHRI.3.2.Beer","DOIUrl":"https://doi.org/10.5898/JHRI.3.2.Beer","url":null,"abstract":"<p><p>A critical construct related to human-robot interaction (HRI) is autonomy, which varies widely across robot platforms. Levels of robot autonomy (LORA), ranging from teleoperation to fully autonomous systems, influence the way in which humans and robots may interact with one another. Thus, there is a need to understand HRI by identifying variables that influence - and are influenced by - robot autonomy. Our overarching goal is to develop a framework for levels of robot autonomy in HRI. To reach this goal, the framework draws links between HRI and human-automation interaction, a field with a long history of studying and understanding human-related variables. The construct of autonomy is reviewed and redefined within the context of HRI. Additionally, the framework proposes a process for determining a robot's autonomy level, by categorizing autonomy along a 10-point taxonomy. The framework is intended to be treated as guidelines to determine autonomy, categorize the LORA along a qualitative taxonomy, and consider which HRI variables (e.g., acceptance, situation awareness, reliability) may be influenced by the LORA.</p>","PeriodicalId":92076,"journal":{"name":"Journal of human-robot interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.5898/JHRI.3.2.Beer","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"35503497","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This article begins by asking: "Why are robots not becoming domestic products?" In addressing this question the author borrows from the science of ecology and biological concepts of evolution and domestication to make an analogy between the shift of habitats that occurs when an organism successfully goes through the process of artificial selection (from natural to domestic). In addition, this paper explores the transition an emerging technology makes when coming out of the laboratory and becoming a suitable product for domestic use, concluding that the majority of proposed domestic robots are essentially maladapted to everyday life. The article then shifts the focus onto design research, primarily speculative design, to ask, "how could robots become domestic products?" The author uses a variety of design projects to describe how alternative approaches to robots can provide new perspectives on technological research and development.
{"title":"Living with robots","authors":"J. Auger","doi":"10.5898/JHRI.3.1.Auger","DOIUrl":"https://doi.org/10.5898/JHRI.3.1.Auger","url":null,"abstract":"This article begins by asking: \"Why are robots not becoming domestic products?\" In addressing this question the author borrows from the science of ecology and biological concepts of evolution and domestication to make an analogy between the shift of habitats that occurs when an organism successfully goes through the process of artificial selection (from natural to domestic). In addition, this paper explores the transition an emerging technology makes when coming out of the laboratory and becoming a suitable product for domestic use, concluding that the majority of proposed domestic robots are essentially maladapted to everyday life. The article then shifts the focus onto design research, primarily speculative design, to ask, \"how could robots become domestic products?\" The author uses a variety of design projects to describe how alternative approaches to robots can provide new perspectives on technological research and development.","PeriodicalId":92076,"journal":{"name":"Journal of human-robot interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.5898/JHRI.3.1.Auger","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71218224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Humans have had a long history of fascination with building intelligent machines that depict themselves or move animatedly. This article explores the history of robotic creatures like Egyptian figurines, Greek mechanical inventions, 18th century ingenious automata, and modern kinetic mise-en-scène and robotic artworks. Several interactive robots showed the potential for emotional interaction between humans and machines. In the context of human-robot interaction, empathy with robots requires further discussions on the interaction design.
{"title":"From mechanical metamorphosis to empathic interaction","authors":"Chang Geun Oh, Jaeheung Park","doi":"10.5898/JHRI.3.1.Oh","DOIUrl":"https://doi.org/10.5898/JHRI.3.1.Oh","url":null,"abstract":"Humans have had a long history of fascination with building intelligent machines that depict themselves or move animatedly. This article explores the history of robotic creatures like Egyptian figurines, Greek mechanical inventions, 18th century ingenious automata, and modern kinetic mise-en-scène and robotic artworks. Several interactive robots showed the potential for emotional interaction between humans and machines. In the context of human-robot interaction, empathy with robots requires further discussions on the interaction design.","PeriodicalId":92076,"journal":{"name":"Journal of human-robot interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.5898/JHRI.3.1.Oh","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71217860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-02-28DOI: 10.5898/JHRI.3.1.Sabanovic
S. Šabanović, S. Reeder, Bobak Kechavarzi
As robots move into everyday environments, we need to understand both the social and the technical constraints and affordances for human-robot interaction. We use in situ evaluation of partially functioning prototypes to inform the design of robotic technologies that fit their intended contexts of use and illustrate this method through a case study of iteratively designing a desktop robot for break management in a computerized office. After an initial exploratory study of the office as context of use, we used comparative semi-controlled evaluations of multiple design alternatives to explore how different robot characteristics, specifically embodiment and social interactivity, are perceived by users and affect their break taking. We found evaluating simple prototypes with varying levels of functionality, even when not robust or "complete," provides opportunities for including users in the design process and for identifying emergent factors that impact robot use. Our case study provides insights into the challenges and best practices for performing iterative prototyping and in situ evaluations of robots, which can inform future development of contextually appropriate robotic technologies.
{"title":"Designing robots in the wild","authors":"S. Šabanović, S. Reeder, Bobak Kechavarzi","doi":"10.5898/JHRI.3.1.Sabanovic","DOIUrl":"https://doi.org/10.5898/JHRI.3.1.Sabanovic","url":null,"abstract":"As robots move into everyday environments, we need to understand both the social and the technical constraints and affordances for human-robot interaction. We use in situ evaluation of partially functioning prototypes to inform the design of robotic technologies that fit their intended contexts of use and illustrate this method through a case study of iteratively designing a desktop robot for break management in a computerized office. After an initial exploratory study of the office as context of use, we used comparative semi-controlled evaluations of multiple design alternatives to explore how different robot characteristics, specifically embodiment and social interactivity, are perceived by users and affect their break taking. We found evaluating simple prototypes with varying levels of functionality, even when not robust or \"complete,\" provides opportunities for including users in the design process and for identifying emergent factors that impact robot use. Our case study provides insights into the challenges and best practices for performing iterative prototyping and in situ evaluations of robots, which can inform future development of contextually appropriate robotic technologies.","PeriodicalId":92076,"journal":{"name":"Journal of human-robot interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.5898/JHRI.3.1.Sabanovic","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71217919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-02-28DOI: 10.5898/JHRI.3.1.Johnson
Matthew Johnson, J. Bradshaw, P. Feltovich, C. Jonker, M. van Riemsdijk, M. Sierhuis
Coactive Design is a new approach to address the increasingly sophisticated roles that people and robots play as the use of robots expands into new, complex domains. The approach is motivated by the desire for robots to perform less like teleoperated tools or independent automatons and more like interdependent teammates. In this article, we describe what it means to be interdependent, why this is important, and the design implications that follow from this perspective. We argue for a human-robot system model that supports interdependence through careful attention to requirements for observability, predictability, and directability. We present a Coactive Design method and show how it can be a useful approach for developers trying to understand how to translate high-level teamwork concepts into reusable control algorithms, interface elements, and behaviors that enable robots to fulfill their envisioned role as teammates. As an example of the coactive design approach, we present our results from the DARPA Virtual Robotics Challenge, a competition designed to spur development of advanced robots that can assist humans in recovering from natural and man-made disasters. Twenty-six teams from eight countries competed in three different tasks providing an excellent evaluation of the relative effectiveness of different approaches to human-machine system design.
{"title":"Coactive design","authors":"Matthew Johnson, J. Bradshaw, P. Feltovich, C. Jonker, M. van Riemsdijk, M. Sierhuis","doi":"10.5898/JHRI.3.1.Johnson","DOIUrl":"https://doi.org/10.5898/JHRI.3.1.Johnson","url":null,"abstract":"Coactive Design is a new approach to address the increasingly sophisticated roles that people and robots play as the use of robots expands into new, complex domains. The approach is motivated by the desire for robots to perform less like teleoperated tools or independent automatons and more like interdependent teammates. In this article, we describe what it means to be interdependent, why this is important, and the design implications that follow from this perspective. We argue for a human-robot system model that supports interdependence through careful attention to requirements for observability, predictability, and directability. We present a Coactive Design method and show how it can be a useful approach for developers trying to understand how to translate high-level teamwork concepts into reusable control algorithms, interface elements, and behaviors that enable robots to fulfill their envisioned role as teammates. As an example of the coactive design approach, we present our results from the DARPA Virtual Robotics Challenge, a competition designed to spur development of advanced robots that can assist humans in recovering from natural and man-made disasters. Twenty-six teams from eight countries competed in three different tasks providing an excellent evaluation of the relative effectiveness of different approaches to human-machine system design.","PeriodicalId":92076,"journal":{"name":"Journal of human-robot interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.5898/JHRI.3.1.Johnson","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71218291","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-02-28DOI: 10.5898/JHRI.3.1.Hoffman
Guy Hoffman, Wendy Ju
This paper makes the case for designing interactive robots with their expressive movement in mind. As people are highly sensitive to physical movement and spatiotemporal affordances, well-designed robot motion can communicate, engage, and offer dynamic possibilities beyond the machines' surface appearance or pragmatic motion paths. We present techniques for movement centric design, including character animation sketches, video prototyping, interactive movement explorations, Wizard of Oz studies, and skeletal prototypes. To illustrate our design approach, we discuss four case studies: a social head for a robotic musician, a robotic speaker dock listening companion, a desktop telepresence robot, and a service robot performing assistive and communicative tasks. We then relate our approach to the design of non-anthropomorphic robots and robotic objects, a design strategy that could facilitate the feasibility of real-world human-robot interaction.
{"title":"Designing robots with movement in mind","authors":"Guy Hoffman, Wendy Ju","doi":"10.5898/JHRI.3.1.Hoffman","DOIUrl":"https://doi.org/10.5898/JHRI.3.1.Hoffman","url":null,"abstract":"This paper makes the case for designing interactive robots with their expressive movement in mind. As people are highly sensitive to physical movement and spatiotemporal affordances, well-designed robot motion can communicate, engage, and offer dynamic possibilities beyond the machines' surface appearance or pragmatic motion paths. We present techniques for movement centric design, including character animation sketches, video prototyping, interactive movement explorations, Wizard of Oz studies, and skeletal prototypes. To illustrate our design approach, we discuss four case studies: a social head for a robotic musician, a robotic speaker dock listening companion, a desktop telepresence robot, and a service robot performing assistive and communicative tasks. We then relate our approach to the design of non-anthropomorphic robots and robotic objects, a design strategy that could facilitate the feasibility of real-world human-robot interaction.","PeriodicalId":92076,"journal":{"name":"Journal of human-robot interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.5898/JHRI.3.1.Hoffman","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71218272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-02-28DOI: 10.5898/JHRI.3.1.Scherer
D. Scherer
This essay provides a perspective on the ongoing convergence of social robots and special effects, animation, animatronics, puppetry techniques, and other entertainment technologies. In this paper, I will address the following design concepts: - Character robot design principles and concepts - Fully realized Character robots as Slaves, as a term and social artifact - The implementation of particular special effect techniques I abstract robots into six categories and focus the discussion on one of these: The Companion type. I suggest that this type of robot might benefit from being what I call a Character robot, achieving this status by possessing the following four traits: 1) Biological in its appearance 2) Clear and simple in its expression 3) Well-animated 4) Recognizable in its personality
{"title":"Movie magic makes better social robots","authors":"D. Scherer","doi":"10.5898/JHRI.3.1.Scherer","DOIUrl":"https://doi.org/10.5898/JHRI.3.1.Scherer","url":null,"abstract":"This essay provides a perspective on the ongoing convergence of social robots and special effects, animation, animatronics, puppetry techniques, and other entertainment technologies. In this paper, I will address the following design concepts: - Character robot design principles and concepts - Fully realized Character robots as Slaves, as a term and social artifact - The implementation of particular special effect techniques I abstract robots into six categories and focus the discussion on one of these: The Companion type. I suggest that this type of robot might benefit from being what I call a Character robot, achieving this status by possessing the following four traits: 1) Biological in its appearance 2) Clear and simple in its expression 3) Well-animated 4) Recognizable in its personality","PeriodicalId":92076,"journal":{"name":"Journal of human-robot interaction","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.5898/JHRI.3.1.Scherer","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71217930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}