首页 > 最新文献

2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)最新文献

英文 中文
A robot to study the development of artwork appreciation through social interactions 一个机器人通过社会互动来研究艺术品欣赏的发展
A. Karaouzene, P. Gaussier, Denis Vidal
In this work, we present a model based on social referencing mechanisms for the artwork appreciation development. We show that trying to reach autonomous robot artwork preferences is an interesting framework for developmental robotics. Therefore, we present an application of our model in a natural environment. The museum as an experiment place, helps us to benefit from a large number of visitors with different backgrounds and personal interests, as well as provides the challenge of the learning and recognition of an important number of artefacts (heterogeneous set of objects in terms of size, culture). To overcome some limitations like the autonomy in the learning, we proposed a first model for cumulative learning using a second-order conditioning approach with very encouraging results.
在这项工作中,我们提出了一个基于社会参考机制的艺术品欣赏发展模型。我们表明,试图达到自主机器人艺术偏好是一个有趣的发展机器人框架。因此,我们提出了该模型在自然环境中的应用。博物馆作为一个实验场所,帮助我们从大量具有不同背景和个人兴趣的游客中受益,并提供了学习和识别重要数量的人工制品(在大小,文化方面异构的物体集合)的挑战。为了克服学习自主性等局限性,我们提出了第一个使用二阶条件反射方法的累积学习模型,并取得了非常令人鼓舞的结果。
{"title":"A robot to study the development of artwork appreciation through social interactions","authors":"A. Karaouzene, P. Gaussier, Denis Vidal","doi":"10.1109/DEVLRN.2013.6652554","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652554","url":null,"abstract":"In this work, we present a model based on social referencing mechanisms for the artwork appreciation development. We show that trying to reach autonomous robot artwork preferences is an interesting framework for developmental robotics. Therefore, we present an application of our model in a natural environment. The museum as an experiment place, helps us to benefit from a large number of visitors with different backgrounds and personal interests, as well as provides the challenge of the learning and recognition of an important number of artefacts (heterogeneous set of objects in terms of size, culture). To overcome some limitations like the autonomy in the learning, we proposed a first model for cumulative learning using a second-order conditioning approach with very encouraging results.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129416301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Learning stable pushing locations 学习稳定的推动位置
Tucker Hermans, Fuxin Li, James M. Rehg, A. Bobick
We present a method by which a robot learns to predict effective push-locations as a function of object shape. The robot performs push experiments at many contact locations on multiple objects and records local and global shape features at each point of contact. The robot observes the outcome trajectories of the manipulations and computes a novel push-stability score for each trial. The robot then learns a regression function in order to predict push effectiveness as a function of object shape. This mapping allows the robot to select effective push locations for subsequent objects whether they are previously manipulated instances, new instances from previously encountered object classes, or entirely novel objects. In the totally novel object case, the local shape property coupled with the overall distribution of the object allows for the discovery of effective push locations. These results are demonstrated on a mobile manipulator robot pushing a variety of household objects on a tabletop surface.
我们提出了一种方法,通过机器人学习预测有效的推位置作为一个函数的对象形状。机器人在多个物体的多个接触点进行推入实验,并记录每个接触点的局部和全局形状特征。机器人观察操作的结果轨迹,并为每次试验计算一个新的推稳定性评分。然后,机器人学习一个回归函数,以预测推送效果作为物体形状的函数。这种映射允许机器人为后续对象选择有效的推送位置,无论它们是以前操作过的实例、以前遇到的对象类的新实例,还是完全新的对象。在完全新对象的情况下,局部形状属性与对象的整体分布相结合,允许发现有效的推送位置。这些结果在移动机械手机器人上进行了演示,机器人在桌面表面上推动各种家居物品。
{"title":"Learning stable pushing locations","authors":"Tucker Hermans, Fuxin Li, James M. Rehg, A. Bobick","doi":"10.1109/DEVLRN.2013.6652539","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652539","url":null,"abstract":"We present a method by which a robot learns to predict effective push-locations as a function of object shape. The robot performs push experiments at many contact locations on multiple objects and records local and global shape features at each point of contact. The robot observes the outcome trajectories of the manipulations and computes a novel push-stability score for each trial. The robot then learns a regression function in order to predict push effectiveness as a function of object shape. This mapping allows the robot to select effective push locations for subsequent objects whether they are previously manipulated instances, new instances from previously encountered object classes, or entirely novel objects. In the totally novel object case, the local shape property coupled with the overall distribution of the object allows for the discovery of effective push locations. These results are demonstrated on a mobile manipulator robot pushing a variety of household objects on a tabletop surface.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121859117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Goal babbling with unknown ranges: A direction-sampling approach 未知范围的目标胡言乱语:一种方向采样方法
Matthias Rolf
Goal babbling is a recent concept for the efficient bootstrapping of sensorimotor coordination that is inspired by infants' early goal-directed movement attempts. Several studies have shown its superior performance compared to random motor babbling. Yet, previous implementations of goal babbling require knowledge of a set of achievable goals in advance. This paper introduces an approach to goal babbling that can bootstrap coordination skills without pre-specifying, or even representing, a set of goals. On the contrary, it can discover the ranges of achievable goals autonomously. This capability is demonstrated in a challenging task with up to 50 degrees of freedom, in which the discovery of possible outcomes is shown to be desperately intractable with random motor babbling.
目标咿呀学语是最近提出的一个概念,旨在有效地引导感觉运动协调,其灵感来自婴儿早期目标导向的运动尝试。几项研究表明,与随机运动牙牙学语相比,其性能优越。然而,之前的目标喋喋不休的实现需要事先了解一组可实现的目标。本文介绍了一种目标喋喋不休的方法,它可以在不预先指定甚至不代表一组目标的情况下引导协调技能。相反,它可以自主地发现可实现目标的范围。这种能力在一个具有高达50个自由度的挑战性任务中得到了证明,在这个任务中,发现可能的结果被证明是极其棘手的,而且是随机的运动牙牙学语。
{"title":"Goal babbling with unknown ranges: A direction-sampling approach","authors":"Matthias Rolf","doi":"10.1109/DEVLRN.2013.6652526","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652526","url":null,"abstract":"Goal babbling is a recent concept for the efficient bootstrapping of sensorimotor coordination that is inspired by infants' early goal-directed movement attempts. Several studies have shown its superior performance compared to random motor babbling. Yet, previous implementations of goal babbling require knowledge of a set of achievable goals in advance. This paper introduces an approach to goal babbling that can bootstrap coordination skills without pre-specifying, or even representing, a set of goals. On the contrary, it can discover the ranges of achievable goals autonomously. This capability is demonstrated in a challenging task with up to 50 degrees of freedom, in which the discovery of possible outcomes is shown to be desperately intractable with random motor babbling.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"692 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126341665","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Reinforcement learning with state-dependent discount factor 状态相关折现因子的强化学习
N. Yoshida, E. Uchibe, K. Doya
Conventional reinforcement learning algorithms have several parameters which determine the feature of learning process, called meta-parameters. In this study, we focus on the discount factor that influences the time scale of the tradeoff between immediate and delayed rewards. The discount factor is usually considered as a constant value, but we introduce the state-dependent discount function and a new optimization criterion for the reinforcement learning algorithm. We first derive a new algorithm under the criterion, named ExQ-learning and we prove that the algorithm converges to the optimal action-value function in the meaning of new criterion w.p.1. We then present a framework to optimize the discount factor and the discount function by using an evolutionary algorithm. In order to validate the proposed method, we conduct a simple computer simulation and show that the proposed algorithm can find an appropriate state-dependent discount function with which performs better than that with a constant discount factor.
传统的强化学习算法有几个参数,这些参数决定了学习过程的特征,称为元参数。在本研究中,我们关注影响即时和延迟奖励权衡的时间尺度的折扣因子。折扣因子通常被认为是一个常数,但我们引入了状态相关的折扣函数和一个新的强化学习算法优化准则。我们首先在该准则下推导了一个新的算法ExQ-learning,并证明了该算法收敛于新准则w.p.1意义下的最优动作值函数。然后,我们提出了一个利用进化算法优化折现因子和折现函数的框架。为了验证所提出的方法,我们进行了简单的计算机模拟,并表明所提出的算法可以找到合适的状态相关折扣函数,该函数比使用恒定折扣因子的算法性能更好。
{"title":"Reinforcement learning with state-dependent discount factor","authors":"N. Yoshida, E. Uchibe, K. Doya","doi":"10.1109/DEVLRN.2013.6652533","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652533","url":null,"abstract":"Conventional reinforcement learning algorithms have several parameters which determine the feature of learning process, called meta-parameters. In this study, we focus on the discount factor that influences the time scale of the tradeoff between immediate and delayed rewards. The discount factor is usually considered as a constant value, but we introduce the state-dependent discount function and a new optimization criterion for the reinforcement learning algorithm. We first derive a new algorithm under the criterion, named ExQ-learning and we prove that the algorithm converges to the optimal action-value function in the meaning of new criterion w.p.1. We then present a framework to optimize the discount factor and the discount function by using an evolutionary algorithm. In order to validate the proposed method, we conduct a simple computer simulation and show that the proposed algorithm can find an appropriate state-dependent discount function with which performs better than that with a constant discount factor.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126497515","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Do beliefs about a robot's capabilities influence alignment to its actions? 对机器人能力的信念会影响其行动的一致性吗?
Anna-Lisa Vollmer, B. Wrede, K. Rohlfing, A. Cangelosi
Interlocutors in a dialog align on many aspects of behavior (word choice, speech rate, syntactic structure, gestures, facial expressions, etc.). Such alignment has been proposed to be the basis for communicating successfully. We believe alignment could be beneficial for smooth human-robot interaction and facilitate robot action learning from demonstration. Recent research put forward a mediated communicative design account of alignment according to which interlocutors align stronger when they believe it will lead to communicative success. Branigan et al. showed that when interacting with an artificial system, participants aligned their lexical choices more to an artificial system they believed to be basic than to one they believed to be advanced. Our work extends these results in two ways: First, instead of an artificial computer dialog system, participants interact with a humanoid robot, the iCub robot. Second, instead of lexical choice, our work investigates alignment in the domain of manual actions. In an action demonstration and matching game, we examine the extent to which participants who believe that they are playing with a basic version or an advanced version of the iCub robot adapt the way they execute actions to what their robot partner has previously shown to them. Our results confirm that alignment also takes place in action demonstration. We were not able to replicate Branigan et al.'s results in general in this setup, but in line with their findings, participants with a low questionnaire score on neuroticism and participants who are familiar with robots aligned their actions more to a robot they believed to be basic than to one they believed to be advanced.
对话中的对话者在行为的许多方面都是一致的(用词、语速、句法结构、手势、面部表情等)。这种一致性被认为是成功沟通的基础。我们相信对齐可以使人机交互更加顺畅,并促进机器人从演示中学习动作。最近的研究提出了一种中介交际设计理论,根据该理论,当对话者认为结盟会导致交际成功时,他们的结盟会更强。布莱尼根等人的研究表明,当参与者与人工系统互动时,他们的词汇选择更倾向于他们认为是基本的人工系统,而不是他们认为是高级的人工系统。我们的工作以两种方式扩展了这些结果:首先,参与者与人形机器人iCub机器人进行交互,而不是人工计算机对话系统。其次,我们研究的不是词汇选择,而是手动操作领域的对齐。在动作演示和匹配游戏中,我们检查了认为自己正在玩iCub机器人的基本版本或高级版本的参与者在多大程度上适应他们执行动作的方式,以适应他们的机器人伙伴之前向他们展示的内容。我们的结果证实了在行动演示中也会发生对齐。我们无法在这个设置中复制Branigan等人的结果,但与他们的发现一致,神经质问卷得分低的参与者和熟悉机器人的参与者更倾向于他们认为基本的机器人,而不是他们认为先进的机器人。
{"title":"Do beliefs about a robot's capabilities influence alignment to its actions?","authors":"Anna-Lisa Vollmer, B. Wrede, K. Rohlfing, A. Cangelosi","doi":"10.1109/DEVLRN.2013.6652521","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652521","url":null,"abstract":"Interlocutors in a dialog align on many aspects of behavior (word choice, speech rate, syntactic structure, gestures, facial expressions, etc.). Such alignment has been proposed to be the basis for communicating successfully. We believe alignment could be beneficial for smooth human-robot interaction and facilitate robot action learning from demonstration. Recent research put forward a mediated communicative design account of alignment according to which interlocutors align stronger when they believe it will lead to communicative success. Branigan et al. showed that when interacting with an artificial system, participants aligned their lexical choices more to an artificial system they believed to be basic than to one they believed to be advanced. Our work extends these results in two ways: First, instead of an artificial computer dialog system, participants interact with a humanoid robot, the iCub robot. Second, instead of lexical choice, our work investigates alignment in the domain of manual actions. In an action demonstration and matching game, we examine the extent to which participants who believe that they are playing with a basic version or an advanced version of the iCub robot adapt the way they execute actions to what their robot partner has previously shown to them. Our results confirm that alignment also takes place in action demonstration. We were not able to replicate Branigan et al.'s results in general in this setup, but in line with their findings, participants with a low questionnaire score on neuroticism and participants who are familiar with robots aligned their actions more to a robot they believed to be basic than to one they believed to be advanced.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123990138","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Autonomous learning of active multi-scale binocular vision 主动多尺度双目视觉的自主学习
L. Lonini, Yu Zhao, Pramod Chandrashekhariah, Bertram E. Shi, J. Triesch
We present a method for autonomously learning representations of visual disparity between images from left and right eye, as well as appropriate vergence movements to fixate objects with both eyes. A sparse coding model (perception) encodes sensory information using binocular basis functions, while a reinforcement learner (behavior) generates the eye movement, according to the sensed disparity. Perception and behavior develop in parallel, by minimizing the same cost function: the reconstruction error of the stimulus by the generative model. In order to efficiently cope with multiple disparity ranges, sparse coding models are learnt at multiple scales, encoding disparities at various resolutions. Similarly, vergence commands are defined on a logarithmic scale to allow both coarse and fine actions. We demonstrate the efficacy of the proposed method using the humanoid robot iCub. We show that the model is fully self-calibrating and does not require any prior information about the camera parameters or the system dynamics.
我们提出了一种自主学习左眼和右眼图像视觉差异表征的方法,以及两只眼睛注视物体的适当收敛运动。稀疏编码模型(感知)利用双眼基函数对感觉信息进行编码,而强化学习模型(行为)根据感知到的视差产生眼动。通过最小化相同的成本函数,即生成模型对刺激的重建误差,感知和行为是并行发展的。为了有效地处理多个视差范围,在多个尺度上学习稀疏编码模型,对不同分辨率的视差进行编码。类似地,收敛命令在对数尺度上定义,以允许粗操作和细操作。我们用仿人机器人iCub验证了该方法的有效性。我们证明了该模型是完全自校准的,并且不需要任何关于相机参数或系统动力学的先验信息。
{"title":"Autonomous learning of active multi-scale binocular vision","authors":"L. Lonini, Yu Zhao, Pramod Chandrashekhariah, Bertram E. Shi, J. Triesch","doi":"10.1109/DEVLRN.2013.6652541","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652541","url":null,"abstract":"We present a method for autonomously learning representations of visual disparity between images from left and right eye, as well as appropriate vergence movements to fixate objects with both eyes. A sparse coding model (perception) encodes sensory information using binocular basis functions, while a reinforcement learner (behavior) generates the eye movement, according to the sensed disparity. Perception and behavior develop in parallel, by minimizing the same cost function: the reconstruction error of the stimulus by the generative model. In order to efficiently cope with multiple disparity ranges, sparse coding models are learnt at multiple scales, encoding disparities at various resolutions. Similarly, vergence commands are defined on a logarithmic scale to allow both coarse and fine actions. We demonstrate the efficacy of the proposed method using the humanoid robot iCub. We show that the model is fully self-calibrating and does not require any prior information about the camera parameters or the system dynamics.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129155063","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Towards understanding the origin of infant directed speech: A vocal robot with infant-like articulation 了解婴儿定向言语的起源:一个具有类似婴儿发音的发声机器人
Yuki Sasamoto, Naoto Nishijima, M. Asada
Infant-Directed Speech (IDS) is the non-standard form of caregivers' speech to their infants. Developmental studies indicate that IDS changes from infant-directed to adult-directed depending on infant's age and/or linguistic level. However, it is still unclear what features in infants cause IDS. This article introduces a vocal robot with an infant-like articulatory system to attack the issue by means of a constructive approach. A preliminary experiment implies that our robot can vocalize structurally similar to infant articulation although it is mechanically rather different.
婴儿指向语是照顾者对婴儿的非标准语言形式。发展研究表明,根据婴儿的年龄和/或语言水平,IDS从婴儿导向转变为成人导向。然而,目前尚不清楚婴儿的哪些特征会导致IDS。本文介绍了一种具有类似婴儿发音系统的发声机器人,通过一种建设性的方法来解决这个问题。一项初步实验表明,我们的机器人可以发出类似于婴儿发音的结构,尽管它在机械上有很大的不同。
{"title":"Towards understanding the origin of infant directed speech: A vocal robot with infant-like articulation","authors":"Yuki Sasamoto, Naoto Nishijima, M. Asada","doi":"10.1109/DEVLRN.2013.6652562","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652562","url":null,"abstract":"Infant-Directed Speech (IDS) is the non-standard form of caregivers' speech to their infants. Developmental studies indicate that IDS changes from infant-directed to adult-directed depending on infant's age and/or linguistic level. However, it is still unclear what features in infants cause IDS. This article introduces a vocal robot with an infant-like articulatory system to attack the issue by means of a constructive approach. A preliminary experiment implies that our robot can vocalize structurally similar to infant articulation although it is mechanically rather different.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127029573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Structural bootstrapping at the sensorimotor level for the fast acquisition of action knowledge for cognitive robots 认知机器人快速获取动作知识的感觉运动水平结构自举
E. Aksoy, M. Tamosiunaite, Rok Vuga, A. Ude, C. Geib, Mark Steedman, F. Wörgötter
Autonomous robots are faced with the problem of encoding complex actions (e.g. complete manipulations) in a generic and generalizable way. Recently we had introduced the Semantic Event Chains (SECs) as a new representation which can be directly computed from a stream of 3D images and is based on changes in the relationships between objects involved in a manipulation. Here we show that the SEC framework can be extended (called “extended SEC”) with action-related information and used to achieve and encode two important cognitive properties relevant for advanced autonomous robots: The extended SEC enables us to determine whether an action representation (1) needs to be newly created and stored in its entirety in the robot's memory or (2) whether one of the already known and memorized action representations just needs to be refined. In human cognition these two processes (1 and 2) are known as accommodation and assimilation. Thus, here we show that the extended SEC representation can be used to realize these processes originally defined by Piaget for the first time in a robotic application. This is of fundamental importance for any cognitive agent as it allows categorizing observed actions in new versus known ones, storing only the relevant aspects.
自主机器人面临着用通用和可推广的方式对复杂动作(如完整操作)进行编码的问题。最近,我们引入了语义事件链(SECs)作为一种新的表示,它可以直接从3D图像流中计算出来,并且基于操作中涉及的对象之间关系的变化。在这里,我们展示了SEC框架可以用与动作相关的信息进行扩展(称为“扩展SEC”),并用于实现和编码与高级自主机器人相关的两个重要认知属性:扩展SEC使我们能够确定一个动作表示(1)是否需要新创建并完整地存储在机器人的记忆中,或者(2)是否一个已知和记忆的动作表示只需要改进。在人类认知中,这两个过程(1和2)被称为适应和同化。因此,我们在这里展示了扩展的SEC表示可以用于实现最初由Piaget在机器人应用中定义的这些过程。这对于任何认知代理都是至关重要的,因为它允许将观察到的行为分为新的和已知的,只存储相关的方面。
{"title":"Structural bootstrapping at the sensorimotor level for the fast acquisition of action knowledge for cognitive robots","authors":"E. Aksoy, M. Tamosiunaite, Rok Vuga, A. Ude, C. Geib, Mark Steedman, F. Wörgötter","doi":"10.1109/DEVLRN.2013.6652537","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652537","url":null,"abstract":"Autonomous robots are faced with the problem of encoding complex actions (e.g. complete manipulations) in a generic and generalizable way. Recently we had introduced the Semantic Event Chains (SECs) as a new representation which can be directly computed from a stream of 3D images and is based on changes in the relationships between objects involved in a manipulation. Here we show that the SEC framework can be extended (called “extended SEC”) with action-related information and used to achieve and encode two important cognitive properties relevant for advanced autonomous robots: The extended SEC enables us to determine whether an action representation (1) needs to be newly created and stored in its entirety in the robot's memory or (2) whether one of the already known and memorized action representations just needs to be refined. In human cognition these two processes (1 and 2) are known as accommodation and assimilation. Thus, here we show that the extended SEC representation can be used to realize these processes originally defined by Piaget for the first time in a robotic application. This is of fundamental importance for any cognitive agent as it allows categorizing observed actions in new versus known ones, storing only the relevant aspects.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122746347","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Temporal emphasis for goal extraction in task demonstration to a humanoid robot by naive users 幼稚用户对仿人机器人任务演示中目标提取的时间重点
Konstantinos Theofilis, K. Lohan, Chrystopher L. Nehaniv, K. Dautenhahn, B. Wrede
Goal extraction in learning by demonstration is a complex problem. A novel approach, inspired by developmental psychology and focused on use in experiments with naive users, is presented in this paper. Participants were presenting a simple task, how to stack three boxes, to the humanoid robot iCub. The stationary states of the task - 1 box, 2 boxes stacked, 3 boxes stacked - were defined and the time span of each state was measured. Analysis of the results showed that there is a significant result that users tend to keep the boxes stationary longer upon completion of the end goal than upon completion of the sub-goals. A simple and straightforward learning algorithm was then used on the demonstration data, using only the time spans of the stationary states. The learning algorithm successfully detected the end goal. These temporal differences, functioning as emphasis, could be used as a complementary mechanism for goal extraction in imitation learning. Furthermore, it is suggested that since a simple, straightforward learning algorithm can use these pauses to recognise the goal state, humans may also be able to use this pause as a complementary mechanism for recognising the goal state of a task.
示范学习中的目标提取是一个复杂的问题。本文提出了一种新的方法,受发展心理学的启发,并专注于在天真用户的实验中使用。参与者向人形机器人iCub展示了一个简单的任务,即如何堆叠三个盒子。定义了1箱、2箱堆叠、3箱堆叠的任务静态状态,并测量了每个状态的时间跨度。对结果的分析表明,有一个显著的结果,即用户在完成最终目标时比在完成子目标时更倾向于保持盒子静止更长时间。然后对演示数据使用了一种简单直接的学习算法,仅使用稳态的时间跨度。学习算法成功检测到最终目标。这些时间差异作为重点,可以作为模仿学习中目标提取的补充机制。此外,有人建议,由于简单,直接的学习算法可以使用这些暂停来识别目标状态,人类也可以使用这种暂停作为识别任务目标状态的补充机制。
{"title":"Temporal emphasis for goal extraction in task demonstration to a humanoid robot by naive users","authors":"Konstantinos Theofilis, K. Lohan, Chrystopher L. Nehaniv, K. Dautenhahn, B. Wrede","doi":"10.1109/DEVLRN.2013.6652536","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652536","url":null,"abstract":"Goal extraction in learning by demonstration is a complex problem. A novel approach, inspired by developmental psychology and focused on use in experiments with naive users, is presented in this paper. Participants were presenting a simple task, how to stack three boxes, to the humanoid robot iCub. The stationary states of the task - 1 box, 2 boxes stacked, 3 boxes stacked - were defined and the time span of each state was measured. Analysis of the results showed that there is a significant result that users tend to keep the boxes stationary longer upon completion of the end goal than upon completion of the sub-goals. A simple and straightforward learning algorithm was then used on the demonstration data, using only the time spans of the stationary states. The learning algorithm successfully detected the end goal. These temporal differences, functioning as emphasis, could be used as a complementary mechanism for goal extraction in imitation learning. Furthermore, it is suggested that since a simple, straightforward learning algorithm can use these pauses to recognise the goal state, humans may also be able to use this pause as a complementary mechanism for recognising the goal state of a task.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115062060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Adaptive reachability assessment in the humanoid robot iCub 仿人机器人iCub的自适应可达性评估
Salomón Ramírez-Contla, D. Marocco
We present a model for reachability assessment implemented in a simulated iCub humanoid robot. The robot uses a neural network both for estimating reachability and as a controller for the arm. During training, multi-modality information including vision and proprioception of the effector's length was provided, along with tactile and postural information. The task was to assess if a target in view was at reach range. After training with data from two different effector's lengths, the system generalised also for a third one, both for producing reaching postures and for assessing reachability. We present preliminary results that show good reachability predictions with a decrease in confidence that display a depth gradient.
我们提出了一个模拟iCub人形机器人可达性评估模型。机器人使用神经网络来估计可达性,并作为手臂的控制器。在训练过程中,提供多模态信息,包括视觉和效应器长度的本体感觉,以及触觉和姿势信息。任务是评估目标是否在可及范围内。在用来自两个不同效应器长度的数据进行训练后,该系统也对第三个效应器进行了一般化,包括产生伸展姿势和评估可达性。我们提出了初步结果,显示出良好的可达性预测,但显示深度梯度的置信度降低。
{"title":"Adaptive reachability assessment in the humanoid robot iCub","authors":"Salomón Ramírez-Contla, D. Marocco","doi":"10.1109/DEVLRN.2013.6652546","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652546","url":null,"abstract":"We present a model for reachability assessment implemented in a simulated iCub humanoid robot. The robot uses a neural network both for estimating reachability and as a controller for the arm. During training, multi-modality information including vision and proprioception of the effector's length was provided, along with tactile and postural information. The task was to assess if a target in view was at reach range. After training with data from two different effector's lengths, the system generalised also for a third one, both for producing reaching postures and for assessing reachability. We present preliminary results that show good reachability predictions with a decrease in confidence that display a depth gradient.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132641434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1