首页 > 最新文献

2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)最新文献

英文 中文
Learning stable pushing locations 学习稳定的推动位置
Tucker Hermans, Fuxin Li, James M. Rehg, A. Bobick
We present a method by which a robot learns to predict effective push-locations as a function of object shape. The robot performs push experiments at many contact locations on multiple objects and records local and global shape features at each point of contact. The robot observes the outcome trajectories of the manipulations and computes a novel push-stability score for each trial. The robot then learns a regression function in order to predict push effectiveness as a function of object shape. This mapping allows the robot to select effective push locations for subsequent objects whether they are previously manipulated instances, new instances from previously encountered object classes, or entirely novel objects. In the totally novel object case, the local shape property coupled with the overall distribution of the object allows for the discovery of effective push locations. These results are demonstrated on a mobile manipulator robot pushing a variety of household objects on a tabletop surface.
我们提出了一种方法,通过机器人学习预测有效的推位置作为一个函数的对象形状。机器人在多个物体的多个接触点进行推入实验,并记录每个接触点的局部和全局形状特征。机器人观察操作的结果轨迹,并为每次试验计算一个新的推稳定性评分。然后,机器人学习一个回归函数,以预测推送效果作为物体形状的函数。这种映射允许机器人为后续对象选择有效的推送位置,无论它们是以前操作过的实例、以前遇到的对象类的新实例,还是完全新的对象。在完全新对象的情况下,局部形状属性与对象的整体分布相结合,允许发现有效的推送位置。这些结果在移动机械手机器人上进行了演示,机器人在桌面表面上推动各种家居物品。
{"title":"Learning stable pushing locations","authors":"Tucker Hermans, Fuxin Li, James M. Rehg, A. Bobick","doi":"10.1109/DEVLRN.2013.6652539","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652539","url":null,"abstract":"We present a method by which a robot learns to predict effective push-locations as a function of object shape. The robot performs push experiments at many contact locations on multiple objects and records local and global shape features at each point of contact. The robot observes the outcome trajectories of the manipulations and computes a novel push-stability score for each trial. The robot then learns a regression function in order to predict push effectiveness as a function of object shape. This mapping allows the robot to select effective push locations for subsequent objects whether they are previously manipulated instances, new instances from previously encountered object classes, or entirely novel objects. In the totally novel object case, the local shape property coupled with the overall distribution of the object allows for the discovery of effective push locations. These results are demonstrated on a mobile manipulator robot pushing a variety of household objects on a tabletop surface.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121859117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Reinforcement learning with state-dependent discount factor 状态相关折现因子的强化学习
N. Yoshida, E. Uchibe, K. Doya
Conventional reinforcement learning algorithms have several parameters which determine the feature of learning process, called meta-parameters. In this study, we focus on the discount factor that influences the time scale of the tradeoff between immediate and delayed rewards. The discount factor is usually considered as a constant value, but we introduce the state-dependent discount function and a new optimization criterion for the reinforcement learning algorithm. We first derive a new algorithm under the criterion, named ExQ-learning and we prove that the algorithm converges to the optimal action-value function in the meaning of new criterion w.p.1. We then present a framework to optimize the discount factor and the discount function by using an evolutionary algorithm. In order to validate the proposed method, we conduct a simple computer simulation and show that the proposed algorithm can find an appropriate state-dependent discount function with which performs better than that with a constant discount factor.
传统的强化学习算法有几个参数,这些参数决定了学习过程的特征,称为元参数。在本研究中,我们关注影响即时和延迟奖励权衡的时间尺度的折扣因子。折扣因子通常被认为是一个常数,但我们引入了状态相关的折扣函数和一个新的强化学习算法优化准则。我们首先在该准则下推导了一个新的算法ExQ-learning,并证明了该算法收敛于新准则w.p.1意义下的最优动作值函数。然后,我们提出了一个利用进化算法优化折现因子和折现函数的框架。为了验证所提出的方法,我们进行了简单的计算机模拟,并表明所提出的算法可以找到合适的状态相关折扣函数,该函数比使用恒定折扣因子的算法性能更好。
{"title":"Reinforcement learning with state-dependent discount factor","authors":"N. Yoshida, E. Uchibe, K. Doya","doi":"10.1109/DEVLRN.2013.6652533","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652533","url":null,"abstract":"Conventional reinforcement learning algorithms have several parameters which determine the feature of learning process, called meta-parameters. In this study, we focus on the discount factor that influences the time scale of the tradeoff between immediate and delayed rewards. The discount factor is usually considered as a constant value, but we introduce the state-dependent discount function and a new optimization criterion for the reinforcement learning algorithm. We first derive a new algorithm under the criterion, named ExQ-learning and we prove that the algorithm converges to the optimal action-value function in the meaning of new criterion w.p.1. We then present a framework to optimize the discount factor and the discount function by using an evolutionary algorithm. In order to validate the proposed method, we conduct a simple computer simulation and show that the proposed algorithm can find an appropriate state-dependent discount function with which performs better than that with a constant discount factor.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126497515","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Autonomous learning of domain models using two-dimensional probability distributions 基于二维概率分布的领域模型自主学习
Witold Sowiski, Frank Guerin
An autonomous agent placed without any prior knowledge in an environment without goals or a reward function will need to develop a model of that environment using an unguided approach by discovering patters occurring in its observations. We expand on a prior algorithm which allows an agent to achieve that by learning clusters in probability distributions of one-dimensional sensory variables and propose a novel quadtree-based algorithm for two dimensions. We then evaluate it in a dynamic continuous domain involving a ball being thrown onto uneven terrain, simulated using a physics engine. Finally, we put forward criteria which can be used to evaluate a domain model without requiring goals and apply them to our work. We show that adding two-dimensional rules to the algorithm improves the model and that such models can be transferred to similar but previously-unseen environments.
一个没有任何先验知识的自主智能体被放置在一个没有目标或奖励功能的环境中,将需要通过发现其观察中出现的模式来使用非引导方法开发该环境的模型。我们扩展了先前的算法,该算法允许智能体通过学习一维感官变量的概率分布中的聚类来实现这一目标,并提出了一种新的基于四叉树的二维算法。然后,我们在一个动态连续域中评估它,包括一个球被扔到不平坦的地形上,使用物理引擎模拟。最后,我们提出了可以在不需要目标的情况下评估领域模型的标准,并将其应用到我们的工作中。我们表明,在算法中添加二维规则可以改善模型,并且这些模型可以转移到类似但以前未见过的环境中。
{"title":"Autonomous learning of domain models using two-dimensional probability distributions","authors":"Witold Sowiski, Frank Guerin","doi":"10.1109/DEVLRN.2013.6652524","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652524","url":null,"abstract":"An autonomous agent placed without any prior knowledge in an environment without goals or a reward function will need to develop a model of that environment using an unguided approach by discovering patters occurring in its observations. We expand on a prior algorithm which allows an agent to achieve that by learning clusters in probability distributions of one-dimensional sensory variables and propose a novel quadtree-based algorithm for two dimensions. We then evaluate it in a dynamic continuous domain involving a ball being thrown onto uneven terrain, simulated using a physics engine. Finally, we put forward criteria which can be used to evaluate a domain model without requiring goals and apply them to our work. We show that adding two-dimensional rules to the algorithm improves the model and that such models can be transferred to similar but previously-unseen environments.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124351884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning versatile sensorimotor coordination with goal babbling and neural associative dynamics 学习多用途的感觉运动协调与目标咿呀学语和神经联想动力学
R. F. Reinhart, Matthias Rolf
We combine an efficient exploration mechanism for bootstrapping of sensorimotor coordination skills with an associative network model to represent multiple coordination styles. Both approaches are integrated into a three-phased process of exploration, i.e. discovery of a new model, consolidation, the long-term storage of multiple models in a dynamical associative network, and exploitation of multiple models by the neural dynamics for versatile sensorimotor coordination. The proposed exploration-consolidation process is demonstrated for a planar robotic manipulator with ten degrees of freedom. Exploitation of sensorimotor coordination from the consolidated neural dynamics features motor hysteresis and additionally comprises a forward model that can be utilized to interpret proprioceptive feedback.
我们将感觉运动协调技能的有效探索机制与表达多种协调风格的关联网络模型相结合。这两种方法都被整合到一个三个阶段的探索过程中,即发现一个新的模型,巩固,多个模型在动态联想网络中的长期存储,以及通过神经动力学利用多个模型进行多功能感觉运动协调。以平面十自由度机械臂为例,对该方法进行了验证。从巩固的神经动力学中开发感觉运动协调的特点是运动滞后,另外还包括一个可以用来解释本体感觉反馈的前向模型。
{"title":"Learning versatile sensorimotor coordination with goal babbling and neural associative dynamics","authors":"R. F. Reinhart, Matthias Rolf","doi":"10.1109/DEVLRN.2013.6652566","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652566","url":null,"abstract":"We combine an efficient exploration mechanism for bootstrapping of sensorimotor coordination skills with an associative network model to represent multiple coordination styles. Both approaches are integrated into a three-phased process of exploration, i.e. discovery of a new model, consolidation, the long-term storage of multiple models in a dynamical associative network, and exploitation of multiple models by the neural dynamics for versatile sensorimotor coordination. The proposed exploration-consolidation process is demonstrated for a planar robotic manipulator with ten degrees of freedom. Exploitation of sensorimotor coordination from the consolidated neural dynamics features motor hysteresis and additionally comprises a forward model that can be utilized to interpret proprioceptive feedback.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124337093","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Do beliefs about a robot's capabilities influence alignment to its actions? 对机器人能力的信念会影响其行动的一致性吗?
Anna-Lisa Vollmer, B. Wrede, K. Rohlfing, A. Cangelosi
Interlocutors in a dialog align on many aspects of behavior (word choice, speech rate, syntactic structure, gestures, facial expressions, etc.). Such alignment has been proposed to be the basis for communicating successfully. We believe alignment could be beneficial for smooth human-robot interaction and facilitate robot action learning from demonstration. Recent research put forward a mediated communicative design account of alignment according to which interlocutors align stronger when they believe it will lead to communicative success. Branigan et al. showed that when interacting with an artificial system, participants aligned their lexical choices more to an artificial system they believed to be basic than to one they believed to be advanced. Our work extends these results in two ways: First, instead of an artificial computer dialog system, participants interact with a humanoid robot, the iCub robot. Second, instead of lexical choice, our work investigates alignment in the domain of manual actions. In an action demonstration and matching game, we examine the extent to which participants who believe that they are playing with a basic version or an advanced version of the iCub robot adapt the way they execute actions to what their robot partner has previously shown to them. Our results confirm that alignment also takes place in action demonstration. We were not able to replicate Branigan et al.'s results in general in this setup, but in line with their findings, participants with a low questionnaire score on neuroticism and participants who are familiar with robots aligned their actions more to a robot they believed to be basic than to one they believed to be advanced.
对话中的对话者在行为的许多方面都是一致的(用词、语速、句法结构、手势、面部表情等)。这种一致性被认为是成功沟通的基础。我们相信对齐可以使人机交互更加顺畅,并促进机器人从演示中学习动作。最近的研究提出了一种中介交际设计理论,根据该理论,当对话者认为结盟会导致交际成功时,他们的结盟会更强。布莱尼根等人的研究表明,当参与者与人工系统互动时,他们的词汇选择更倾向于他们认为是基本的人工系统,而不是他们认为是高级的人工系统。我们的工作以两种方式扩展了这些结果:首先,参与者与人形机器人iCub机器人进行交互,而不是人工计算机对话系统。其次,我们研究的不是词汇选择,而是手动操作领域的对齐。在动作演示和匹配游戏中,我们检查了认为自己正在玩iCub机器人的基本版本或高级版本的参与者在多大程度上适应他们执行动作的方式,以适应他们的机器人伙伴之前向他们展示的内容。我们的结果证实了在行动演示中也会发生对齐。我们无法在这个设置中复制Branigan等人的结果,但与他们的发现一致,神经质问卷得分低的参与者和熟悉机器人的参与者更倾向于他们认为基本的机器人,而不是他们认为先进的机器人。
{"title":"Do beliefs about a robot's capabilities influence alignment to its actions?","authors":"Anna-Lisa Vollmer, B. Wrede, K. Rohlfing, A. Cangelosi","doi":"10.1109/DEVLRN.2013.6652521","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652521","url":null,"abstract":"Interlocutors in a dialog align on many aspects of behavior (word choice, speech rate, syntactic structure, gestures, facial expressions, etc.). Such alignment has been proposed to be the basis for communicating successfully. We believe alignment could be beneficial for smooth human-robot interaction and facilitate robot action learning from demonstration. Recent research put forward a mediated communicative design account of alignment according to which interlocutors align stronger when they believe it will lead to communicative success. Branigan et al. showed that when interacting with an artificial system, participants aligned their lexical choices more to an artificial system they believed to be basic than to one they believed to be advanced. Our work extends these results in two ways: First, instead of an artificial computer dialog system, participants interact with a humanoid robot, the iCub robot. Second, instead of lexical choice, our work investigates alignment in the domain of manual actions. In an action demonstration and matching game, we examine the extent to which participants who believe that they are playing with a basic version or an advanced version of the iCub robot adapt the way they execute actions to what their robot partner has previously shown to them. Our results confirm that alignment also takes place in action demonstration. We were not able to replicate Branigan et al.'s results in general in this setup, but in line with their findings, participants with a low questionnaire score on neuroticism and participants who are familiar with robots aligned their actions more to a robot they believed to be basic than to one they believed to be advanced.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123990138","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Autonomous learning of active multi-scale binocular vision 主动多尺度双目视觉的自主学习
L. Lonini, Yu Zhao, Pramod Chandrashekhariah, Bertram E. Shi, J. Triesch
We present a method for autonomously learning representations of visual disparity between images from left and right eye, as well as appropriate vergence movements to fixate objects with both eyes. A sparse coding model (perception) encodes sensory information using binocular basis functions, while a reinforcement learner (behavior) generates the eye movement, according to the sensed disparity. Perception and behavior develop in parallel, by minimizing the same cost function: the reconstruction error of the stimulus by the generative model. In order to efficiently cope with multiple disparity ranges, sparse coding models are learnt at multiple scales, encoding disparities at various resolutions. Similarly, vergence commands are defined on a logarithmic scale to allow both coarse and fine actions. We demonstrate the efficacy of the proposed method using the humanoid robot iCub. We show that the model is fully self-calibrating and does not require any prior information about the camera parameters or the system dynamics.
我们提出了一种自主学习左眼和右眼图像视觉差异表征的方法,以及两只眼睛注视物体的适当收敛运动。稀疏编码模型(感知)利用双眼基函数对感觉信息进行编码,而强化学习模型(行为)根据感知到的视差产生眼动。通过最小化相同的成本函数,即生成模型对刺激的重建误差,感知和行为是并行发展的。为了有效地处理多个视差范围,在多个尺度上学习稀疏编码模型,对不同分辨率的视差进行编码。类似地,收敛命令在对数尺度上定义,以允许粗操作和细操作。我们用仿人机器人iCub验证了该方法的有效性。我们证明了该模型是完全自校准的,并且不需要任何关于相机参数或系统动力学的先验信息。
{"title":"Autonomous learning of active multi-scale binocular vision","authors":"L. Lonini, Yu Zhao, Pramod Chandrashekhariah, Bertram E. Shi, J. Triesch","doi":"10.1109/DEVLRN.2013.6652541","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652541","url":null,"abstract":"We present a method for autonomously learning representations of visual disparity between images from left and right eye, as well as appropriate vergence movements to fixate objects with both eyes. A sparse coding model (perception) encodes sensory information using binocular basis functions, while a reinforcement learner (behavior) generates the eye movement, according to the sensed disparity. Perception and behavior develop in parallel, by minimizing the same cost function: the reconstruction error of the stimulus by the generative model. In order to efficiently cope with multiple disparity ranges, sparse coding models are learnt at multiple scales, encoding disparities at various resolutions. Similarly, vergence commands are defined on a logarithmic scale to allow both coarse and fine actions. We demonstrate the efficacy of the proposed method using the humanoid robot iCub. We show that the model is fully self-calibrating and does not require any prior information about the camera parameters or the system dynamics.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129155063","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Towards understanding the origin of infant directed speech: A vocal robot with infant-like articulation 了解婴儿定向言语的起源:一个具有类似婴儿发音的发声机器人
Yuki Sasamoto, Naoto Nishijima, M. Asada
Infant-Directed Speech (IDS) is the non-standard form of caregivers' speech to their infants. Developmental studies indicate that IDS changes from infant-directed to adult-directed depending on infant's age and/or linguistic level. However, it is still unclear what features in infants cause IDS. This article introduces a vocal robot with an infant-like articulatory system to attack the issue by means of a constructive approach. A preliminary experiment implies that our robot can vocalize structurally similar to infant articulation although it is mechanically rather different.
婴儿指向语是照顾者对婴儿的非标准语言形式。发展研究表明,根据婴儿的年龄和/或语言水平,IDS从婴儿导向转变为成人导向。然而,目前尚不清楚婴儿的哪些特征会导致IDS。本文介绍了一种具有类似婴儿发音系统的发声机器人,通过一种建设性的方法来解决这个问题。一项初步实验表明,我们的机器人可以发出类似于婴儿发音的结构,尽管它在机械上有很大的不同。
{"title":"Towards understanding the origin of infant directed speech: A vocal robot with infant-like articulation","authors":"Yuki Sasamoto, Naoto Nishijima, M. Asada","doi":"10.1109/DEVLRN.2013.6652562","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652562","url":null,"abstract":"Infant-Directed Speech (IDS) is the non-standard form of caregivers' speech to their infants. Developmental studies indicate that IDS changes from infant-directed to adult-directed depending on infant's age and/or linguistic level. However, it is still unclear what features in infants cause IDS. This article introduces a vocal robot with an infant-like articulatory system to attack the issue by means of a constructive approach. A preliminary experiment implies that our robot can vocalize structurally similar to infant articulation although it is mechanically rather different.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127029573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Structural bootstrapping at the sensorimotor level for the fast acquisition of action knowledge for cognitive robots 认知机器人快速获取动作知识的感觉运动水平结构自举
E. Aksoy, M. Tamosiunaite, Rok Vuga, A. Ude, C. Geib, Mark Steedman, F. Wörgötter
Autonomous robots are faced with the problem of encoding complex actions (e.g. complete manipulations) in a generic and generalizable way. Recently we had introduced the Semantic Event Chains (SECs) as a new representation which can be directly computed from a stream of 3D images and is based on changes in the relationships between objects involved in a manipulation. Here we show that the SEC framework can be extended (called “extended SEC”) with action-related information and used to achieve and encode two important cognitive properties relevant for advanced autonomous robots: The extended SEC enables us to determine whether an action representation (1) needs to be newly created and stored in its entirety in the robot's memory or (2) whether one of the already known and memorized action representations just needs to be refined. In human cognition these two processes (1 and 2) are known as accommodation and assimilation. Thus, here we show that the extended SEC representation can be used to realize these processes originally defined by Piaget for the first time in a robotic application. This is of fundamental importance for any cognitive agent as it allows categorizing observed actions in new versus known ones, storing only the relevant aspects.
自主机器人面临着用通用和可推广的方式对复杂动作(如完整操作)进行编码的问题。最近,我们引入了语义事件链(SECs)作为一种新的表示,它可以直接从3D图像流中计算出来,并且基于操作中涉及的对象之间关系的变化。在这里,我们展示了SEC框架可以用与动作相关的信息进行扩展(称为“扩展SEC”),并用于实现和编码与高级自主机器人相关的两个重要认知属性:扩展SEC使我们能够确定一个动作表示(1)是否需要新创建并完整地存储在机器人的记忆中,或者(2)是否一个已知和记忆的动作表示只需要改进。在人类认知中,这两个过程(1和2)被称为适应和同化。因此,我们在这里展示了扩展的SEC表示可以用于实现最初由Piaget在机器人应用中定义的这些过程。这对于任何认知代理都是至关重要的,因为它允许将观察到的行为分为新的和已知的,只存储相关的方面。
{"title":"Structural bootstrapping at the sensorimotor level for the fast acquisition of action knowledge for cognitive robots","authors":"E. Aksoy, M. Tamosiunaite, Rok Vuga, A. Ude, C. Geib, Mark Steedman, F. Wörgötter","doi":"10.1109/DEVLRN.2013.6652537","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652537","url":null,"abstract":"Autonomous robots are faced with the problem of encoding complex actions (e.g. complete manipulations) in a generic and generalizable way. Recently we had introduced the Semantic Event Chains (SECs) as a new representation which can be directly computed from a stream of 3D images and is based on changes in the relationships between objects involved in a manipulation. Here we show that the SEC framework can be extended (called “extended SEC”) with action-related information and used to achieve and encode two important cognitive properties relevant for advanced autonomous robots: The extended SEC enables us to determine whether an action representation (1) needs to be newly created and stored in its entirety in the robot's memory or (2) whether one of the already known and memorized action representations just needs to be refined. In human cognition these two processes (1 and 2) are known as accommodation and assimilation. Thus, here we show that the extended SEC representation can be used to realize these processes originally defined by Piaget for the first time in a robotic application. This is of fundamental importance for any cognitive agent as it allows categorizing observed actions in new versus known ones, storing only the relevant aspects.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122746347","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Temporal emphasis for goal extraction in task demonstration to a humanoid robot by naive users 幼稚用户对仿人机器人任务演示中目标提取的时间重点
Konstantinos Theofilis, K. Lohan, Chrystopher L. Nehaniv, K. Dautenhahn, B. Wrede
Goal extraction in learning by demonstration is a complex problem. A novel approach, inspired by developmental psychology and focused on use in experiments with naive users, is presented in this paper. Participants were presenting a simple task, how to stack three boxes, to the humanoid robot iCub. The stationary states of the task - 1 box, 2 boxes stacked, 3 boxes stacked - were defined and the time span of each state was measured. Analysis of the results showed that there is a significant result that users tend to keep the boxes stationary longer upon completion of the end goal than upon completion of the sub-goals. A simple and straightforward learning algorithm was then used on the demonstration data, using only the time spans of the stationary states. The learning algorithm successfully detected the end goal. These temporal differences, functioning as emphasis, could be used as a complementary mechanism for goal extraction in imitation learning. Furthermore, it is suggested that since a simple, straightforward learning algorithm can use these pauses to recognise the goal state, humans may also be able to use this pause as a complementary mechanism for recognising the goal state of a task.
示范学习中的目标提取是一个复杂的问题。本文提出了一种新的方法,受发展心理学的启发,并专注于在天真用户的实验中使用。参与者向人形机器人iCub展示了一个简单的任务,即如何堆叠三个盒子。定义了1箱、2箱堆叠、3箱堆叠的任务静态状态,并测量了每个状态的时间跨度。对结果的分析表明,有一个显著的结果,即用户在完成最终目标时比在完成子目标时更倾向于保持盒子静止更长时间。然后对演示数据使用了一种简单直接的学习算法,仅使用稳态的时间跨度。学习算法成功检测到最终目标。这些时间差异作为重点,可以作为模仿学习中目标提取的补充机制。此外,有人建议,由于简单,直接的学习算法可以使用这些暂停来识别目标状态,人类也可以使用这种暂停作为识别任务目标状态的补充机制。
{"title":"Temporal emphasis for goal extraction in task demonstration to a humanoid robot by naive users","authors":"Konstantinos Theofilis, K. Lohan, Chrystopher L. Nehaniv, K. Dautenhahn, B. Wrede","doi":"10.1109/DEVLRN.2013.6652536","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652536","url":null,"abstract":"Goal extraction in learning by demonstration is a complex problem. A novel approach, inspired by developmental psychology and focused on use in experiments with naive users, is presented in this paper. Participants were presenting a simple task, how to stack three boxes, to the humanoid robot iCub. The stationary states of the task - 1 box, 2 boxes stacked, 3 boxes stacked - were defined and the time span of each state was measured. Analysis of the results showed that there is a significant result that users tend to keep the boxes stationary longer upon completion of the end goal than upon completion of the sub-goals. A simple and straightforward learning algorithm was then used on the demonstration data, using only the time spans of the stationary states. The learning algorithm successfully detected the end goal. These temporal differences, functioning as emphasis, could be used as a complementary mechanism for goal extraction in imitation learning. Furthermore, it is suggested that since a simple, straightforward learning algorithm can use these pauses to recognise the goal state, humans may also be able to use this pause as a complementary mechanism for recognising the goal state of a task.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115062060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Adaptive reachability assessment in the humanoid robot iCub 仿人机器人iCub的自适应可达性评估
Salomón Ramírez-Contla, D. Marocco
We present a model for reachability assessment implemented in a simulated iCub humanoid robot. The robot uses a neural network both for estimating reachability and as a controller for the arm. During training, multi-modality information including vision and proprioception of the effector's length was provided, along with tactile and postural information. The task was to assess if a target in view was at reach range. After training with data from two different effector's lengths, the system generalised also for a third one, both for producing reaching postures and for assessing reachability. We present preliminary results that show good reachability predictions with a decrease in confidence that display a depth gradient.
我们提出了一个模拟iCub人形机器人可达性评估模型。机器人使用神经网络来估计可达性,并作为手臂的控制器。在训练过程中,提供多模态信息,包括视觉和效应器长度的本体感觉,以及触觉和姿势信息。任务是评估目标是否在可及范围内。在用来自两个不同效应器长度的数据进行训练后,该系统也对第三个效应器进行了一般化,包括产生伸展姿势和评估可达性。我们提出了初步结果,显示出良好的可达性预测,但显示深度梯度的置信度降低。
{"title":"Adaptive reachability assessment in the humanoid robot iCub","authors":"Salomón Ramírez-Contla, D. Marocco","doi":"10.1109/DEVLRN.2013.6652546","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652546","url":null,"abstract":"We present a model for reachability assessment implemented in a simulated iCub humanoid robot. The robot uses a neural network both for estimating reachability and as a controller for the arm. During training, multi-modality information including vision and proprioception of the effector's length was provided, along with tactile and postural information. The task was to assess if a target in view was at reach range. After training with data from two different effector's lengths, the system generalised also for a third one, both for producing reaching postures and for assessing reachability. We present preliminary results that show good reachability predictions with a decrease in confidence that display a depth gradient.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132641434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1