首页 > 最新文献

2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)最新文献

英文 中文
Learning spatial relations between objects from 3D scenes 从3D场景中学习物体之间的空间关系
Severin Fichtl, J. W. Alexander, Frank Guerin, Wail Mustafa, D. Kraft, N. Krüger
In this work, we learn a limited number of abstractions which can then be used to form preconditions for motor actions. These abstractions take the form of spatial relations amongst objects. We consider three “classes” of spatial relation: The objects either are separated from, on-top of, or inside each other. We have tackled this same problem in previous work (Fichtl et al., 2013). Here we report on recent improved results using a novel application of histograms to visually recognise a spatial relation between objects in the environment. Using this histogram based approach we are able to report a very high rate of success when the system is asked to recognise a spatial relation.
在这项工作中,我们学习了有限数量的抽象概念,然后可以用来形成运动动作的先决条件。这些抽象采用对象之间空间关系的形式。我们考虑空间关系的三“类”:对象要么是彼此分离的,要么是彼此之上的,要么是彼此内部的。我们在之前的工作中解决了同样的问题(Fichtl et al., 2013)。在这里,我们报告了最近使用直方图的新应用来视觉识别环境中物体之间的空间关系的改进结果。使用这种基于直方图的方法,当系统被要求识别空间关系时,我们能够报告非常高的成功率。
{"title":"Learning spatial relations between objects from 3D scenes","authors":"Severin Fichtl, J. W. Alexander, Frank Guerin, Wail Mustafa, D. Kraft, N. Krüger","doi":"10.1109/DEVLRN.2013.6652552","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652552","url":null,"abstract":"In this work, we learn a limited number of abstractions which can then be used to form preconditions for motor actions. These abstractions take the form of spatial relations amongst objects. We consider three “classes” of spatial relation: The objects either are separated from, on-top of, or inside each other. We have tackled this same problem in previous work (Fichtl et al., 2013). Here we report on recent improved results using a novel application of histograms to visually recognise a spatial relation between objects in the environment. Using this histogram based approach we are able to report a very high rate of success when the system is asked to recognise a spatial relation.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128481259","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Building specific contexts for on-line learning of dynamical tasks through non-verbal interaction 通过非语言互动建立动态任务在线学习的特定语境
A. D. Rengervé, Souheil Hanoune, P. Andry, M. Quoy, P. Gaussier
Trajectories can be encoded as attraction basin resulting from recruited associations between visually based localization and orientations to follow (low level behaviors). Navigation to different places according to some other multimodal information needs a particular learning. We propose a minimal model explaining such a behavior adaptation from non-verbal interaction with a teacher. Specific contexts can be recruited to prevent the behaviors to activate in cases the interaction showed they were inadequate. Still, the model is compatible with the recruitment of new low level behaviors. The tests done in simulation show the capabilities of the architecture, the limitations regarding the generalization and the learning speed. We also discuss the possible evolutions towards more bio-inspired models.
轨迹可以被编码为吸引盆地,这是由基于视觉的定位和跟随方向(低级行为)之间的关联产生的。根据其他一些多模式信息导航到不同的地方需要特殊的学习。我们提出了一个最小的模型来解释这种与老师的非语言互动的行为适应。可以利用特定的情境来防止在互动显示不充分的情况下激活行为。尽管如此,该模型与新的低水平行为的招募是兼容的。仿真测试表明了该体系结构的能力、泛化和学习速度方面的局限性。我们还讨论了向更多生物启发模型的可能进化。
{"title":"Building specific contexts for on-line learning of dynamical tasks through non-verbal interaction","authors":"A. D. Rengervé, Souheil Hanoune, P. Andry, M. Quoy, P. Gaussier","doi":"10.1109/DEVLRN.2013.6652564","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652564","url":null,"abstract":"Trajectories can be encoded as attraction basin resulting from recruited associations between visually based localization and orientations to follow (low level behaviors). Navigation to different places according to some other multimodal information needs a particular learning. We propose a minimal model explaining such a behavior adaptation from non-verbal interaction with a teacher. Specific contexts can be recruited to prevent the behaviors to activate in cases the interaction showed they were inadequate. Still, the model is compatible with the recruitment of new low level behaviors. The tests done in simulation show the capabilities of the architecture, the limitations regarding the generalization and the learning speed. We also discuss the possible evolutions towards more bio-inspired models.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126673079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
The significance of social input, early motion experiences, and attentional selection 社会输入、早期运动经验和注意选择的意义
Joseph M. Burling, Hanako Yoshida, Y. Nagai
Before babies acquire an adult-like visual capacity, they participate in a social world as a human learning system which promotes social activities around them and in turn dramatically alters their own social participation. Visual input becomes more dynamic as they gain self-generated movement, and such movement has a potential role in learning. The present study specifically looks at the expected change in motion of the early visual input that infants are exposed to, and the corresponding attentional coordination within the specific context of parent-infant interactions. The results will be discussed in terms of the significance of social input for development.
在婴儿获得像成年人一样的视觉能力之前,他们作为一个人类学习系统参与社会世界,这促进了他们周围的社会活动,反过来又极大地改变了他们自己的社会参与。当他们获得自我产生的运动时,视觉输入变得更加动态,这种运动在学习中具有潜在的作用。本研究特别关注婴儿接触到的早期视觉输入的预期运动变化,以及在父母-婴儿互动的特定背景下相应的注意协调。将根据社会投入对发展的重要性来讨论这些结果。
{"title":"The significance of social input, early motion experiences, and attentional selection","authors":"Joseph M. Burling, Hanako Yoshida, Y. Nagai","doi":"10.1109/DEVLRN.2013.6652556","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652556","url":null,"abstract":"Before babies acquire an adult-like visual capacity, they participate in a social world as a human learning system which promotes social activities around them and in turn dramatically alters their own social participation. Visual input becomes more dynamic as they gain self-generated movement, and such movement has a potential role in learning. The present study specifically looks at the expected change in motion of the early visual input that infants are exposed to, and the corresponding attentional coordination within the specific context of parent-infant interactions. The results will be discussed in terms of the significance of social input for development.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121747729","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Learning semantic components from subsymbolic multimodal perception 从亚符号多模态感知中学习语义成分
Olivier Mangin, Pierre-Yves Oudeyer
Perceptual systems often include sensors from several modalities. However, existing robots do not yet sufficiently discover patterns that are spread over the flow of multimodal data they receive. In this paper we present a framework that learns a dictionary of words from full spoken utterances, together with a set of gestures from human demonstrations and the semantic connection between words and gestures. We explain how to use a nonnegative matrix factorization algorithm to learn a dictionary of components that represent meaningful elements present in the multimodal perception, without providing the system with a symbolic representation of the semantics. We illustrate this framework by showing how a learner discovers word-like components from observation of gestures made by a human together with spoken descriptions of the gestures, and how it captures the semantic association between the two.
感知系统通常包括来自不同模态的传感器。然而,现有的机器人还不能充分发现分布在它们接收到的多模态数据流中的模式。在本文中,我们提出了一个框架,从完整的口头话语中学习单词字典,从人类演示中学习一组手势,以及单词和手势之间的语义联系。我们解释了如何使用非负矩阵分解算法来学习表示多模态感知中存在的有意义元素的组件字典,而不向系统提供语义的符号表示。我们通过展示学习者如何通过观察人类的手势以及对手势的口头描述来发现类词成分,以及它如何捕捉两者之间的语义关联来说明这个框架。
{"title":"Learning semantic components from subsymbolic multimodal perception","authors":"Olivier Mangin, Pierre-Yves Oudeyer","doi":"10.1109/DEVLRN.2013.6652563","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652563","url":null,"abstract":"Perceptual systems often include sensors from several modalities. However, existing robots do not yet sufficiently discover patterns that are spread over the flow of multimodal data they receive. In this paper we present a framework that learns a dictionary of words from full spoken utterances, together with a set of gestures from human demonstrations and the semantic connection between words and gestures. We explain how to use a nonnegative matrix factorization algorithm to learn a dictionary of components that represent meaningful elements present in the multimodal perception, without providing the system with a symbolic representation of the semantics. We illustrate this framework by showing how a learner discovers word-like components from observation of gestures made by a human together with spoken descriptions of the gestures, and how it captures the semantic association between the two.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"19 12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130033685","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
Autonomous reuse of motor exploration trajectories 运动探索轨迹的自主重用
Fabien C. Y. Benureau, Pierre-Yves Oudeyer
We present an algorithm for transferring exploration strategies between tasks that share a common motor space in the context of lifelong autonomous learning in robotics. The algorithm does not transfer observations, or make assumptions about how the learning is conducted. Instead, only selected motor commands are transferred between tasks, chosen autonomously according to an empirical measure of learning progress. We show that on a wide variety of variations from a source task, such as changing the object the robot is interacting with or altering the morphology of the robot, this simple and flexible transfer method increases early performance significantly in the new task. We also provide examples of situations where the transfer is not helpful.
在机器人终身自主学习的背景下,我们提出了一种在共享共同运动空间的任务之间转移探索策略的算法。该算法不会转移观察结果,也不会对如何进行学习做出假设。相反,只有选定的运动命令在任务之间传递,根据学习进度的经验衡量自主选择。我们表明,在源任务的各种变化中,例如改变机器人与之交互的对象或改变机器人的形态,这种简单而灵活的转移方法显着提高了新任务的早期性能。我们还提供了一些情况的例子,其中转移是没有帮助的。
{"title":"Autonomous reuse of motor exploration trajectories","authors":"Fabien C. Y. Benureau, Pierre-Yves Oudeyer","doi":"10.1109/DEVLRN.2013.6652567","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652567","url":null,"abstract":"We present an algorithm for transferring exploration strategies between tasks that share a common motor space in the context of lifelong autonomous learning in robotics. The algorithm does not transfer observations, or make assumptions about how the learning is conducted. Instead, only selected motor commands are transferred between tasks, chosen autonomously according to an empirical measure of learning progress. We show that on a wide variety of variations from a source task, such as changing the object the robot is interacting with or altering the morphology of the robot, this simple and flexible transfer method increases early performance significantly in the new task. We also provide examples of situations where the transfer is not helpful.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"124 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116156700","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A generative probabilistic framework for learning spatial language 空间语言学习的生成概率框架
C. Dawson, Jeremy B. Wright, Antons Rebguns, M. Valenzuela-Escarcega, Daniel Fried, P. Cohen
The language of space and spatial relations is a rich source of abstract semantic structure. We develop a probabilistic model that learns to understand utterances that describe spatial configurations of objects in a tabletop scene by seeking the meaning that best explains the sentence chosen. The inference problem is simplified by assuming that sentences express symbolic representations of (latent) semantic relations between referents and landmarks in space, and that given these symbolic representations, utterances and physical locations are conditionally independent. As such, the inference problem factors into a symbol-grounding component (linking propositions to physical locations) and a symbol-translation component (linking propositions to parse trees). We evaluate the model by eliciting production and comprehension data from human English speakers and find that our system recovers the referent of spatial utterances at a level of proficiency approaching human performance.
空间语言和空间关系是抽象语义结构的丰富来源。我们开发了一个概率模型,通过寻找最能解释所选句子的含义来学习理解描述桌面场景中物体空间配置的话语。通过假设句子表达了空间中指称物和地标之间(潜在的)语义关系的符号表征,并且假设这些符号表征,话语和物理位置是条件独立的,可以简化推理问题。因此,推理问题分为符号基础组件(将命题与物理位置连接起来)和符号翻译组件(将命题与解析树连接起来)。我们通过提取人类英语使用者的生产和理解数据来评估该模型,发现我们的系统在接近人类表现的熟练程度上恢复了空间话语的参考。
{"title":"A generative probabilistic framework for learning spatial language","authors":"C. Dawson, Jeremy B. Wright, Antons Rebguns, M. Valenzuela-Escarcega, Daniel Fried, P. Cohen","doi":"10.1109/DEVLRN.2013.6652560","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652560","url":null,"abstract":"The language of space and spatial relations is a rich source of abstract semantic structure. We develop a probabilistic model that learns to understand utterances that describe spatial configurations of objects in a tabletop scene by seeking the meaning that best explains the sentence chosen. The inference problem is simplified by assuming that sentences express symbolic representations of (latent) semantic relations between referents and landmarks in space, and that given these symbolic representations, utterances and physical locations are conditionally independent. As such, the inference problem factors into a symbol-grounding component (linking propositions to physical locations) and a symbol-translation component (linking propositions to parse trees). We evaluate the model by eliciting production and comprehension data from human English speakers and find that our system recovers the referent of spatial utterances at a level of proficiency approaching human performance.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"238 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132583270","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Extracting image features in static images for depth estimation 在静态图像中提取图像特征进行深度估计
M. Ogino, Junji Suzuki, M. Asada
Human feels three-dimensional effect for static image with the cues of various kinds of image features such as relative sizes of objects, up and down, rules of perspective, texture gradient, and shadow. The features are called pictorial depth cues. Human is thought to learn to extract these features as important cues for depth estimation in the developmental process. In this paper, we make a hypothesis that pictorial depth cues are acquired so that disparities can be predicted well and make a model that extracts features appropriate for depth estimation from static images. Random forest network is trained to extract important ones among a large amount image features so as to estimate motion and stereo disparities. The experiments with simulation and real environments show high correlation between estimated and real disparities.
人通过物体的相对大小、上下、透视规则、纹理渐变、阴影等各种图像特征的提示,对静态图像产生三维效果。这些特征被称为图像深度线索。人类在发育过程中学习提取这些特征作为深度估计的重要线索。在本文中,我们假设图像深度线索被获取,从而可以很好地预测差异,并建立了一个从静态图像中提取适合深度估计的特征的模型。训练随机森林网络从大量图像特征中提取重要特征,从而估计运动和立体差异。在模拟和真实环境下进行的实验表明,估计差值与实际差值高度相关。
{"title":"Extracting image features in static images for depth estimation","authors":"M. Ogino, Junji Suzuki, M. Asada","doi":"10.1109/DEVLRN.2013.6652551","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652551","url":null,"abstract":"Human feels three-dimensional effect for static image with the cues of various kinds of image features such as relative sizes of objects, up and down, rules of perspective, texture gradient, and shadow. The features are called pictorial depth cues. Human is thought to learn to extract these features as important cues for depth estimation in the developmental process. In this paper, we make a hypothesis that pictorial depth cues are acquired so that disparities can be predicted well and make a model that extracts features appropriate for depth estimation from static images. Random forest network is trained to extract important ones among a large amount image features so as to estimate motion and stereo disparities. The experiments with simulation and real environments show high correlation between estimated and real disparities.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126726676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Aquila 2.0 software architecture for cognitive robotics 认知机器人的Aquila 2.0软件架构
M. Peniak, Anthony F. Morse, A. Cangelosi
The modelling of the integration of various cognitive skills and modalities requires complex and computationally intensive algorithms running in parallel while controlling high-performance systems. The distribution of processing across many computers has certainly advanced our software ecosystem and opened up research to new possibilities. While this was an essential move, we are aspiring to augment the field of cognitive robotics by providing Aquila 2.0, a novel hi-performance software architecture utilising cross-platform, heterogeneous CPU-GPU modules loosely coupled with GUIs used for module management and data visualisation.
各种认知技能和模式的整合建模需要复杂和计算密集型的算法并行运行,同时控制高性能系统。跨多台计算机的处理分布无疑推动了我们的软件生态系统,并为研究开辟了新的可能性。虽然这是一个必要的举措,但我们渴望通过提供Aquila 2.0来增强认知机器人领域,Aquila 2.0是一种新型的高性能软件架构,利用跨平台、异构CPU-GPU模块与用于模块管理和数据可视化的gui松散耦合。
{"title":"Aquila 2.0 software architecture for cognitive robotics","authors":"M. Peniak, Anthony F. Morse, A. Cangelosi","doi":"10.1109/DEVLRN.2013.6652565","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652565","url":null,"abstract":"The modelling of the integration of various cognitive skills and modalities requires complex and computationally intensive algorithms running in parallel while controlling high-performance systems. The distribution of processing across many computers has certainly advanced our software ecosystem and opened up research to new possibilities. While this was an essential move, we are aspiring to augment the field of cognitive robotics by providing Aquila 2.0, a novel hi-performance software architecture utilising cross-platform, heterogeneous CPU-GPU modules loosely coupled with GUIs used for module management and data visualisation.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123169777","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Emergence of flexible prediction-based discrete decision making and continuous motion generation through actor-Q-learning 基于柔性预测的离散决策和通过actor- q学习的连续运动生成的出现
K. Shibata, Kenta Goto
In this paper, the authors first point the importance of three factors for filling the gap between humans and robots in the flexibility in the real world. Those are (1)parallel processing, (2)emergence through learning and solving “what” problems, and (3)abstraction and generalization on the abstract space. To explore the possibility of human-like flexibility in robots, a prediction-required task in which an agent (robot) gets a reward by capturing a moving target that sometimes becomes invisible was learned by reinforcement learning using a recurrent neural network. Even though the agent did not know in advance that “prediction is required” or “what information should be predicted”, appropriate discrete decision making, in which `capture' or `move' was chosen, and also continuous motion generation in two-dimensional space, could be acquired. Furthermore, in this task, the target sometimes changed its moving direction randomly when it became visible again from invisible state. Then the agent could change its moving direction promptly and appropriately without introducing any special architecture or technique. Such emergent property is what general parallel processing systems such as Subsumption architecture do not have, and the authors believe it is a key to solve the “Frame Problem” fundamentally.
在本文中,作者首先指出了三个因素对于填补现实世界中人类与机器人在灵活性方面的差距的重要性。它们是:(1)并行处理;(2)通过学习和解决“什么”问题而出现;(3)抽象空间的抽象和泛化。为了探索机器人具有类似人类的灵活性的可能性,通过使用循环神经网络的强化学习来学习一个需要预测的任务,其中代理(机器人)通过捕获有时变得不可见的移动目标获得奖励。即使agent事先不知道“需要预测”或“应该预测什么信息”,也可以获得适当的离散决策,选择“捕获”或“移动”,并在二维空间中连续生成运动。此外,在该任务中,当目标从不可见状态变为可见状态时,有时会随机改变其运动方向。这样,智能体就可以在不引入任何特殊架构或技术的情况下,迅速而适当地改变其移动方向。这种涌现性是一般的并行处理系统如包容架构所不具备的,是从根本上解决“框架问题”的关键。
{"title":"Emergence of flexible prediction-based discrete decision making and continuous motion generation through actor-Q-learning","authors":"K. Shibata, Kenta Goto","doi":"10.1109/DEVLRN.2013.6652559","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652559","url":null,"abstract":"In this paper, the authors first point the importance of three factors for filling the gap between humans and robots in the flexibility in the real world. Those are (1)parallel processing, (2)emergence through learning and solving “what” problems, and (3)abstraction and generalization on the abstract space. To explore the possibility of human-like flexibility in robots, a prediction-required task in which an agent (robot) gets a reward by capturing a moving target that sometimes becomes invisible was learned by reinforcement learning using a recurrent neural network. Even though the agent did not know in advance that “prediction is required” or “what information should be predicted”, appropriate discrete decision making, in which `capture' or `move' was chosen, and also continuous motion generation in two-dimensional space, could be acquired. Furthermore, in this task, the target sometimes changed its moving direction randomly when it became visible again from invisible state. Then the agent could change its moving direction promptly and appropriately without introducing any special architecture or technique. Such emergent property is what general parallel processing systems such as Subsumption architecture do not have, and the authors believe it is a key to solve the “Frame Problem” fundamentally.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127952319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Learning to recognize objects through curiosity-driven manipulation with the iCub humanoid robot 学习识别物体通过好奇心驱动操作与iCub人形机器人
S. Nguyen, S. Ivaldi, Natalia Lyubova, Alain Droniou, Damien Gérardeaux-Viret, David Filliat, V. Padois, Olivier Sigaud, Pierre-Yves Oudeyer
In this paper we address the problem of learning to recognize objects by manipulation in a developmental robotics scenario. In a life-long learning perspective, a humanoid robot should be capable of improving its knowledge of objects with active perception. Our approach stems from the cognitive development of infants, exploiting active curiosity-driven manipulation to improve perceptual learning of objects. These functionalities are implemented as perception, control and active exploration modules as part of the Cognitive Architecture of the MACSi project. In this paper we integrate these functionalities into an active perception system which learns to recognise objects through manipulation. Our work in this paper integrates a bottom-up vision system, a control system of a complex robot system and a top-down interactive exploration method, which actively chooses an exploration method to collect data and whether interacting with humans is profitable or not. Experimental results show that the humanoid robot iCub can learn to recognize 3D objects by manipulation and in interaction with teachers by choosing the adequate exploration strategy to enhance competence progress and by focusing its efforts on the most complex tasks. Thus the learner can learn interactively with humans by actively self-regulating its requests for help.
在本文中,我们解决了在发展机器人场景中通过操纵来学习识别物体的问题。从终身学习的角度来看,人形机器人应该能够通过主动感知来提高其对物体的知识。我们的方法源于婴儿的认知发展,利用主动的好奇心驱动的操作来提高对物体的感知学习。这些功能被实现为感知、控制和主动探索模块,作为MACSi项目认知架构的一部分。在本文中,我们将这些功能集成到一个主动感知系统中,该系统通过操纵来学习识别物体。我们在本文中的工作集成了自下而上的视觉系统、复杂机器人系统的控制系统和自上而下的交互式探索方法,主动选择一种探索方法来收集数据,以及与人类互动是否有利可图。实验结果表明,仿人机器人iCub通过选择适当的探索策略来提高能力进步,并将精力集中在最复杂的任务上,可以通过操纵和与教师的互动来学习识别3D物体。因此,学习者可以通过主动自我调节其请求帮助的方式与人类进行互动学习。
{"title":"Learning to recognize objects through curiosity-driven manipulation with the iCub humanoid robot","authors":"S. Nguyen, S. Ivaldi, Natalia Lyubova, Alain Droniou, Damien Gérardeaux-Viret, David Filliat, V. Padois, Olivier Sigaud, Pierre-Yves Oudeyer","doi":"10.1109/DEVLRN.2013.6652525","DOIUrl":"https://doi.org/10.1109/DEVLRN.2013.6652525","url":null,"abstract":"In this paper we address the problem of learning to recognize objects by manipulation in a developmental robotics scenario. In a life-long learning perspective, a humanoid robot should be capable of improving its knowledge of objects with active perception. Our approach stems from the cognitive development of infants, exploiting active curiosity-driven manipulation to improve perceptual learning of objects. These functionalities are implemented as perception, control and active exploration modules as part of the Cognitive Architecture of the MACSi project. In this paper we integrate these functionalities into an active perception system which learns to recognise objects through manipulation. Our work in this paper integrates a bottom-up vision system, a control system of a complex robot system and a top-down interactive exploration method, which actively chooses an exploration method to collect data and whether interacting with humans is profitable or not. Experimental results show that the humanoid robot iCub can learn to recognize 3D objects by manipulation and in interaction with teachers by choosing the adequate exploration strategy to enhance competence progress and by focusing its efforts on the most complex tasks. Thus the learner can learn interactively with humans by actively self-regulating its requests for help.","PeriodicalId":106997,"journal":{"name":"2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)","volume":"2 19","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131943364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
期刊
2013 IEEE Third Joint International Conference on Development and Learning and Epigenetic Robotics (ICDL)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1