Regions of interest in observing robot hand movement by a cooperative robot

Toyomi Fujita, C. Privitera
{"title":"Regions of interest in observing robot hand movement by a cooperative robot","authors":"Toyomi Fujita, C. Privitera","doi":"10.1109/ICAWST.2013.6765457","DOIUrl":null,"url":null,"abstract":"This paper presents a method for generating regions-of-interest in a scene of robot hand movement observed by a cooperative robot for action recognition. In a cooperative work, a robot needs to be aware of an action of its partner robot by detecting some regions-of-interest in the visual field like human visual scanpath. To generate regions-of-interests by a robot, we have applied image processing algorithms based on active top-down feature patterns and bottom-up spatial kernels. The algorithms have produced energy maps from the images observed by the robot and they were combined with different weights to generate algorithmic regions-of-interests. They were compared with human regions-of-interest measured by a psychophysical experiment and the algorithmic predictability of scanpath was evaluated using a positional similarity index. Experimental results showed that presented method is applicable to the detection of regions-of-interests in hand movement.","PeriodicalId":68697,"journal":{"name":"炎黄地理","volume":"21 1","pages":"318-321"},"PeriodicalIF":0.0000,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"炎黄地理","FirstCategoryId":"1089","ListUrlMain":"https://doi.org/10.1109/ICAWST.2013.6765457","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

This paper presents a method for generating regions-of-interest in a scene of robot hand movement observed by a cooperative robot for action recognition. In a cooperative work, a robot needs to be aware of an action of its partner robot by detecting some regions-of-interest in the visual field like human visual scanpath. To generate regions-of-interests by a robot, we have applied image processing algorithms based on active top-down feature patterns and bottom-up spatial kernels. The algorithms have produced energy maps from the images observed by the robot and they were combined with different weights to generate algorithmic regions-of-interests. They were compared with human regions-of-interest measured by a psychophysical experiment and the algorithmic predictability of scanpath was evaluated using a positional similarity index. Experimental results showed that presented method is applicable to the detection of regions-of-interests in hand movement.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
合作机器人观察机械手运动的兴趣区域
提出了一种在协作机器人观察到的机械手运动场景中生成兴趣区域的方法,用于动作识别。在协同工作中,机器人需要像人类的视觉扫描路径一样,通过检测视野中的兴趣区域来感知同伴的动作。为了生成机器人感兴趣的区域,我们应用了基于主动自顶向下特征模式和自底向上空间核的图像处理算法。这些算法从机器人观察到的图像中生成能量图,并将它们与不同的权重组合在一起,生成算法感兴趣的区域。将它们与心理物理实验测量的人类感兴趣区域进行比较,并使用位置相似性指数评估扫描路径的算法可预测性。实验结果表明,该方法适用于手部运动兴趣区域的检测。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
784
期刊最新文献
Make decision boundary smoother by transition learning Neurophysiological evidence of the cognitive cycle and the emergence of awareness An efficient implementation of normalized cross-correlation image matching based on pyramid A hybrid recommender system based non-common items in social media "Canderoid": A mobile system to remotely monitor travelling status of the elderly with dementia
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1