情感框架下支持决策的手势动作识别新工具

Vitoantonio Bevilacqua, D. Barone, Francesco Cipriani, Gaetano D'Onghia, Giuseppe Mastrandrea, G. Mastronardi, M. Suma, Dario D'Ambruoso
{"title":"情感框架下支持决策的手势动作识别新工具","authors":"Vitoantonio Bevilacqua, D. Barone, Francesco Cipriani, Gaetano D'Onghia, Giuseppe Mastrandrea, G. Mastronardi, M. Suma, Dario D'Ambruoso","doi":"10.1109/INISTA.2014.6873616","DOIUrl":null,"url":null,"abstract":"Introduction and objective: the purpose of this work is to design and implement an innovative tool to recognize 16 different human gestural actions and use them to predict 7 different emotional states. The solution proposed in this paper is based on RGB and depth information of 2D/3D images acquired from a commercial RGB-D sensor called Kinect. Materials: the dataset is a collection of several human actions made by different actors. Each action is performed by each actor for three times in each video. 20 actors perform 16 different actions, both seated and upright, totalling 40 videos per actor. Methods: human gestural actions are recognized by means feature extractions as angles and distances related to joints of human skeleton from RGB and depth images. Emotions are selected according to the state-of-the-art. Experimental results: despite truly similar actions, the overall-accuracy reached is approximately 80%. Conclusions and future works: the proposed work seems to be back-ground- and speed-independent, and it will be used in the future as part of a multimodal emotion recognition software based on facial expressions and speech analysis as well.","PeriodicalId":339652,"journal":{"name":"2014 IEEE International Symposium on Innovations in Intelligent Systems and Applications (INISTA) Proceedings","volume":"239 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":"{\"title\":\"A new tool for gestural action recognition to support decisions in emotional framework\",\"authors\":\"Vitoantonio Bevilacqua, D. Barone, Francesco Cipriani, Gaetano D'Onghia, Giuseppe Mastrandrea, G. Mastronardi, M. Suma, Dario D'Ambruoso\",\"doi\":\"10.1109/INISTA.2014.6873616\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Introduction and objective: the purpose of this work is to design and implement an innovative tool to recognize 16 different human gestural actions and use them to predict 7 different emotional states. The solution proposed in this paper is based on RGB and depth information of 2D/3D images acquired from a commercial RGB-D sensor called Kinect. Materials: the dataset is a collection of several human actions made by different actors. Each action is performed by each actor for three times in each video. 20 actors perform 16 different actions, both seated and upright, totalling 40 videos per actor. Methods: human gestural actions are recognized by means feature extractions as angles and distances related to joints of human skeleton from RGB and depth images. Emotions are selected according to the state-of-the-art. Experimental results: despite truly similar actions, the overall-accuracy reached is approximately 80%. Conclusions and future works: the proposed work seems to be back-ground- and speed-independent, and it will be used in the future as part of a multimodal emotion recognition software based on facial expressions and speech analysis as well.\",\"PeriodicalId\":339652,\"journal\":{\"name\":\"2014 IEEE International Symposium on Innovations in Intelligent Systems and Applications (INISTA) Proceedings\",\"volume\":\"239 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2014-06-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"8\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2014 IEEE International Symposium on Innovations in Intelligent Systems and Applications (INISTA) Proceedings\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/INISTA.2014.6873616\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2014 IEEE International Symposium on Innovations in Intelligent Systems and Applications (INISTA) Proceedings","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/INISTA.2014.6873616","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 8

摘要

简介及目的:本工作的目的是设计并实现一个创新的工具来识别16种不同的人类手势动作,并利用它们来预测7种不同的情绪状态。本文提出的解决方案基于从商用RGB- d传感器Kinect获取的2D/3D图像的RGB和深度信息。材料:数据集是由不同参与者所做的几个人类行为的集合。每个动作由每个演员在每个视频中表演三次。20名演员表演16种不同的动作,有坐着的,也有直立的,每个演员总共40个视频。方法:从RGB图像和深度图像中提取与人体骨骼关节相关的角度和距离等特征来识别人体手势动作。情感是根据技术水平来选择的。实验结果:尽管动作非常相似,但总体准确率达到约80%。结论和未来的工作:所提出的工作似乎是背景和速度无关的,它将在未来作为基于面部表情和语音分析的多模态情感识别软件的一部分使用。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
A new tool for gestural action recognition to support decisions in emotional framework
Introduction and objective: the purpose of this work is to design and implement an innovative tool to recognize 16 different human gestural actions and use them to predict 7 different emotional states. The solution proposed in this paper is based on RGB and depth information of 2D/3D images acquired from a commercial RGB-D sensor called Kinect. Materials: the dataset is a collection of several human actions made by different actors. Each action is performed by each actor for three times in each video. 20 actors perform 16 different actions, both seated and upright, totalling 40 videos per actor. Methods: human gestural actions are recognized by means feature extractions as angles and distances related to joints of human skeleton from RGB and depth images. Emotions are selected according to the state-of-the-art. Experimental results: despite truly similar actions, the overall-accuracy reached is approximately 80%. Conclusions and future works: the proposed work seems to be back-ground- and speed-independent, and it will be used in the future as part of a multimodal emotion recognition software based on facial expressions and speech analysis as well.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
A Multi-Objective Graph-based Genetic Algorithm for image segmentation Threat assessment for GPS navigation Elastic constant identification of laminated composite beam with metaheuristic algorithms Optimization of waiting and journey time in group elevator system using genetic algorithm Multilayer medium technique for nondestructive EM-properties measurement of radar absorbing materials using flanged rectangular waveguide sensor and FDTD method
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1