Vitoantonio Bevilacqua, D. Barone, Francesco Cipriani, Gaetano D'Onghia, Giuseppe Mastrandrea, G. Mastronardi, M. Suma, Dario D'Ambruoso
{"title":"情感框架下支持决策的手势动作识别新工具","authors":"Vitoantonio Bevilacqua, D. Barone, Francesco Cipriani, Gaetano D'Onghia, Giuseppe Mastrandrea, G. Mastronardi, M. Suma, Dario D'Ambruoso","doi":"10.1109/INISTA.2014.6873616","DOIUrl":null,"url":null,"abstract":"Introduction and objective: the purpose of this work is to design and implement an innovative tool to recognize 16 different human gestural actions and use them to predict 7 different emotional states. The solution proposed in this paper is based on RGB and depth information of 2D/3D images acquired from a commercial RGB-D sensor called Kinect. Materials: the dataset is a collection of several human actions made by different actors. Each action is performed by each actor for three times in each video. 20 actors perform 16 different actions, both seated and upright, totalling 40 videos per actor. Methods: human gestural actions are recognized by means feature extractions as angles and distances related to joints of human skeleton from RGB and depth images. Emotions are selected according to the state-of-the-art. Experimental results: despite truly similar actions, the overall-accuracy reached is approximately 80%. Conclusions and future works: the proposed work seems to be back-ground- and speed-independent, and it will be used in the future as part of a multimodal emotion recognition software based on facial expressions and speech analysis as well.","PeriodicalId":339652,"journal":{"name":"2014 IEEE International Symposium on Innovations in Intelligent Systems and Applications (INISTA) Proceedings","volume":"239 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":"{\"title\":\"A new tool for gestural action recognition to support decisions in emotional framework\",\"authors\":\"Vitoantonio Bevilacqua, D. Barone, Francesco Cipriani, Gaetano D'Onghia, Giuseppe Mastrandrea, G. Mastronardi, M. Suma, Dario D'Ambruoso\",\"doi\":\"10.1109/INISTA.2014.6873616\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Introduction and objective: the purpose of this work is to design and implement an innovative tool to recognize 16 different human gestural actions and use them to predict 7 different emotional states. The solution proposed in this paper is based on RGB and depth information of 2D/3D images acquired from a commercial RGB-D sensor called Kinect. Materials: the dataset is a collection of several human actions made by different actors. Each action is performed by each actor for three times in each video. 20 actors perform 16 different actions, both seated and upright, totalling 40 videos per actor. Methods: human gestural actions are recognized by means feature extractions as angles and distances related to joints of human skeleton from RGB and depth images. Emotions are selected according to the state-of-the-art. Experimental results: despite truly similar actions, the overall-accuracy reached is approximately 80%. Conclusions and future works: the proposed work seems to be back-ground- and speed-independent, and it will be used in the future as part of a multimodal emotion recognition software based on facial expressions and speech analysis as well.\",\"PeriodicalId\":339652,\"journal\":{\"name\":\"2014 IEEE International Symposium on Innovations in Intelligent Systems and Applications (INISTA) Proceedings\",\"volume\":\"239 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2014-06-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"8\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2014 IEEE International Symposium on Innovations in Intelligent Systems and Applications (INISTA) Proceedings\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/INISTA.2014.6873616\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2014 IEEE International Symposium on Innovations in Intelligent Systems and Applications (INISTA) Proceedings","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/INISTA.2014.6873616","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A new tool for gestural action recognition to support decisions in emotional framework
Introduction and objective: the purpose of this work is to design and implement an innovative tool to recognize 16 different human gestural actions and use them to predict 7 different emotional states. The solution proposed in this paper is based on RGB and depth information of 2D/3D images acquired from a commercial RGB-D sensor called Kinect. Materials: the dataset is a collection of several human actions made by different actors. Each action is performed by each actor for three times in each video. 20 actors perform 16 different actions, both seated and upright, totalling 40 videos per actor. Methods: human gestural actions are recognized by means feature extractions as angles and distances related to joints of human skeleton from RGB and depth images. Emotions are selected according to the state-of-the-art. Experimental results: despite truly similar actions, the overall-accuracy reached is approximately 80%. Conclusions and future works: the proposed work seems to be back-ground- and speed-independent, and it will be used in the future as part of a multimodal emotion recognition software based on facial expressions and speech analysis as well.