{"title":"Human Action Recognition Based on STDMI-HOG and STjoint Feature","authors":"Qianhan Wx, Qian Huan, Xing Ll","doi":"10.1109/PIC53636.2021.9687036","DOIUrl":null,"url":null,"abstract":"More and more attention has been focused on the human action recognition domain. The existing methods are mostly based on single-mode data. However, single-mode data lacks adequate information. So, it is necessary to propose methods based on multimode data. In this paper, we extract two kinds of features from depth videos and skeleton sequences, named STDMI-HOG and STjoint feature respectively. STDMI-HOG is extracted from a new depth feature map Spatial-Temporal Depth Motion Image by Histogram of Oriented Gradient. STjoint feature is extracted from skeleton sequences by ST-GCN extractor. Then two kinds of features are connected to make up a one-dimensional vector. Finally, SVM classifies the actions according to the feature vector. To evaluate the performance, several experiments are conducted on two public datasets: the MSR Action3D dataset and the UTD-MHAD dataset. The accuracy of our method on two datasets is compared with the existing methods, and the experiments prove the outperformance of our method.","PeriodicalId":297239,"journal":{"name":"2021 IEEE International Conference on Progress in Informatics and Computing (PIC)","volume":"65 2 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Conference on Progress in Informatics and Computing (PIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/PIC53636.2021.9687036","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
More and more attention has been focused on the human action recognition domain. The existing methods are mostly based on single-mode data. However, single-mode data lacks adequate information. So, it is necessary to propose methods based on multimode data. In this paper, we extract two kinds of features from depth videos and skeleton sequences, named STDMI-HOG and STjoint feature respectively. STDMI-HOG is extracted from a new depth feature map Spatial-Temporal Depth Motion Image by Histogram of Oriented Gradient. STjoint feature is extracted from skeleton sequences by ST-GCN extractor. Then two kinds of features are connected to make up a one-dimensional vector. Finally, SVM classifies the actions according to the feature vector. To evaluate the performance, several experiments are conducted on two public datasets: the MSR Action3D dataset and the UTD-MHAD dataset. The accuracy of our method on two datasets is compared with the existing methods, and the experiments prove the outperformance of our method.