Md Atiqur Rahman Ahad, J. Tan, Hyoungseop Kim, S. Ishikawa
{"title":"SURF-based spatio-temporal history image method for action representation","authors":"Md Atiqur Rahman Ahad, J. Tan, Hyoungseop Kim, S. Ishikawa","doi":"10.1109/ICIT.2011.5754412","DOIUrl":null,"url":null,"abstract":"Researches on action understanding and analysis are very crucial for various applications in computer vision. However, these face numerous challenges to represent and recognize different complex actions. This paper presents a noble spatio-temporal 3D (XYT) method for recognizing various complex activities, with a blend of local and global feature-based approach for motion representation. We incorporate SURF (Speeded-Up Robust Features), which is a scale- and rotation-invariant interest point detector and descriptor. Based on the interest points, optical flow-based directional motion history and energy images are developed. In this approach, the flow-based motion vectors are split into four different channels. From these channels, the corresponding four directional templates are computed. 56-D feature vector is calculated according to the Hu invariants for each action. k-nearest neighbor classification scheme is employed for recognition. We employ leave-one-out cross-validation method for partitioning scheme. We apply our method to outdoor dataset and we achieve satisfactory recognition results. We compare our method with some of other approaches and show that our method outperforms them.","PeriodicalId":356868,"journal":{"name":"2011 IEEE International Conference on Industrial Technology","volume":"114 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2011-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2011 IEEE International Conference on Industrial Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICIT.2011.5754412","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
Researches on action understanding and analysis are very crucial for various applications in computer vision. However, these face numerous challenges to represent and recognize different complex actions. This paper presents a noble spatio-temporal 3D (XYT) method for recognizing various complex activities, with a blend of local and global feature-based approach for motion representation. We incorporate SURF (Speeded-Up Robust Features), which is a scale- and rotation-invariant interest point detector and descriptor. Based on the interest points, optical flow-based directional motion history and energy images are developed. In this approach, the flow-based motion vectors are split into four different channels. From these channels, the corresponding four directional templates are computed. 56-D feature vector is calculated according to the Hu invariants for each action. k-nearest neighbor classification scheme is employed for recognition. We employ leave-one-out cross-validation method for partitioning scheme. We apply our method to outdoor dataset and we achieve satisfactory recognition results. We compare our method with some of other approaches and show that our method outperforms them.