{"title":"Temporal Superpixels based Human Action Localization","authors":"Sami Ullah, Najmul Hassan, Naeem Bhatti","doi":"10.1109/ICET.2018.8603608","DOIUrl":null,"url":null,"abstract":"In this paper, we present human action localization in videos. It is a challenging task to localize actions performed in videos where the foreground areas containing video object as well as the background areas depict motion simultaneously. We set the main objective to identify action related spatio-temporal areas in a video. The proposed approach consists of two main steps. At first, we perform saturation (S plane of HSV space) based background subtraction to obtain foreground in video frames. Secondly, we compute the optical flow of video frames. Performing the superpixels segmentation of each video frame, we classify the superpixels as temporal or stationary using optical flow information and the extracted foreground. The collection of the classified temporal superpixels provide the required action localization in videos. We present qualitative as well as quantitative evaluation of our approach using UCF sports and Weizmann actions dataset. The visual results and quantitative performance measures show that the proposed approach well localizes the action areas in the taken action sequences for evaluation.","PeriodicalId":443353,"journal":{"name":"2018 14th International Conference on Emerging Technologies (ICET)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 14th International Conference on Emerging Technologies (ICET)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICET.2018.8603608","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
In this paper, we present human action localization in videos. It is a challenging task to localize actions performed in videos where the foreground areas containing video object as well as the background areas depict motion simultaneously. We set the main objective to identify action related spatio-temporal areas in a video. The proposed approach consists of two main steps. At first, we perform saturation (S plane of HSV space) based background subtraction to obtain foreground in video frames. Secondly, we compute the optical flow of video frames. Performing the superpixels segmentation of each video frame, we classify the superpixels as temporal or stationary using optical flow information and the extracted foreground. The collection of the classified temporal superpixels provide the required action localization in videos. We present qualitative as well as quantitative evaluation of our approach using UCF sports and Weizmann actions dataset. The visual results and quantitative performance measures show that the proposed approach well localizes the action areas in the taken action sequences for evaluation.