Fast Video Classification based on unidirectional temporal differences based dynamic spatial selection with custom loss function and new class suggestion
{"title":"Fast Video Classification based on unidirectional temporal differences based dynamic spatial selection with custom loss function and new class suggestion","authors":"Prashant Kaushik, V. Saxena","doi":"10.1109/ICDT57929.2023.10150644","DOIUrl":null,"url":null,"abstract":"With the new and emerging usages of faster video classification and identifications of new classes has pushed the research in this direction. Be it like similar video detection, percentage of similarity, anomaly detection or finding something trending in the current videos. Use of identification of objects in frames and features in surrounding has proven its advantages for video classifications specially for short video similarity detection. Use of temporally important objects and actions has also proved advantages for video classifications. However existing methods takes huge computation to train the model and does not detect the possibility of new classes. To address this scenario for faster video classification and reducing the training time and computation cost, we propose one directional temporal difference of frames and selectively selecting the spatial information with custom loss function. This allows faster training of the models and has a capability of detecting the new classes in the production videos. This new class detection will provide us new ways looking at video data and thus new kinds of platform conceptualization. Experiments were conducted in UCF and MSVD datasets. Validations were done using statistical methods like f-test etc. Validation for being faster in training are done using comparison of state of the art methods. The novelty of the work lies in the processing of video data for similarity detection in short video and new kinds of intelligence extraction. Which is generated from regression values for possible new classes of video.","PeriodicalId":266681,"journal":{"name":"2023 International Conference on Disruptive Technologies (ICDT)","volume":"01 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Conference on Disruptive Technologies (ICDT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDT57929.2023.10150644","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
With the new and emerging usages of faster video classification and identifications of new classes has pushed the research in this direction. Be it like similar video detection, percentage of similarity, anomaly detection or finding something trending in the current videos. Use of identification of objects in frames and features in surrounding has proven its advantages for video classifications specially for short video similarity detection. Use of temporally important objects and actions has also proved advantages for video classifications. However existing methods takes huge computation to train the model and does not detect the possibility of new classes. To address this scenario for faster video classification and reducing the training time and computation cost, we propose one directional temporal difference of frames and selectively selecting the spatial information with custom loss function. This allows faster training of the models and has a capability of detecting the new classes in the production videos. This new class detection will provide us new ways looking at video data and thus new kinds of platform conceptualization. Experiments were conducted in UCF and MSVD datasets. Validations were done using statistical methods like f-test etc. Validation for being faster in training are done using comparison of state of the art methods. The novelty of the work lies in the processing of video data for similarity detection in short video and new kinds of intelligence extraction. Which is generated from regression values for possible new classes of video.