Xinyu Lin, V. Kitanovski, Qianni Zhang, E. Izquierdo
{"title":"Enhanced multi-view dancing videos synchronisation","authors":"Xinyu Lin, V. Kitanovski, Qianni Zhang, E. Izquierdo","doi":"10.1109/WIAMIS.2012.6226773","DOIUrl":null,"url":null,"abstract":"This paper describes a system for automatically synchronising multi-view video sequences of Salsa dancing recorded with multimodal capturing platform. The multimodal capturing setup consists of audiovisual streams along with depth maps and inertial measurements. Part of the dataset was video sequences captured from machine vision cameras and Microsoft Kinect sensor that were not temporal synchronised during the capturing stage. As an essential step, we proposed efficient solutions for synchronisation of these data based on co-occurrence appearance changes. In order to improve the accuracy, the proposed system employed state-of-art body detection and tracking algorithm to obtain Region of Interest, within which the appearance changes are analysed. The accurately synchronised video set can then be further analysed and augmented for visualisation and evaluation of dancing performance.","PeriodicalId":346777,"journal":{"name":"2012 13th International Workshop on Image Analysis for Multimedia Interactive Services","volume":"3 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2012 13th International Workshop on Image Analysis for Multimedia Interactive Services","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WIAMIS.2012.6226773","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
This paper describes a system for automatically synchronising multi-view video sequences of Salsa dancing recorded with multimodal capturing platform. The multimodal capturing setup consists of audiovisual streams along with depth maps and inertial measurements. Part of the dataset was video sequences captured from machine vision cameras and Microsoft Kinect sensor that were not temporal synchronised during the capturing stage. As an essential step, we proposed efficient solutions for synchronisation of these data based on co-occurrence appearance changes. In order to improve the accuracy, the proposed system employed state-of-art body detection and tracking algorithm to obtain Region of Interest, within which the appearance changes are analysed. The accurately synchronised video set can then be further analysed and augmented for visualisation and evaluation of dancing performance.