{"title":"MOT-H: A Multi-Target Tracking Dataset Based on Horizontal View","authors":"Bixuan Zhang, Yuefeng Zhang","doi":"10.1145/3569966.3570064","DOIUrl":null,"url":null,"abstract":"The computer vision field is quickly developing, including multiple object tracking, as the big data age approaches. The majority of the effort is focused on tracking methods while less attention is paid to the most important aspect, data. After an analysis of existing datasets, we find that they commonly ignore the breakpoint problem in tracking and have low image quality. Thus we present the dataset named Multiple Object Tracking on Horizontal view (MOT-H). MOT-H is meticulously annotated on crowded scenes from the horizontal view, with the primary goal of proving anti-jamming performance against complicated occlusion or even complete occlusion. The breakpoint issue is emphasized, which means the target object temporarily leaves the scene and returns after a while. The proposed MOT-H dataset has 10 sequences, 20,311 frames, and 337,440 annotation boxes in total, with all pictures having the resolution of 3840 × 2160 and being filmed at 30 frames per second (fps). We establish a fair benchmark for the future object tracking method development. The whole dataset can be found at: https://drive.google.com/drive/folders/1SCUJAdbqXQStyV-F2M9UyGfsuCaxR73a?usp=sharing.","PeriodicalId":145580,"journal":{"name":"Proceedings of the 5th International Conference on Computer Science and Software Engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 5th International Conference on Computer Science and Software Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3569966.3570064","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The computer vision field is quickly developing, including multiple object tracking, as the big data age approaches. The majority of the effort is focused on tracking methods while less attention is paid to the most important aspect, data. After an analysis of existing datasets, we find that they commonly ignore the breakpoint problem in tracking and have low image quality. Thus we present the dataset named Multiple Object Tracking on Horizontal view (MOT-H). MOT-H is meticulously annotated on crowded scenes from the horizontal view, with the primary goal of proving anti-jamming performance against complicated occlusion or even complete occlusion. The breakpoint issue is emphasized, which means the target object temporarily leaves the scene and returns after a while. The proposed MOT-H dataset has 10 sequences, 20,311 frames, and 337,440 annotation boxes in total, with all pictures having the resolution of 3840 × 2160 and being filmed at 30 frames per second (fps). We establish a fair benchmark for the future object tracking method development. The whole dataset can be found at: https://drive.google.com/drive/folders/1SCUJAdbqXQStyV-F2M9UyGfsuCaxR73a?usp=sharing.