{"title":"虚拟PTZ摄像机在360视频中的行人跟踪","authors":"Vito Monteleone, Liliana Lo Presti, M. Cascia","doi":"10.1109/RTSI.2018.8548499","DOIUrl":null,"url":null,"abstract":"Since the data acquired by a PTZ camera change while adjusting the pan, tilt and zoom parameters, the results of tracking algorithms are difficult to reproduce; such difficulty limits the development and the comparison of tracking algorithms with PTZ cameras. The recently introduced 360-degree cameras acquire spherical views of the environment, generally stored as equirectangular images. Each pixel of an equirectangular image corresponds to a point on the spherical surface. A gnomonic projection can be used to project the points on the spherical surface onto a plane tangent to the sphere. Such tangent plane can be interpreted as the image plane of a virtual PTZ camera oriented towards the point of tangency. This paper proposes a framework to simulate PTZ cameras from 360-degree video enabling, in this way, the development and comparison of PTZ-based tracking algorithms. Furthermore, within the above mentioned framework, this paper presents a novel pedestrian tracking algorithm for 360-degree videos. The proposed algorithm aims at estimating the pan, tilt and zoom parameters required to control the virtual camera in such a way that the target is always at the center of the virtual camera view. The proposed method belongs to the category of tracking-by-detection algorithms; it also exploits the use of a dynamic memory to store the appearance models of the best past target detections. Preliminary results on a publicly available benchmark demonstrate the viability of the proposed approach.","PeriodicalId":363896,"journal":{"name":"2018 IEEE 4th International Forum on Research and Technology for Society and Industry (RTSI)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":"{\"title\":\"Pedestrian Tracking in 360 Video by Virtual PTZ Cameras\",\"authors\":\"Vito Monteleone, Liliana Lo Presti, M. Cascia\",\"doi\":\"10.1109/RTSI.2018.8548499\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Since the data acquired by a PTZ camera change while adjusting the pan, tilt and zoom parameters, the results of tracking algorithms are difficult to reproduce; such difficulty limits the development and the comparison of tracking algorithms with PTZ cameras. The recently introduced 360-degree cameras acquire spherical views of the environment, generally stored as equirectangular images. Each pixel of an equirectangular image corresponds to a point on the spherical surface. A gnomonic projection can be used to project the points on the spherical surface onto a plane tangent to the sphere. Such tangent plane can be interpreted as the image plane of a virtual PTZ camera oriented towards the point of tangency. This paper proposes a framework to simulate PTZ cameras from 360-degree video enabling, in this way, the development and comparison of PTZ-based tracking algorithms. Furthermore, within the above mentioned framework, this paper presents a novel pedestrian tracking algorithm for 360-degree videos. The proposed algorithm aims at estimating the pan, tilt and zoom parameters required to control the virtual camera in such a way that the target is always at the center of the virtual camera view. The proposed method belongs to the category of tracking-by-detection algorithms; it also exploits the use of a dynamic memory to store the appearance models of the best past target detections. Preliminary results on a publicly available benchmark demonstrate the viability of the proposed approach.\",\"PeriodicalId\":363896,\"journal\":{\"name\":\"2018 IEEE 4th International Forum on Research and Technology for Society and Industry (RTSI)\",\"volume\":\"30 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"6\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2018 IEEE 4th International Forum on Research and Technology for Society and Industry (RTSI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/RTSI.2018.8548499\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE 4th International Forum on Research and Technology for Society and Industry (RTSI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/RTSI.2018.8548499","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Pedestrian Tracking in 360 Video by Virtual PTZ Cameras
Since the data acquired by a PTZ camera change while adjusting the pan, tilt and zoom parameters, the results of tracking algorithms are difficult to reproduce; such difficulty limits the development and the comparison of tracking algorithms with PTZ cameras. The recently introduced 360-degree cameras acquire spherical views of the environment, generally stored as equirectangular images. Each pixel of an equirectangular image corresponds to a point on the spherical surface. A gnomonic projection can be used to project the points on the spherical surface onto a plane tangent to the sphere. Such tangent plane can be interpreted as the image plane of a virtual PTZ camera oriented towards the point of tangency. This paper proposes a framework to simulate PTZ cameras from 360-degree video enabling, in this way, the development and comparison of PTZ-based tracking algorithms. Furthermore, within the above mentioned framework, this paper presents a novel pedestrian tracking algorithm for 360-degree videos. The proposed algorithm aims at estimating the pan, tilt and zoom parameters required to control the virtual camera in such a way that the target is always at the center of the virtual camera view. The proposed method belongs to the category of tracking-by-detection algorithms; it also exploits the use of a dynamic memory to store the appearance models of the best past target detections. Preliminary results on a publicly available benchmark demonstrate the viability of the proposed approach.