{"title":"一种基于多目标跟踪的摄像机运动引导方法","authors":"Puchun Liu, Bo Li, Sheng Bi, Muye Li, Chen Zheng","doi":"10.1109/CYBER55403.2022.9907417","DOIUrl":null,"url":null,"abstract":"Multi-object tracking (MOT) attracts great attention in computer vision while playing an increasingly essential role in manufacturing and life. There is currently an urgent requirement for practical implementation of MOT algorithms to guide production in industry. Nevertheless, the algorithm guiding camera movement is prone to be naive in reality to some degree, which is not conducive to handle complex situations and may cause damage. In this paper, we propose a deep-learning-based MOT method to guide camera movement. In particular, jointly learnt detector and embedding model (JDE) is adopted to extract the features of live stream through data preprocessing and training, which detects the spacial and temporal information of objects, for instance, pedestrians. Moreover, we propose a cognitive-based network segmentation method which makes edge-cloud collaboration possible. Additionally, object location information provided by deep learning allows object clustering and weight allocation, followed by utilizing PID algorithm to guide camera motion. Our method is compared with conventional model through several metrics, especially Euclidean Average Distance, which indicates the effectiveness, reliability and robustness of our model.","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"24 10","pages":"150-155"},"PeriodicalIF":1.5000,"publicationDate":"2022-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Camera Movement Guidance Method based on Multi-Object Tracking\",\"authors\":\"Puchun Liu, Bo Li, Sheng Bi, Muye Li, Chen Zheng\",\"doi\":\"10.1109/CYBER55403.2022.9907417\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multi-object tracking (MOT) attracts great attention in computer vision while playing an increasingly essential role in manufacturing and life. There is currently an urgent requirement for practical implementation of MOT algorithms to guide production in industry. Nevertheless, the algorithm guiding camera movement is prone to be naive in reality to some degree, which is not conducive to handle complex situations and may cause damage. In this paper, we propose a deep-learning-based MOT method to guide camera movement. In particular, jointly learnt detector and embedding model (JDE) is adopted to extract the features of live stream through data preprocessing and training, which detects the spacial and temporal information of objects, for instance, pedestrians. Moreover, we propose a cognitive-based network segmentation method which makes edge-cloud collaboration possible. Additionally, object location information provided by deep learning allows object clustering and weight allocation, followed by utilizing PID algorithm to guide camera motion. Our method is compared with conventional model through several metrics, especially Euclidean Average Distance, which indicates the effectiveness, reliability and robustness of our model.\",\"PeriodicalId\":34110,\"journal\":{\"name\":\"IET Cybersystems and Robotics\",\"volume\":\"24 10\",\"pages\":\"150-155\"},\"PeriodicalIF\":1.5000,\"publicationDate\":\"2022-07-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IET Cybersystems and Robotics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CYBER55403.2022.9907417\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IET Cybersystems and Robotics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CYBER55403.2022.9907417","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
A Camera Movement Guidance Method based on Multi-Object Tracking
Multi-object tracking (MOT) attracts great attention in computer vision while playing an increasingly essential role in manufacturing and life. There is currently an urgent requirement for practical implementation of MOT algorithms to guide production in industry. Nevertheless, the algorithm guiding camera movement is prone to be naive in reality to some degree, which is not conducive to handle complex situations and may cause damage. In this paper, we propose a deep-learning-based MOT method to guide camera movement. In particular, jointly learnt detector and embedding model (JDE) is adopted to extract the features of live stream through data preprocessing and training, which detects the spacial and temporal information of objects, for instance, pedestrians. Moreover, we propose a cognitive-based network segmentation method which makes edge-cloud collaboration possible. Additionally, object location information provided by deep learning allows object clustering and weight allocation, followed by utilizing PID algorithm to guide camera motion. Our method is compared with conventional model through several metrics, especially Euclidean Average Distance, which indicates the effectiveness, reliability and robustness of our model.