{"title":"FACT:用于多目标跟踪的特征自适应持续学习跟踪器","authors":"Rongzihan Song, Zhenyu Weng, Huiping Zhuang, Jinchang Ren, Yongming Chen, Zhiping Lin","doi":"arxiv-2409.07904","DOIUrl":null,"url":null,"abstract":"Multiple object tracking (MOT) involves identifying multiple targets and\nassigning them corresponding IDs within a video sequence, where occlusions are\noften encountered. Recent methods address occlusions using appearance cues\nthrough online learning techniques to improve adaptivity or offline learning\ntechniques to utilize temporal information from videos. However, most existing\nonline learning-based MOT methods are unable to learn from all past tracking\ninformation to improve adaptivity on long-term occlusions while maintaining\nreal-time tracking speed. On the other hand, temporal information-based offline\nlearning methods maintain a long-term memory to store past tracking\ninformation, but this approach restricts them to use only local past\ninformation during tracking. To address these challenges, we propose a new MOT\nframework called the Feature Adaptive Continual-learning Tracker (FACT), which\nenables real-time tracking and feature learning for targets by utilizing all\npast tracking information. We demonstrate that the framework can be integrated\nwith various state-of-the-art feature-based trackers, thereby improving their\ntracking ability. Specifically, we develop the feature adaptive\ncontinual-learning (FAC) module, a neural network that can be trained online to\nlearn features adaptively using all past tracking information during tracking.\nMoreover, we also introduce a two-stage association module specifically\ndesigned for the proposed continual learning-based tracking. Extensive\nexperiment results demonstrate that the proposed method achieves\nstate-of-the-art online tracking performance on MOT17 and MOT20 benchmarks. The\ncode will be released upon acceptance.","PeriodicalId":501130,"journal":{"name":"arXiv - CS - Computer Vision and Pattern Recognition","volume":"87 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"FACT: Feature Adaptive Continual-learning Tracker for Multiple Object Tracking\",\"authors\":\"Rongzihan Song, Zhenyu Weng, Huiping Zhuang, Jinchang Ren, Yongming Chen, Zhiping Lin\",\"doi\":\"arxiv-2409.07904\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multiple object tracking (MOT) involves identifying multiple targets and\\nassigning them corresponding IDs within a video sequence, where occlusions are\\noften encountered. Recent methods address occlusions using appearance cues\\nthrough online learning techniques to improve adaptivity or offline learning\\ntechniques to utilize temporal information from videos. However, most existing\\nonline learning-based MOT methods are unable to learn from all past tracking\\ninformation to improve adaptivity on long-term occlusions while maintaining\\nreal-time tracking speed. On the other hand, temporal information-based offline\\nlearning methods maintain a long-term memory to store past tracking\\ninformation, but this approach restricts them to use only local past\\ninformation during tracking. To address these challenges, we propose a new MOT\\nframework called the Feature Adaptive Continual-learning Tracker (FACT), which\\nenables real-time tracking and feature learning for targets by utilizing all\\npast tracking information. We demonstrate that the framework can be integrated\\nwith various state-of-the-art feature-based trackers, thereby improving their\\ntracking ability. Specifically, we develop the feature adaptive\\ncontinual-learning (FAC) module, a neural network that can be trained online to\\nlearn features adaptively using all past tracking information during tracking.\\nMoreover, we also introduce a two-stage association module specifically\\ndesigned for the proposed continual learning-based tracking. Extensive\\nexperiment results demonstrate that the proposed method achieves\\nstate-of-the-art online tracking performance on MOT17 and MOT20 benchmarks. The\\ncode will be released upon acceptance.\",\"PeriodicalId\":501130,\"journal\":{\"name\":\"arXiv - CS - Computer Vision and Pattern Recognition\",\"volume\":\"87 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Computer Vision and Pattern Recognition\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.07904\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computer Vision and Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07904","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
多目标跟踪(MOT)涉及在视频序列中识别多个目标并为其分配相应的 ID,而在视频序列中经常会遇到遮挡物。最近的方法通过在线学习技术来提高适应性,或通过离线学习技术来利用视频中的时间信息,从而利用外观线索来解决遮挡问题。然而,大多数现有的基于在线学习的 MOT 方法都无法从所有过去的跟踪信息中学习,从而在保持实时跟踪速度的同时提高对长期遮挡的适应性。另一方面,基于时间信息的离线学习方法会保留一个长期存储器来存储过去的跟踪信息,但这种方法限制了它们在跟踪过程中只能使用局部的过去信息。为了应对这些挑战,我们提出了一种名为 "特征自适应持续学习跟踪器"(FACT)的新型 MOT 框架,通过利用所有过去的跟踪信息,实现对目标的实时跟踪和特征学习。我们证明,该框架可以与各种最先进的基于特征的跟踪器集成,从而提高它们的跟踪能力。具体来说,我们开发了特征自适应持续学习(FAC)模块,这是一个可在线训练的神经网络,可在跟踪过程中利用所有过去的跟踪信息自适应地学习特征。广泛的实验结果表明,所提出的方法在 MOT17 和 MOT20 基准上实现了最先进的在线跟踪性能。代码将在验收通过后发布。
FACT: Feature Adaptive Continual-learning Tracker for Multiple Object Tracking
Multiple object tracking (MOT) involves identifying multiple targets and
assigning them corresponding IDs within a video sequence, where occlusions are
often encountered. Recent methods address occlusions using appearance cues
through online learning techniques to improve adaptivity or offline learning
techniques to utilize temporal information from videos. However, most existing
online learning-based MOT methods are unable to learn from all past tracking
information to improve adaptivity on long-term occlusions while maintaining
real-time tracking speed. On the other hand, temporal information-based offline
learning methods maintain a long-term memory to store past tracking
information, but this approach restricts them to use only local past
information during tracking. To address these challenges, we propose a new MOT
framework called the Feature Adaptive Continual-learning Tracker (FACT), which
enables real-time tracking and feature learning for targets by utilizing all
past tracking information. We demonstrate that the framework can be integrated
with various state-of-the-art feature-based trackers, thereby improving their
tracking ability. Specifically, we develop the feature adaptive
continual-learning (FAC) module, a neural network that can be trained online to
learn features adaptively using all past tracking information during tracking.
Moreover, we also introduce a two-stage association module specifically
designed for the proposed continual learning-based tracking. Extensive
experiment results demonstrate that the proposed method achieves
state-of-the-art online tracking performance on MOT17 and MOT20 benchmarks. The
code will be released upon acceptance.