Multi-Object Tracking (MOT) aims to detect and associate objects across frames while maintaining consistent IDs. While some approaches leverage both strong and weak cues alongside camera compensation to improve association, they struggle in scenarios involving high object density or nonlinear motion. To address these challenges, we propose RAP-SORT, a novel MOT framework that introduces four key innovations. First, the Robust Tracklet Confidence Modeling (RTCM) module models trajectory confidence by smoothing updates and applying second-order difference adjustments for low-confidence cases. Second, the Advanced Observation-Centric Recovery (AOCR) module facilitates trajectory recovery via linear interpolation and backtracking. Third, the Pseudo-Depth IoU (PDIoU) metric integrates height and depth cues into IoU calculations for enhanced spatial awareness. Finally, the Window Denoising (WD) module is tailored for the DanceTrack dataset, effectively mitigating the creation of new tracks caused by misdetections. RAP-SORT sets a new state-of-the-art on the DanceTrack and MOT20 benchmarks, achieving HOTA scores of 66.7 and 64.2, surpassing the previous best by 1.0 and 0.3, respectively, while also delivering competitive performance on MOT17. Code and models will be available soon at https://github.com/levi5611/RAP-SORT.
扫码关注我们
求助内容:
应助结果提醒方式:
