Cunshu Pan , Zhenhua Dai , Yuhao Zhang , Heshan Zhang , Mengwei Fan , Jin Xu
{"title":"An approach for accurately extracting vehicle trajectory from aerial videos based on computer vision","authors":"Cunshu Pan , Zhenhua Dai , Yuhao Zhang , Heshan Zhang , Mengwei Fan , Jin Xu","doi":"10.1016/j.measurement.2024.116212","DOIUrl":null,"url":null,"abstract":"<div><div>Vehicle trajectory data holds valuable information for advanced driving development and traffic analysis. While unmanned aerial vehicle (UAV) offer a broader perspective, the detection of small-scale vehicles in video frames still suffers from low accuracy or is even missed. This study proposes a comprehensive technical framework for accurate vehicle trajectory extraction, encompassing six main components: video stabilization, vehicle detection, vehicle tracking, lane marking detection, coordinate transformation, and data denoising. To mitigate video jitter, the SURF and FLANN stabilization algorithms are utilized. An enhanced detector based on You Only Look Once X (YOLOX) is employed for multi-target vehicle detection, incorporating a shallow feature extraction module within the detection head to improve the performance for low-level and small-scale features. Efficient Channel Attention (ECA) modules are integrated before the neck to further boost the expressiveness. Additionally, a sliding window inference method is applied at the input stage to prevent compression of high-resolution video frames. The Savitzky-Golay filter is used for trajectory noise reduction. Verification results demonstrate that the improved YOLOX achieves a mean average precision (mAP) of 88.7 %, an enhancement of 5.6 % over the original model. When compared to advanced YOLOv7 and YOLOv8 models, the proposed method increases mAP@50 by 7.63 % and 1.07 %, respectively. The Mostly Tracked (MT) trajectories metric reaches 98.9 %, and the root-mean-square error of one-sided localization is approximately 0.05 m. These results confirm that the proposed framework is an effective tool for high-accuracy vehicle trajectory data collection in traffic studies. Additionally, a vehicle trajectory dataset has been developed and is publicly accessible at <span><span>www.cqskyeyex.com</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18349,"journal":{"name":"Measurement","volume":"242 ","pages":"Article 116212"},"PeriodicalIF":5.2000,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Measurement","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0263224124020979","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0
Abstract
Vehicle trajectory data holds valuable information for advanced driving development and traffic analysis. While unmanned aerial vehicle (UAV) offer a broader perspective, the detection of small-scale vehicles in video frames still suffers from low accuracy or is even missed. This study proposes a comprehensive technical framework for accurate vehicle trajectory extraction, encompassing six main components: video stabilization, vehicle detection, vehicle tracking, lane marking detection, coordinate transformation, and data denoising. To mitigate video jitter, the SURF and FLANN stabilization algorithms are utilized. An enhanced detector based on You Only Look Once X (YOLOX) is employed for multi-target vehicle detection, incorporating a shallow feature extraction module within the detection head to improve the performance for low-level and small-scale features. Efficient Channel Attention (ECA) modules are integrated before the neck to further boost the expressiveness. Additionally, a sliding window inference method is applied at the input stage to prevent compression of high-resolution video frames. The Savitzky-Golay filter is used for trajectory noise reduction. Verification results demonstrate that the improved YOLOX achieves a mean average precision (mAP) of 88.7 %, an enhancement of 5.6 % over the original model. When compared to advanced YOLOv7 and YOLOv8 models, the proposed method increases mAP@50 by 7.63 % and 1.07 %, respectively. The Mostly Tracked (MT) trajectories metric reaches 98.9 %, and the root-mean-square error of one-sided localization is approximately 0.05 m. These results confirm that the proposed framework is an effective tool for high-accuracy vehicle trajectory data collection in traffic studies. Additionally, a vehicle trajectory dataset has been developed and is publicly accessible at www.cqskyeyex.com.
期刊介绍:
Contributions are invited on novel achievements in all fields of measurement and instrumentation science and technology. Authors are encouraged to submit novel material, whose ultimate goal is an advancement in the state of the art of: measurement and metrology fundamentals, sensors, measurement instruments, measurement and estimation techniques, measurement data processing and fusion algorithms, evaluation procedures and methodologies for plants and industrial processes, performance analysis of systems, processes and algorithms, mathematical models for measurement-oriented purposes, distributed measurement systems in a connected world.