Asmaa Hasan Alrubaie, Maisa'a Abid Ali Khodhe, A. Abdulameer
{"title":"改进 YOLOV7 算法以检测录制视频中的物体","authors":"Asmaa Hasan Alrubaie, Maisa'a Abid Ali Khodhe, A. Abdulameer","doi":"10.24996/ijs.2024.65.2.35","DOIUrl":null,"url":null,"abstract":" Object detection algorithms play an important role in detecting people in surveillance videos. In recent years, with the rapid development of deep learning, the performance of object detection has improved by leaps and bounds, and the scheme of object detection by the YOLOV7 algorithm has also been born. Traditional object detection methods often fail to achieve a balance between speed and accuracy. To address these issues, in this research, an improved YOLOv7 algorithm performance is proposed to get the best speed-to-accuracy balance compared to state-of-the-art object detection within recorded videos using an effective compression method. This method calculates the difference between frames of video, and by using the zero difference approach by removing the duplicate frames from the recorded video and choosing only the meaningful frames based on many variables, including frame size, frame details, and the distance of the frames, influence the choice of a meaningful frame, and this will reduce the size of the video by eliminating the frames comparable to those chosen. \n Additionally, any other datasets or pre-trained weights have not been used; YOLOv7 has been exclusively trained on the MS COCO dataset from scratch. In order to ensure the effectiveness of this approach, numerous detection systems are used in this work. Additionally, positive performance results to reduce the processing time required for object detection have been attained.","PeriodicalId":14698,"journal":{"name":"Iraqi Journal of Science","volume":"19 5","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-02-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Improving the YOLOV7 Algorithm for Object Detection within Recorded Videos\",\"authors\":\"Asmaa Hasan Alrubaie, Maisa'a Abid Ali Khodhe, A. Abdulameer\",\"doi\":\"10.24996/ijs.2024.65.2.35\",\"DOIUrl\":null,\"url\":null,\"abstract\":\" Object detection algorithms play an important role in detecting people in surveillance videos. In recent years, with the rapid development of deep learning, the performance of object detection has improved by leaps and bounds, and the scheme of object detection by the YOLOV7 algorithm has also been born. Traditional object detection methods often fail to achieve a balance between speed and accuracy. To address these issues, in this research, an improved YOLOv7 algorithm performance is proposed to get the best speed-to-accuracy balance compared to state-of-the-art object detection within recorded videos using an effective compression method. This method calculates the difference between frames of video, and by using the zero difference approach by removing the duplicate frames from the recorded video and choosing only the meaningful frames based on many variables, including frame size, frame details, and the distance of the frames, influence the choice of a meaningful frame, and this will reduce the size of the video by eliminating the frames comparable to those chosen. \\n Additionally, any other datasets or pre-trained weights have not been used; YOLOv7 has been exclusively trained on the MS COCO dataset from scratch. In order to ensure the effectiveness of this approach, numerous detection systems are used in this work. Additionally, positive performance results to reduce the processing time required for object detection have been attained.\",\"PeriodicalId\":14698,\"journal\":{\"name\":\"Iraqi Journal of Science\",\"volume\":\"19 5\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-02-29\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Iraqi Journal of Science\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.24996/ijs.2024.65.2.35\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"Earth and Planetary Sciences\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Iraqi Journal of Science","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.24996/ijs.2024.65.2.35","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"Earth and Planetary Sciences","Score":null,"Total":0}
引用次数: 0
摘要
物体检测算法在监控视频中的人员检测中发挥着重要作用。近年来,随着深度学习的飞速发展,物体检测的性能有了突飞猛进的提高,YOLOV7算法的物体检测方案也应运而生。传统的物体检测方法往往无法实现速度与精度的平衡。为了解决这些问题,本研究提出了一种改进的 YOLOv7 算法性能,通过一种有效的压缩方法,在录制的视频中与最先进的物体检测方法相比,获得速度与精度的最佳平衡。该方法计算视频帧之间的差值,通过使用零差值方法,从录制的视频中删除重复帧,并根据帧大小、帧细节和帧间距离等影响有意义帧选择的多种变量,只选择有意义的帧,这样就可以通过删除与所选帧相似的帧来减少视频的大小。 此外,YOLOv7 没有使用任何其他数据集或预先训练好的权重;YOLOv7 完全是在 MS COCO 数据集上从头开始训练的。为了确保这种方法的有效性,在这项工作中使用了许多检测系统。此外,在减少物体检测所需的处理时间方面也取得了积极的成果。
Improving the YOLOV7 Algorithm for Object Detection within Recorded Videos
Object detection algorithms play an important role in detecting people in surveillance videos. In recent years, with the rapid development of deep learning, the performance of object detection has improved by leaps and bounds, and the scheme of object detection by the YOLOV7 algorithm has also been born. Traditional object detection methods often fail to achieve a balance between speed and accuracy. To address these issues, in this research, an improved YOLOv7 algorithm performance is proposed to get the best speed-to-accuracy balance compared to state-of-the-art object detection within recorded videos using an effective compression method. This method calculates the difference between frames of video, and by using the zero difference approach by removing the duplicate frames from the recorded video and choosing only the meaningful frames based on many variables, including frame size, frame details, and the distance of the frames, influence the choice of a meaningful frame, and this will reduce the size of the video by eliminating the frames comparable to those chosen.
Additionally, any other datasets or pre-trained weights have not been used; YOLOv7 has been exclusively trained on the MS COCO dataset from scratch. In order to ensure the effectiveness of this approach, numerous detection systems are used in this work. Additionally, positive performance results to reduce the processing time required for object detection have been attained.