首页 > 最新文献

2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)最新文献

英文 中文
Attribute-Based Facial Image Manipulation on Latent Space 基于隐空间属性的人脸图像处理
Chien-Hung Lin, Yiyun Pan, Ja-Ling Wu
Using machine learning to generate images has become more mature, especially the images produced using a Generative Adversarial Network. Unfortunately, the complicated architecture of those models makes it difficult for us to ensure the output images’ diversity and controllability without introducing little embarrassment in implementation. Therefore, some researchers try to edit the latent codes generated by a given learning model directly on the latent space for manipulating the output image by simply inputting the new latent codes into the original model without changing the model’s structure and learned parameters. However, the methods mentioned above faced the problems that the size of latent space cannot be too large or the trouble-some of features entanglement. In this work, we propose an approach to conquer the problems mentioned above, which is to compress the original latent space to better the applicability and usability of the methods limited by the size of the latent space. Compared with the existing methods, this method can be applied to more models and still reach the target of image manipulation.
使用机器学习生成图像已经变得更加成熟,特别是使用生成对抗网络生成的图像。不幸的是,这些模型的复杂架构使得我们很难在保证输出图像的多样性和可控性的同时,在实现上不带来一点尴尬。因此,一些研究人员试图在不改变模型结构和学习参数的情况下,将给定学习模型生成的潜码直接在潜空间上进行编辑,以便对输出图像进行操作。然而,上述方法都面临着潜在空间不能太大或特征纠缠的麻烦等问题。在这项工作中,我们提出了一种解决上述问题的方法,即压缩原始潜在空间,以提高受潜在空间大小限制的方法的适用性和可用性。与现有方法相比,该方法可以应用于更多的模型,并且仍然达到了图像处理的目标。
{"title":"Attribute-Based Facial Image Manipulation on Latent Space","authors":"Chien-Hung Lin, Yiyun Pan, Ja-Ling Wu","doi":"10.1109/AVSS52988.2021.9663845","DOIUrl":"https://doi.org/10.1109/AVSS52988.2021.9663845","url":null,"abstract":"Using machine learning to generate images has become more mature, especially the images produced using a Generative Adversarial Network. Unfortunately, the complicated architecture of those models makes it difficult for us to ensure the output images’ diversity and controllability without introducing little embarrassment in implementation. Therefore, some researchers try to edit the latent codes generated by a given learning model directly on the latent space for manipulating the output image by simply inputting the new latent codes into the original model without changing the model’s structure and learned parameters. However, the methods mentioned above faced the problems that the size of latent space cannot be too large or the trouble-some of features entanglement. In this work, we propose an approach to conquer the problems mentioned above, which is to compress the original latent space to better the applicability and usability of the methods limited by the size of the latent space. Compared with the existing methods, this method can be applied to more models and still reach the target of image manipulation.","PeriodicalId":246327,"journal":{"name":"2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124927304","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Bridging the Invisible and Visible World: Translation between RGB and IR Images through Contour Cycle GAN 桥接不可见和可见的世界:通过轮廓循环GAN在RGB和IR图像之间的转换
Yawen Lu, G. Lu
Infrared Radiation (IR) images that capture the emitted IR signals from surrounding environment have been widely applied to pedestrian detection and video surveillance. However, there are not many textures that appeared on thermal images as compared to RGB images, which brings enormous challenges and difficulties in various tasks. Visible images cannot capture scenes in the dark and night environment due to the lack of light. In this paper, we propose a Contour GAN-based framework to learn the cross-domain representation and also map IR images with visible images. In contrast to existing structures of image translation that focus on spectral consistency, our framework also introduces strong spatial constraints, with further spectral enhancement by illuminance contrast and consistency constraints. Designating our method for IR and RGB image translation, it can generate high-quality translated images. Extensive experiments on near IR (NIR) and far IR (thermal) datasets demonstrate superior performance for quantitative and visual results.
红外图像捕获周围环境发出的红外信号,已广泛应用于行人检测和视频监控中。然而,与RGB图像相比,热图像上出现的纹理并不多,这给各种任务带来了巨大的挑战和困难。由于光线不足,可见光图像无法捕捉黑暗和夜间环境中的场景。在本文中,我们提出了一个基于轮廓gan的框架来学习跨域表示,并将红外图像与可见图像进行映射。与现有的专注于光谱一致性的图像平移结构不同,我们的框架还引入了强空间约束,通过照度对比和一致性约束进一步增强光谱。将我们的方法用于红外和RGB图像的翻译,可以生成高质量的翻译图像。在近红外(NIR)和远红外(热)数据集上进行的大量实验表明,在定量和视觉结果方面具有优越的性能。
{"title":"Bridging the Invisible and Visible World: Translation between RGB and IR Images through Contour Cycle GAN","authors":"Yawen Lu, G. Lu","doi":"10.1109/AVSS52988.2021.9663750","DOIUrl":"https://doi.org/10.1109/AVSS52988.2021.9663750","url":null,"abstract":"Infrared Radiation (IR) images that capture the emitted IR signals from surrounding environment have been widely applied to pedestrian detection and video surveillance. However, there are not many textures that appeared on thermal images as compared to RGB images, which brings enormous challenges and difficulties in various tasks. Visible images cannot capture scenes in the dark and night environment due to the lack of light. In this paper, we propose a Contour GAN-based framework to learn the cross-domain representation and also map IR images with visible images. In contrast to existing structures of image translation that focus on spectral consistency, our framework also introduces strong spatial constraints, with further spectral enhancement by illuminance contrast and consistency constraints. Designating our method for IR and RGB image translation, it can generate high-quality translated images. Extensive experiments on near IR (NIR) and far IR (thermal) datasets demonstrate superior performance for quantitative and visual results.","PeriodicalId":246327,"journal":{"name":"2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129859257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Drone-vs-Bird Detection Challenge at IEEE AVSS2021 IEEE AVSS2021无人机对鸟探测挑战赛
A. Coluccia, A. Fascista, Arne Schumann, L. Sommer, A. Dimou, D. Zarpalas, F. C. Akyon, Ogulcan Eryuksel, Kamil Anil Ozfuttu, S. Altinuc, Fardad Dadboud, Vaibhav Patel, Varun Mehta, M. Bolic, I. Mantegh
This paper presents the 4-th edition of the “drone-vs-bird” detection challenge, launched in conjunction with the the 17-th IEEE International Conference on Advanced Video and Signal-based Surveillance (AVSS). The objective of the challenge is to tackle the problem of detecting the presence of one or more drones in video scenes where birds may suddenly appear, taking into account some important effects such as the background and foreground motion. The proposed solutions should identify and localize drones in the scene only when they are actually present, without being confused by the presence of birds and the dynamic nature of the captured scenes. The paper illustrates the results of the challenge on the 2021 dataset, which has been further extended compared to the previous edition run in 2020.
本文介绍了与第17届IEEE高级视频和基于信号的监控(AVSS)国际会议一起推出的第4版“无人机对鸟”检测挑战。该挑战的目标是解决在鸟类可能突然出现的视频场景中检测一个或多个无人机存在的问题,同时考虑到一些重要的影响,如背景和前景运动。所提出的解决方案应该只有在无人机实际存在时才能识别和定位场景中的无人机,而不会被鸟类的存在和捕获场景的动态特性所迷惑。本文说明了2021年数据集的挑战结果,与2020年运行的上一版本相比,该数据集得到了进一步扩展。
{"title":"Drone-vs-Bird Detection Challenge at IEEE AVSS2021","authors":"A. Coluccia, A. Fascista, Arne Schumann, L. Sommer, A. Dimou, D. Zarpalas, F. C. Akyon, Ogulcan Eryuksel, Kamil Anil Ozfuttu, S. Altinuc, Fardad Dadboud, Vaibhav Patel, Varun Mehta, M. Bolic, I. Mantegh","doi":"10.1109/AVSS52988.2021.9663844","DOIUrl":"https://doi.org/10.1109/AVSS52988.2021.9663844","url":null,"abstract":"This paper presents the 4-th edition of the “drone-vs-bird” detection challenge, launched in conjunction with the the 17-th IEEE International Conference on Advanced Video and Signal-based Surveillance (AVSS). The objective of the challenge is to tackle the problem of detecting the presence of one or more drones in video scenes where birds may suddenly appear, taking into account some important effects such as the background and foreground motion. The proposed solutions should identify and localize drones in the scene only when they are actually present, without being confused by the presence of birds and the dynamic nature of the captured scenes. The paper illustrates the results of the challenge on the 2021 dataset, which has been further extended compared to the previous edition run in 2020.","PeriodicalId":246327,"journal":{"name":"2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)","volume":"104 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127954465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
[AVSS 2021 Front cover] [AVSS 2021封面]
{"title":"[AVSS 2021 Front cover]","authors":"","doi":"10.1109/avss52988.2021.9663795","DOIUrl":"https://doi.org/10.1109/avss52988.2021.9663795","url":null,"abstract":"","PeriodicalId":246327,"journal":{"name":"2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129251010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Performance of Crowd-Specific Detectors in Multi-Pedestrian Tracking 人群特定检测器在多行人跟踪中的性能研究
Daniel Stadler, J. Beyerer
In recent years, several methods and datasets have been proposed to push the performance of pedestrian detection in crowded scenarios. In this study, three crowd-specific detectors are combined with a general tracking-by-detection approach to evaluate their applicability in multi-pedestrian tracking. Investigating the relation between detection and tracking accuracy, we make the interesting observation that in spite of a high detection capability, the performance in tracking can be poor and analyze the reasons behind that. However, one of the examined approaches can significantly boost the tracking performance on two benchmarks under different training configurations. It is shown that combining crowd-specific detectors with a simple tracking pipeline can achieve promising results, especially in challenging scenes with heavy occlusion. Although our tracker only relies on motion cues and no visual information is considered, applying the strong detections from the crowd-specific model, state-of-the-art results on the challenging MOT17 and MOT20 benchmarks are obtained.
近年来,人们提出了几种方法和数据集来提高拥挤场景下行人检测的性能。在本研究中,将三种特定人群的检测器与一般的逐检测跟踪方法相结合,以评估它们在多行人跟踪中的适用性。通过研究检测和跟踪精度之间的关系,我们发现了一个有趣的现象,即尽管检测能力很高,但跟踪性能却很差,并分析了其背后的原因。然而,其中一种方法可以在不同的训练配置下显著提高两个基准的跟踪性能。研究表明,将人群特定检测器与简单的跟踪管道相结合可以取得很好的效果,特别是在具有严重遮挡的挑战性场景中。虽然我们的跟踪器只依赖于运动线索,没有考虑视觉信息,但应用来自人群特定模型的强检测,在具有挑战性的MOT17和MOT20基准上获得了最先进的结果。
{"title":"On the Performance of Crowd-Specific Detectors in Multi-Pedestrian Tracking","authors":"Daniel Stadler, J. Beyerer","doi":"10.1109/AVSS52988.2021.9663836","DOIUrl":"https://doi.org/10.1109/AVSS52988.2021.9663836","url":null,"abstract":"In recent years, several methods and datasets have been proposed to push the performance of pedestrian detection in crowded scenarios. In this study, three crowd-specific detectors are combined with a general tracking-by-detection approach to evaluate their applicability in multi-pedestrian tracking. Investigating the relation between detection and tracking accuracy, we make the interesting observation that in spite of a high detection capability, the performance in tracking can be poor and analyze the reasons behind that. However, one of the examined approaches can significantly boost the tracking performance on two benchmarks under different training configurations. It is shown that combining crowd-specific detectors with a simple tracking pipeline can achieve promising results, especially in challenging scenes with heavy occlusion. Although our tracker only relies on motion cues and no visual information is considered, applying the strong detections from the crowd-specific model, state-of-the-art results on the challenging MOT17 and MOT20 benchmarks are obtained.","PeriodicalId":246327,"journal":{"name":"2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)","volume":"394 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115310435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Deep Learning for Body Parts Detection using HRNet and EfficientNet 基于HRNet和EfficientNet的身体部位检测深度学习
Miniar Ben Gamra, M. Akhloufi, Chunpeng Wang, Shuo Liu
Human body parts detection is an important field of research in computer vision. It can serve as an essential tool in surveillance systems and used to automatically detect and moderate non-appropriate online content such as nudity, child pornography, violence, etc. In this work, we introduce a novel two-step framework to define ten body parts using joints localization. A new architecture with EfficientNet as a backbone is proposed and compared to HRNet for the first step of pose estimation. The resulting joints are then used as an input to the second step, where a set of rules is applied to connect the appropriate joints and to define each body part. The developed algorithms were tested using MPII human pose benchmark. The proposed approach achieved a very interesting performance with a 90.13% Probability of Correct Keypoint (PCK) for the pose estimation and an average of 89.80% of mean Average Precision (mAP) for the body parts detection.
人体部位检测是计算机视觉的一个重要研究领域。它可以作为监视系统中的重要工具,用于自动发现和调节不适当的在线内容,如裸体、儿童色情、暴力等。在这项工作中,我们引入了一个新的两步框架,利用关节定位来定义十个身体部位。提出了一种以effentnet为骨干的新体系结构,并将其与HRNet进行了第一步姿态估计的比较。然后将得到的关节用作第二步的输入,在第二步中,应用一组规则来连接适当的关节并定义每个身体部位。采用MPII人体姿态基准对所开发的算法进行了测试。该方法取得了非常有趣的性能,姿态估计的正确关键点概率(PCK)为90.13%,身体部位检测的平均平均精度(mAP)为89.80%。
{"title":"Deep Learning for Body Parts Detection using HRNet and EfficientNet","authors":"Miniar Ben Gamra, M. Akhloufi, Chunpeng Wang, Shuo Liu","doi":"10.1109/AVSS52988.2021.9663785","DOIUrl":"https://doi.org/10.1109/AVSS52988.2021.9663785","url":null,"abstract":"Human body parts detection is an important field of research in computer vision. It can serve as an essential tool in surveillance systems and used to automatically detect and moderate non-appropriate online content such as nudity, child pornography, violence, etc. In this work, we introduce a novel two-step framework to define ten body parts using joints localization. A new architecture with EfficientNet as a backbone is proposed and compared to HRNet for the first step of pose estimation. The resulting joints are then used as an input to the second step, where a set of rules is applied to connect the appropriate joints and to define each body part. The developed algorithms were tested using MPII human pose benchmark. The proposed approach achieved a very interesting performance with a 90.13% Probability of Correct Keypoint (PCK) for the pose estimation and an average of 89.80% of mean Average Precision (mAP) for the body parts detection.","PeriodicalId":246327,"journal":{"name":"2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116824354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Seismic Sensor based Human Activity Recognition Framework using Deep Learning 基于深度学习的地震传感器人体活动识别框架
Priyanka Choudhary, Neeraj Goel, Mukesh Saini
Activity recognition has gained attention due to the rapid development of microelectromechanical sensors. Numerous human-centric applications in healthcare, security, and smart environments can benefit from an efficient human activity recognition system. In this paper, we demonstrate the use of a seismic sensor for human activity recognition. Traditionally, researchers have relied on handcrafted features to identify the target activity, but these features may be inefficient in complex and noisy environments. The proposed framework employs an autoencoder to map the activity into a compact representative descriptor. Further, an Artificial Neural Network (ANN) classifier is trained on the extracted descriptors. We compare the proposed framework with multiple machine learning classifiers and a state-of-the-art framework on different evaluation metrics. On 5-fold cross-validation, the proposed approach outperforms the state-of-the-art in terms of precision and recall by an average of 10.68 and 23.36%, respectively. We also collected a dataset to assess the efficacy of the proposed seismic sensor-based activity recognition. The dataset is collected in a variety of challenging environments, such as variable grass length, soil moisture content, and the passing of unwanted vehicles nearby.
由于微机电传感器的迅速发展,活动识别受到了广泛的关注。在医疗保健、安全和智能环境中,许多以人为中心的应用程序都可以从高效的人类活动识别系统中受益。在本文中,我们演示了地震传感器对人类活动识别的使用。传统上,研究人员依靠手工制作的特征来识别目标活动,但这些特征在复杂和嘈杂的环境中可能效率低下。提出的框架使用一个自动编码器将活动映射到一个紧凑的代表性描述符。进一步,在提取的描述符上训练人工神经网络(ANN)分类器。我们将提出的框架与多个机器学习分类器和不同评估指标的最先进框架进行比较。在5倍交叉验证中,该方法的准确率和召回率平均分别高出10.68和23.36%。我们还收集了一个数据集来评估所提出的基于地震传感器的活动识别的有效性。该数据集是在各种具有挑战性的环境中收集的,例如可变的草长、土壤含水量以及附近不需要的车辆经过。
{"title":"A Seismic Sensor based Human Activity Recognition Framework using Deep Learning","authors":"Priyanka Choudhary, Neeraj Goel, Mukesh Saini","doi":"10.1109/AVSS52988.2021.9663747","DOIUrl":"https://doi.org/10.1109/AVSS52988.2021.9663747","url":null,"abstract":"Activity recognition has gained attention due to the rapid development of microelectromechanical sensors. Numerous human-centric applications in healthcare, security, and smart environments can benefit from an efficient human activity recognition system. In this paper, we demonstrate the use of a seismic sensor for human activity recognition. Traditionally, researchers have relied on handcrafted features to identify the target activity, but these features may be inefficient in complex and noisy environments. The proposed framework employs an autoencoder to map the activity into a compact representative descriptor. Further, an Artificial Neural Network (ANN) classifier is trained on the extracted descriptors. We compare the proposed framework with multiple machine learning classifiers and a state-of-the-art framework on different evaluation metrics. On 5-fold cross-validation, the proposed approach outperforms the state-of-the-art in terms of precision and recall by an average of 10.68 and 23.36%, respectively. We also collected a dataset to assess the efficacy of the proposed seismic sensor-based activity recognition. The dataset is collected in a variety of challenging environments, such as variable grass length, soil moisture content, and the passing of unwanted vehicles nearby.","PeriodicalId":246327,"journal":{"name":"2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123889405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Introspective Closed-Loop Perception for Energy-efficient Sensors 节能传感器的内省闭环感知
Kruttidipta Samal, M. Wolf, S. Mukhopadhyay
Task-driven closed-loop perception-sensing systems have shown considerable energy savings over traditional open-loop systems. Prior works on such systems have used simple feedback signals such as object detections and tracking which led to poor perception quality. This paper proposes an improved approach based on perceptual risk. First, a method is proposed to estimate the risk of failure to detect a target of interest. The risk estimate is used as a signal in a feedback system to determine how sensor resources are utilized. Two feedback algorithms are proposed: one based on proportional/integral methods and the other based on 0/1 (bang-bang) methods. These feedback algorithms are compared based on the efficiency with which they use available sensor resources as well as their absolute detection rates. Experiments on two real-world autonomous driving datasets show that the proposed system has better object detection recall and lower marginal cost of prediction than prior work.
与传统的开环系统相比,任务驱动的闭环感知传感系统显示出相当大的能量节约。在这类系统上,先前的工作使用简单的反馈信号,如物体检测和跟踪,导致感知质量差。本文提出了一种基于感知风险的改进方法。首先,提出了一种估计目标检测失败风险的方法。在反馈系统中,风险评估作为一个信号来决定如何利用传感器资源。提出了两种反馈算法:一种基于比例/积分法,另一种基于0/1 (bang-bang)法。这些反馈算法是基于效率,他们利用可用的传感器资源以及他们的绝对检测率进行比较。在两个真实自动驾驶数据集上的实验表明,该系统具有更好的目标检测召回率和更低的预测边际成本。
{"title":"Introspective Closed-Loop Perception for Energy-efficient Sensors","authors":"Kruttidipta Samal, M. Wolf, S. Mukhopadhyay","doi":"10.1109/AVSS52988.2021.9663801","DOIUrl":"https://doi.org/10.1109/AVSS52988.2021.9663801","url":null,"abstract":"Task-driven closed-loop perception-sensing systems have shown considerable energy savings over traditional open-loop systems. Prior works on such systems have used simple feedback signals such as object detections and tracking which led to poor perception quality. This paper proposes an improved approach based on perceptual risk. First, a method is proposed to estimate the risk of failure to detect a target of interest. The risk estimate is used as a signal in a feedback system to determine how sensor resources are utilized. Two feedback algorithms are proposed: one based on proportional/integral methods and the other based on 0/1 (bang-bang) methods. These feedback algorithms are compared based on the efficiency with which they use available sensor resources as well as their absolute detection rates. Experiments on two real-world autonomous driving datasets show that the proposed system has better object detection recall and lower marginal cost of prediction than prior work.","PeriodicalId":246327,"journal":{"name":"2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116459629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
PETS2021: Through-foliage detection and tracking challenge and evaluation PETS2021:穿透叶探测和跟踪挑战与评估
Luis Patino, Jonathan N. Boyle, J. Ferryman, Jonas Auer, Julian Pegoraro, R. Pflugfelder, Mertcan Cokbas, J. Konrad, P. Ishwar, Giulia Slavic, L. Marcenaro, Yifan Jiang, Youngsaeng Jin, Hanseok Ko, Guangliang Zhao, Guy Ben-Yosef, Jianwei Qiu
This paper presents the outcomes of the PETS2021 challenge held in conjunction with AVSS2021 and sponsored by the EU FOLDOUT project. The challenge comprises the publication of a novel video surveillance dataset on through-foliage detection, the defined challenges addressing person detection and tracking in fragmented occlusion scenarios, and quantitative and qualitative performance evaluation of challenge results submitted by six worldwide participants. The results show that while several detection and tracking methods achieve overall good results, through-foliage detection and tracking remains a challenging task for surveillance systems especially as it serves as the input to behaviour (threat) recognition.
本文介绍了由欧盟FOLDOUT项目赞助的与AVSS2021一起举行的PETS2021挑战的结果。挑战包括发布一个关于树叶检测的新型视频监控数据集,解决碎片遮挡场景中人员检测和跟踪的定义挑战,以及对六个全球参与者提交的挑战结果进行定量和定性的性能评估。结果表明,虽然几种检测和跟踪方法总体上取得了良好的效果,但对于监视系统来说,穿透叶检测和跟踪仍然是一项具有挑战性的任务,特别是当它作为行为(威胁)识别的输入时。
{"title":"PETS2021: Through-foliage detection and tracking challenge and evaluation","authors":"Luis Patino, Jonathan N. Boyle, J. Ferryman, Jonas Auer, Julian Pegoraro, R. Pflugfelder, Mertcan Cokbas, J. Konrad, P. Ishwar, Giulia Slavic, L. Marcenaro, Yifan Jiang, Youngsaeng Jin, Hanseok Ko, Guangliang Zhao, Guy Ben-Yosef, Jianwei Qiu","doi":"10.1109/AVSS52988.2021.9663837","DOIUrl":"https://doi.org/10.1109/AVSS52988.2021.9663837","url":null,"abstract":"This paper presents the outcomes of the PETS2021 challenge held in conjunction with AVSS2021 and sponsored by the EU FOLDOUT project. The challenge comprises the publication of a novel video surveillance dataset on through-foliage detection, the defined challenges addressing person detection and tracking in fragmented occlusion scenarios, and quantitative and qualitative performance evaluation of challenge results submitted by six worldwide participants. The results show that while several detection and tracking methods achieve overall good results, through-foliage detection and tracking remains a challenging task for surveillance systems especially as it serves as the input to behaviour (threat) recognition.","PeriodicalId":246327,"journal":{"name":"2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130080733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Fire Detection Model Based on Tiny-YOLOv3 with Hyperparameters Improvement 基于超参数改进的Tiny-YOLOv3火灾探测模型
Zeineb Daoud, A. B. Hamida, C. Amar
Fires are the most devastating disasters that the world can face. Thereby, it is crucial to exactly identify fire areas in video surveillance scenes, to overcome the shortcomings of the existing fire detection methods. Recently, deep learning models have been widely used for fire recognition applications. Indeed, a novel deep fire detection method is introduced in this paper. An improved fire model based on tiny-YOLOv3 (You Only Look Once version 3) network is developed in order to enhance the detection accuracy. The main idea is the tiny-YOLOv3 improvement according to the refined proposed training hyperparameters. The generated model is trained and evaluated on the constructed and manually labeled dataset. Results show that applying the proposed training heuristics with the tiny-YOLOv3 network improves the fire detection performance with 81.65% of mean Average Precision (mAP). Also, the designed model outperforms the related works with a detection precision of 97.6%.
火灾是世界面临的最具破坏性的灾难。因此,在视频监控场景中准确识别火灾区域,克服现有火灾探测方法的不足至关重要。近年来,深度学习模型被广泛应用于火灾识别领域。实际上,本文提出了一种新的深火探测方法。为了提高探测精度,开发了基于tiny-YOLOv3 (You Only Look Once version 3)网络的改进火灾模型。主要思想是根据提出的训练超参数进行微小的yolov3改进。生成的模型在构建和手动标记的数据集上进行训练和评估。结果表明,将所提出的训练启发式算法应用于tiny-YOLOv3网络,可以将火灾探测性能提高到平均精度(mAP)的81.65%。同时,该模型的检测精度达到97.6%,优于相关工作。
{"title":"A Fire Detection Model Based on Tiny-YOLOv3 with Hyperparameters Improvement","authors":"Zeineb Daoud, A. B. Hamida, C. Amar","doi":"10.1109/AVSS52988.2021.9663822","DOIUrl":"https://doi.org/10.1109/AVSS52988.2021.9663822","url":null,"abstract":"Fires are the most devastating disasters that the world can face. Thereby, it is crucial to exactly identify fire areas in video surveillance scenes, to overcome the shortcomings of the existing fire detection methods. Recently, deep learning models have been widely used for fire recognition applications. Indeed, a novel deep fire detection method is introduced in this paper. An improved fire model based on tiny-YOLOv3 (You Only Look Once version 3) network is developed in order to enhance the detection accuracy. The main idea is the tiny-YOLOv3 improvement according to the refined proposed training hyperparameters. The generated model is trained and evaluated on the constructed and manually labeled dataset. Results show that applying the proposed training heuristics with the tiny-YOLOv3 network improves the fire detection performance with 81.65% of mean Average Precision (mAP). Also, the designed model outperforms the related works with a detection precision of 97.6%.","PeriodicalId":246327,"journal":{"name":"2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116134207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1