首页 > 最新文献

2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)最新文献

英文 中文
Near-Lossless Coding of Plenoptic Camera Sensor Images for Archiving Light Field Array of Views 全光相机传感器图像存档的近无损编码
E. Palma, I. Tabus
In this paper we propose a near-lossless encoder for sensor images acquired by plenoptic cameras, and we investigate its usage for encoding in an archive all information needed for reconstructing high quality versions of the light field (LF) array of views(AoV). The near-lossless encoding of the plenoptic camera sensor image is realized by a modified version of the recently published sparse relevant regressors and contexts (SRRC) encoder. The lossy reconstruction is obtained in two nested loops: the outer one operates over the sensor image patches (each patch corresponding to a microlens image), and the inner loop operates over the pixels in the patch. In the latter, we enforce the SRRC predictors to use the already reconstructed lossy version of the sensor image. Then, we examine the usage of the near-lossless SRRC (NL-SRRC) codec as a building block for an archiving scheme including all information needed for running the plenoptic processing pipeline and obtaining the LF-AoV. Finally, we replace in the archiving scheme the NL-SRRC codec with other state of the art lossy codecs and compare the results, which show that NL-SRRC based archiving scheme achieves better performance for the range of high bitrates.
在本文中,我们提出了一种近乎无损的编码器,用于全光学相机获取的传感器图像,并研究了它在档案中编码重建高质量版本光场(LF)视图阵列(AoV)所需的所有信息的用途。全光相机传感器图像的近无损编码是通过对最近发表的稀疏相关回归和上下文(SRRC)编码器的改进实现的。有损重构是在两个嵌套的循环中获得的:外层循环对传感器图像补丁(每个补丁对应一个微透镜图像)进行操作,内层循环对补丁中的像素进行操作。在后者中,我们强制SRRC预测器使用已经重建的传感器图像的有损版本。然后,我们研究了近无损SRRC (NL-SRRC)编解码器作为归档方案的构建块的使用,包括运行全光处理管道和获得LF-AoV所需的所有信息。最后,我们将NL-SRRC编解码器替换为其他最先进的有损编解码器,并对结果进行了比较,结果表明基于NL-SRRC的归档方案在高比特率范围内具有更好的性能。
{"title":"Near-Lossless Coding of Plenoptic Camera Sensor Images for Archiving Light Field Array of Views","authors":"E. Palma, I. Tabus","doi":"10.1109/IPTA54936.2022.9784151","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784151","url":null,"abstract":"In this paper we propose a near-lossless encoder for sensor images acquired by plenoptic cameras, and we investigate its usage for encoding in an archive all information needed for reconstructing high quality versions of the light field (LF) array of views(AoV). The near-lossless encoding of the plenoptic camera sensor image is realized by a modified version of the recently published sparse relevant regressors and contexts (SRRC) encoder. The lossy reconstruction is obtained in two nested loops: the outer one operates over the sensor image patches (each patch corresponding to a microlens image), and the inner loop operates over the pixels in the patch. In the latter, we enforce the SRRC predictors to use the already reconstructed lossy version of the sensor image. Then, we examine the usage of the near-lossless SRRC (NL-SRRC) codec as a building block for an archiving scheme including all information needed for running the plenoptic processing pipeline and obtaining the LF-AoV. Finally, we replace in the archiving scheme the NL-SRRC codec with other state of the art lossy codecs and compare the results, which show that NL-SRRC based archiving scheme achieves better performance for the range of high bitrates.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"147 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124430325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Scale Space Radon Transform for Non Overlapping Thick Ellipses Detection 非重叠厚椭圆检测的尺度空间Radon变换
A. Goumeidane, D. Ziou, Nafaa Nacereddine
This paper presents a new elliptical structure detection method, combining the advantages of the multiscale Hessian, and the scale space Radon transform (SSRT) for an ellipse. The advantage of the former is twofold: highlighting the lines defining the ellipses present in the image and reducing the search space for these ellipses in the SSRT space, which will discard the false SSRT maxima. The subsequent application of the SSRT permits, in turn, to alleviate the computation load and to obtain, moreover, a good detection of thick ellipses when they are not threadlike. Experiments carried out on synthetic and real images have shown good detection of thick ellipses, with low computational overhead compared to the Elliptical Radon transform.
结合椭圆的多尺度Hessian和尺度空间Radon变换(SSRT)的优点,提出了一种新的椭圆结构检测方法。前者的优点是双重的:突出显示定义图像中存在的椭圆的线条,并减少这些椭圆在SSRT空间中的搜索空间,这将丢弃假的SSRT最大值。SSRT的后续应用反过来又可以减轻计算负荷,并且可以很好地检测非线状的厚椭圆。在合成图像和真实图像上进行的实验表明,与椭圆Radon变换相比,该方法具有较好的厚椭圆检测效果,且计算量较低。
{"title":"Scale Space Radon Transform for Non Overlapping Thick Ellipses Detection","authors":"A. Goumeidane, D. Ziou, Nafaa Nacereddine","doi":"10.1109/IPTA54936.2022.9784129","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784129","url":null,"abstract":"This paper presents a new elliptical structure detection method, combining the advantages of the multiscale Hessian, and the scale space Radon transform (SSRT) for an ellipse. The advantage of the former is twofold: highlighting the lines defining the ellipses present in the image and reducing the search space for these ellipses in the SSRT space, which will discard the false SSRT maxima. The subsequent application of the SSRT permits, in turn, to alleviate the computation load and to obtain, moreover, a good detection of thick ellipses when they are not threadlike. Experiments carried out on synthetic and real images have shown good detection of thick ellipses, with low computational overhead compared to the Elliptical Radon transform.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121063571","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
On the relevance of edge-conditioned convolution for GNN-based semantic image segmentation using spatial relationships 基于空间关系的gnn语义图像分割中边缘条件卷积的相关性研究
P. Coupeau, Jean-Baptiste Fasquel, M. Dinomais
This paper addresses the fundamental task of semantic image segmentation by exploiting structural information (spatial relationships between image regions). To perform such task, we propose to combine a deep neural network (CNN) with inexact “many-to-one-or-none” graph matching where graphs encode efficiently class probabilities a nd structural information related to regions segmented by the CNN. In order to achieve node classification, a basic 2 -layer graph neural network (GNN) based on the edge-conditioned convolution operator (ECConv), managing both node and edge attributes, is considered. Prelim-inary experiments are performed on both a synthetic dataset and a public dataset of face images (FASSEG). Our approach is shown to be resilient to small training datasets that often limit the performance of deep learning thanks to a preprocessing task of graph coarsening. Results show that the proposal reaches a perfect accuracy on synthetic dataset and improves performance of the CNN by 6% (bounding box dice index) on FASSEG. Moreover, it enhances by 27% the initial Hausdorff distance (i.e. with CNN only) using the entire training dataset and by 41% with only 75% of training samples.
本文通过利用结构信息(图像区域之间的空间关系)来解决语义图像分割的基本任务。为了完成这样的任务,我们提出将深度神经网络(CNN)与非精确的“多对一或无”图匹配结合起来,其中图有效地编码了与CNN分割的区域相关的分类概率和结构信息。为了实现节点分类,考虑了一种基于边缘条件卷积算子(ECConv)的基本2层图神经网络(GNN),同时管理节点和边缘属性。在人脸图像的合成数据集和公共数据集(FASSEG)上进行了初步实验。我们的方法被证明对小型训练数据集具有弹性,由于图粗化的预处理任务,这些数据集通常限制了深度学习的性能。结果表明,该方法在合成数据集上达到了很好的准确率,并将CNN在FASSEG上的性能提高了6%(边界盒骰子指数)。此外,它使用整个训练数据集将初始Hausdorff距离(即仅使用CNN)增强了27%,仅使用75%的训练样本将初始Hausdorff距离增强了41%。
{"title":"On the relevance of edge-conditioned convolution for GNN-based semantic image segmentation using spatial relationships","authors":"P. Coupeau, Jean-Baptiste Fasquel, M. Dinomais","doi":"10.1109/IPTA54936.2022.9784143","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784143","url":null,"abstract":"This paper addresses the fundamental task of semantic image segmentation by exploiting structural information (spatial relationships between image regions). To perform such task, we propose to combine a deep neural network (CNN) with inexact “many-to-one-or-none” graph matching where graphs encode efficiently class probabilities a nd structural information related to regions segmented by the CNN. In order to achieve node classification, a basic 2 -layer graph neural network (GNN) based on the edge-conditioned convolution operator (ECConv), managing both node and edge attributes, is considered. Prelim-inary experiments are performed on both a synthetic dataset and a public dataset of face images (FASSEG). Our approach is shown to be resilient to small training datasets that often limit the performance of deep learning thanks to a preprocessing task of graph coarsening. Results show that the proposal reaches a perfect accuracy on synthetic dataset and improves performance of the CNN by 6% (bounding box dice index) on FASSEG. Moreover, it enhances by 27% the initial Hausdorff distance (i.e. with CNN only) using the entire training dataset and by 41% with only 75% of training samples.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128180085","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Special Session 3: Visual Computing in Digital Humanities 专题会议3:数字人文学科中的视觉计算
{"title":"Special Session 3: Visual Computing in Digital Humanities","authors":"","doi":"10.1109/ipta54936.2022.9784150","DOIUrl":"https://doi.org/10.1109/ipta54936.2022.9784150","url":null,"abstract":"","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114230944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Draco-Based Selective Crypto-Compression Method of 3D objects 基于draco的3D对象选择性加密压缩方法
Bianca Janse van Rensburg, W. Puech, Jean-Pierre Pedeboy
3D objects play an essential part in many differ-ent domains. They are considered to be important assets and therefore need to be secured. Sometimes, certain users may have the right to access only a part of the 3D object, for example the shape of the object, but not the content. Recently, selective encryption schemes have been proposed in order to allow this. 3D objects are often stored on the cloud or need to be rendered in real time, and therefore compression is important. In this paper, we propose a selective crypto-compression method for 3D objects. This is based on the Draco compression method, which was developed by Google for 3D objects. Our method is format compliant, and to the best of our knowledge, we are the first to propose a selective cry to-compression method for 3D objects.
3D对象在许多不同的领域发挥着重要作用。它们被认为是重要的资产,因此需要得到保护。有时,某些用户可能只有权访问3D对象的一部分,例如对象的形状,而不是内容。最近,为了实现这一点,提出了选择性加密方案。3D对象通常存储在云中或需要实时渲染,因此压缩很重要。本文提出了一种针对三维物体的选择性加密压缩方法。这是基于Draco压缩方法,这是由谷歌开发的3D对象。我们的方法是格式兼容的,并且据我们所知,我们是第一个提出3D对象的选择性呼叫压缩方法的人。
{"title":"Draco-Based Selective Crypto-Compression Method of 3D objects","authors":"Bianca Janse van Rensburg, W. Puech, Jean-Pierre Pedeboy","doi":"10.1109/IPTA54936.2022.9784116","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784116","url":null,"abstract":"3D objects play an essential part in many differ-ent domains. They are considered to be important assets and therefore need to be secured. Sometimes, certain users may have the right to access only a part of the 3D object, for example the shape of the object, but not the content. Recently, selective encryption schemes have been proposed in order to allow this. 3D objects are often stored on the cloud or need to be rendered in real time, and therefore compression is important. In this paper, we propose a selective crypto-compression method for 3D objects. This is based on the Draco compression method, which was developed by Google for 3D objects. Our method is format compliant, and to the best of our knowledge, we are the first to propose a selective cry to-compression method for 3D objects.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"192 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114838654","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Complex Texture Features Learned by Applying Randomized Neural Network on Graphs 基于随机神经网络的复杂纹理特征学习
Kallil M. C. Zielinski, L. C. Ribas, Leonardo F. S. Scabini, O. Bruno
Since the 1960s, texture has become one of the most-studied visual attribute of images for analysis and classification tasks. Among many different approaches such as statistical, spectral, structural and model-based, there are also methods that rely on analyzing the image complexity and learning techniques. These recent approaches are receiving attention for its promising results in the past few years. This paper proposes a method that combines complex networks and randomized neural networks. In the proposed approach, the texture image is modeled as a complex network, and the information measures obtained from the topological properties of the network are then used to train the RNN in order to learn a representation of the modeled image. Our proposal has proven to perform well in comparison to other literature approaches in two different texture databases. Our method also achieved a high performance in a very challenging biological problem of plant species recognition. Thus, the method is a promising option for different tasks of image analysis.
自20世纪60年代以来,纹理已成为研究最多的图像视觉属性之一,用于分析和分类任务。在统计、光谱、结构和基于模型的方法中,也有依赖于分析图像复杂性和学习技术的方法。这些最近的方法在过去几年中因其令人鼓舞的结果而受到关注。本文提出了一种将复杂网络与随机神经网络相结合的方法。在该方法中,将纹理图像建模为一个复杂网络,然后使用从网络拓扑属性中获得的信息度量来训练RNN,以学习建模图像的表示。在两个不同的纹理数据库中,与其他文献方法相比,我们的建议已经被证明表现良好。我们的方法在植物物种识别这一极具挑战性的生物学问题上也取得了很高的性能。因此,该方法对于不同的图像分析任务是一个很有前途的选择。
{"title":"Complex Texture Features Learned by Applying Randomized Neural Network on Graphs","authors":"Kallil M. C. Zielinski, L. C. Ribas, Leonardo F. S. Scabini, O. Bruno","doi":"10.1109/IPTA54936.2022.9784123","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784123","url":null,"abstract":"Since the 1960s, texture has become one of the most-studied visual attribute of images for analysis and classification tasks. Among many different approaches such as statistical, spectral, structural and model-based, there are also methods that rely on analyzing the image complexity and learning techniques. These recent approaches are receiving attention for its promising results in the past few years. This paper proposes a method that combines complex networks and randomized neural networks. In the proposed approach, the texture image is modeled as a complex network, and the information measures obtained from the topological properties of the network are then used to train the RNN in order to learn a representation of the modeled image. Our proposal has proven to perform well in comparison to other literature approaches in two different texture databases. Our method also achieved a high performance in a very challenging biological problem of plant species recognition. Thus, the method is a promising option for different tasks of image analysis.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"131 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114362360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A complex network approach for fish species recognition based on otolith shape 基于耳石形状的复杂网络鱼种识别方法
L. C. Ribas, Leonardo F. S. Scabini, O. Bruno
Fish otolith recognition is an essential task to study the evolution and food chains in paleontological and ecological sciences. One of the approaches to this problem is to automatically analyze the shape of otolith contour present in images. In this paper, we explore a state-of-the-art shape analysis method called “angular descriptors of complex networks (ADCN)” applied to the classification of otolith images for fish species recognition. The ADCN method models the otolith contour as a graph, or complex network, and computes angular properties from its connections for shape characterization. The ADCN method is evaluated in an otolith image dataset composed of 14 fish species from three families. Up to 95.71% of accuracy is achieved, which surpasses other literature methods and confirms that the ADCN method can be an important tool for such biological problems.
鱼类耳石识别是古生物学和生态学研究鱼类进化和食物链的一项重要任务。解决这一问题的方法之一是对图像中存在的耳石轮廓形状进行自动分析。在本文中,我们探索了一种最先进的形状分析方法,称为“复杂网络的角描述子(ADCN)”,应用于鱼类种类识别的耳石图像分类。ADCN方法将耳石轮廓建模为一个图或复杂网络,并从其连接计算角度特性以进行形状表征。在由3科14种鱼类组成的耳石图像数据集中对ADCN方法进行了评估。准确率高达95.71%,超过了其他文献方法,证实了ADCN方法可以成为解决此类生物学问题的重要工具。
{"title":"A complex network approach for fish species recognition based on otolith shape","authors":"L. C. Ribas, Leonardo F. S. Scabini, O. Bruno","doi":"10.1109/IPTA54936.2022.9784114","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784114","url":null,"abstract":"Fish otolith recognition is an essential task to study the evolution and food chains in paleontological and ecological sciences. One of the approaches to this problem is to automatically analyze the shape of otolith contour present in images. In this paper, we explore a state-of-the-art shape analysis method called “angular descriptors of complex networks (ADCN)” applied to the classification of otolith images for fish species recognition. The ADCN method models the otolith contour as a graph, or complex network, and computes angular properties from its connections for shape characterization. The ADCN method is evaluated in an otolith image dataset composed of 14 fish species from three families. Up to 95.71% of accuracy is achieved, which surpasses other literature methods and confirms that the ADCN method can be an important tool for such biological problems.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130534150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Oral Session 3 口头会议3
{"title":"Oral Session 3","authors":"","doi":"10.1109/ipta54936.2022.9784131","DOIUrl":"https://doi.org/10.1109/ipta54936.2022.9784131","url":null,"abstract":"","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116300380","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards Fast and Accurate Intimate Contact Recognition through Video Analysis 通过视频分析实现快速准确的亲密接触识别
Yuhao Luo, Hengjing Zhang, Hengchang Liu
Intimate contact recognition has gained more attention in academia field in recent years due to the outbreak of Covid-19. However, state of the art solutions suffer from either inefficient accuracy or high cost. In this paper, we propose a novel method for COVID-19 intimate contact recognition in public spaces through video camera networks (CCTV). This method leverages distance detection and re-Identification algorithms, so pedestrians in close contact are re-identified, their identity information is obtained and stored in a database to realize contact tracing. We compare different social distance detection algorithms and the Faster-RCNN model outperforms other al-ternatives in terms of running speed. We also evaluate our Re-Identification model on two types of indicators in the PETS2009 dataset: mAP reaches 85.1%; rank-1, rank-5, and rank-10 reach 97.8%, 98.9%, and 98.9%, respectively. Experimental results demonstrate that our solution can be effectively applied in public places to realize fast and accurate automatic contact tracing.
近年来,新型冠状病毒感染症(Covid-19)的爆发,引起了学术界的广泛关注。然而,最先进的解决方案要么精度低,要么成本高。本文提出了一种利用闭路电视网络(CCTV)识别公共空间COVID-19亲密接触者的新方法。该方法利用距离检测和再识别算法,对近距离接触的行人进行再识别,获取其身份信息并存储在数据库中,实现接触追踪。我们比较了不同的社交距离检测算法,fast - rcnn模型在运行速度方面优于其他替代算法。在PETS2009数据集的两类指标上对我们的再识别模型进行了评价:mAP达到85.1%;Rank-1, rank-5, rank-10分别达到97.8%,98.9%,98.9%。实验结果表明,该方法可以有效地应用于公共场所,实现快速、准确的自动接触追踪。
{"title":"Towards Fast and Accurate Intimate Contact Recognition through Video Analysis","authors":"Yuhao Luo, Hengjing Zhang, Hengchang Liu","doi":"10.1109/IPTA54936.2022.9784118","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784118","url":null,"abstract":"Intimate contact recognition has gained more attention in academia field in recent years due to the outbreak of Covid-19. However, state of the art solutions suffer from either inefficient accuracy or high cost. In this paper, we propose a novel method for COVID-19 intimate contact recognition in public spaces through video camera networks (CCTV). This method leverages distance detection and re-Identification algorithms, so pedestrians in close contact are re-identified, their identity information is obtained and stored in a database to realize contact tracing. We compare different social distance detection algorithms and the Faster-RCNN model outperforms other al-ternatives in terms of running speed. We also evaluate our Re-Identification model on two types of indicators in the PETS2009 dataset: mAP reaches 85.1%; rank-1, rank-5, and rank-10 reach 97.8%, 98.9%, and 98.9%, respectively. Experimental results demonstrate that our solution can be effectively applied in public places to realize fast and accurate automatic contact tracing.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114606186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
End-to-End Video Snapshot Compressive Imaging using Video Transformers 端到端视频快照压缩成像使用视频变压器
Wael Saideni, F. Courrèges, D. Helbert, J. Cances
This paper presents a novel reconstruction algorithm for video Snapshot Compressive Imaging (SCI). Inspired by recent research works on Transformers and Self-Attention mechanism in computer vision, we propose the first video SCI reconstruction algorithm built upon Transformers to capture long-range spatio-temporal dependencies enabling the deep learning of feature maps. Our approach is based on a Spatiotempo-ral Convolutional Multi-head Attention (ST-ConvMHA) which enable to exploit the spatial and temporal information of the video scenes instead of using fully-connected attention layers. To evaluate the performances of our approach, we train our algorithm on DAVIS2017 dataset and we test the trained models on six benchmark datasets. The obtained results in terms of PSNR, SSIM and especially reconstruction time prove the ability of using our reconstruction approach for real-time applications. We truly believe that our research will motivate future works for more video reconstruction approaches.
提出了一种新的视频快照压缩成像(SCI)重构算法。受最近计算机视觉中变形金刚和自注意机制的研究工作的启发,我们提出了第一个基于变形金刚的视频SCI重建算法,以捕获远程时空依赖关系,实现特征图的深度学习。我们的方法是基于时空卷积多头注意(ST-ConvMHA),它能够利用视频场景的空间和时间信息,而不是使用完全连接的注意层。为了评估我们的方法的性能,我们在DAVIS2017数据集上训练我们的算法,并在六个基准数据集上测试训练好的模型。在PSNR, SSIM,特别是重构时间方面的结果证明了我们的重构方法在实时应用中的能力。我们真的相信我们的研究将激励更多的视频重建方法的未来工作。
{"title":"End-to-End Video Snapshot Compressive Imaging using Video Transformers","authors":"Wael Saideni, F. Courrèges, D. Helbert, J. Cances","doi":"10.1109/IPTA54936.2022.9784128","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784128","url":null,"abstract":"This paper presents a novel reconstruction algorithm for video Snapshot Compressive Imaging (SCI). Inspired by recent research works on Transformers and Self-Attention mechanism in computer vision, we propose the first video SCI reconstruction algorithm built upon Transformers to capture long-range spatio-temporal dependencies enabling the deep learning of feature maps. Our approach is based on a Spatiotempo-ral Convolutional Multi-head Attention (ST-ConvMHA) which enable to exploit the spatial and temporal information of the video scenes instead of using fully-connected attention layers. To evaluate the performances of our approach, we train our algorithm on DAVIS2017 dataset and we test the trained models on six benchmark datasets. The obtained results in terms of PSNR, SSIM and especially reconstruction time prove the ability of using our reconstruction approach for real-time applications. We truly believe that our research will motivate future works for more video reconstruction approaches.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122072721","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1