首页 > 最新文献

2020 25th International Conference on Pattern Recognition (ICPR)最新文献

英文 中文
Temporal Pattern Detection in Time-Varying Graphical Models 时变图形模型中的时间模式检测
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9413203
Federico Tomasi, Veronica Tozzo, A. Barla
Graphical models allow to describe the interplay among variables of a system through a compact representation, suitable when relations evolve over time. For example, in a biological setting, genes interact differently depending on external environmental or metabolic factors. To incorporate this dynamics a viable strategy is to estimate a sequence of temporally related graphs assuming similarity among samples in different time points. While adjacent time points may direct the analysis towards a robust estimate of the underlying graph, the resulting model will not incorporate long-term or recurrent temporal relationships. In this work we propose a dynamical network inference model that leverages on kernels to consider general temporal patterns (such as circadian rhythms or seasonality). We show how our approach may also be exploited when the recurrent patterns are unknown, by coupling the network inference with a clustering procedure that detects possibly non-consecutive similar networks. Such clusters are then used to build similarity kernels. The convexity of the functional is determined by whether we impose or infer the kernel. In the first case, the optimisation algorithm exploits efficiently proximity operators with closed-form solutions. In the other case, we resort to an alternating minimisation procedure which jointly learns the temporal kernel and the underlying network. Extensive analysis on synthetic data shows the efficacy of our models compared to state-of-the-art methods. Finally, we applied our approach on two realworld applications to show how considering long-term patterns is fundamental to have insights on the behaviour of a complex system.
图形模型允许通过紧凑的表示来描述系统变量之间的相互作用,适用于关系随时间演变的情况。例如,在生物环境中,基因的相互作用取决于外部环境或代谢因素。为了整合这种动态,一个可行的策略是估计一系列时间相关的图,假设不同时间点的样本之间具有相似性。虽然相邻的时间点可能会将分析导向对底层图的可靠估计,但最终的模型将不包含长期或反复出现的时间关系。在这项工作中,我们提出了一个动态网络推理模型,该模型利用核来考虑一般的时间模式(如昼夜节律或季节性)。我们展示了如何在循环模式未知的情况下利用我们的方法,通过将网络推理与检测可能非连续相似网络的聚类过程相结合。然后使用这些聚类来构建相似核。函数的凸性取决于我们是否施加或推断核。在第一种情况下,优化算法有效地利用接近算子与封闭形式的解决方案。在另一种情况下,我们采用交替最小化过程,该过程联合学习时间核和底层网络。对合成数据的广泛分析表明,与最先进的方法相比,我们的模型更有效。最后,我们将我们的方法应用于两个实际应用程序,以说明考虑长期模式对于了解复杂系统的行为是多么重要。
{"title":"Temporal Pattern Detection in Time-Varying Graphical Models","authors":"Federico Tomasi, Veronica Tozzo, A. Barla","doi":"10.1109/ICPR48806.2021.9413203","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9413203","url":null,"abstract":"Graphical models allow to describe the interplay among variables of a system through a compact representation, suitable when relations evolve over time. For example, in a biological setting, genes interact differently depending on external environmental or metabolic factors. To incorporate this dynamics a viable strategy is to estimate a sequence of temporally related graphs assuming similarity among samples in different time points. While adjacent time points may direct the analysis towards a robust estimate of the underlying graph, the resulting model will not incorporate long-term or recurrent temporal relationships. In this work we propose a dynamical network inference model that leverages on kernels to consider general temporal patterns (such as circadian rhythms or seasonality). We show how our approach may also be exploited when the recurrent patterns are unknown, by coupling the network inference with a clustering procedure that detects possibly non-consecutive similar networks. Such clusters are then used to build similarity kernels. The convexity of the functional is determined by whether we impose or infer the kernel. In the first case, the optimisation algorithm exploits efficiently proximity operators with closed-form solutions. In the other case, we resort to an alternating minimisation procedure which jointly learns the temporal kernel and the underlying network. Extensive analysis on synthetic data shows the efficacy of our models compared to state-of-the-art methods. Finally, we applied our approach on two realworld applications to show how considering long-term patterns is fundamental to have insights on the behaviour of a complex system.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"1993 1","pages":"4481-4488"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82388610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
DeepBEV: A Conditional Adversarial Network for Bird's Eye View Generation DeepBEV:一种用于鸟瞰生成的条件对抗网络
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412516
Helmi Fraser, Sen Wang
Obtaining a meaningful, interpretable yet compact representation of the immediate surroundings of an autonomous vehicle is paramount for effective operation as well as safety. This paper proposes a solution to this by representing semantically important objects from a top-down, ego-centric bird's eye view. The novelty in this work is from formulating this problem as an adversarial learning task, tasking a generator model to produce bird's eye view representations which are plausible enough to be mistaken as a ground truth sample. This is achieved by using a Wasserstein Generative Adversarial Network based model conditioned on object detections from monocular RGB images and the corresponding bounding boxes. Extensive experiments show our model is more robust to novel data compared to strictly supervised benchmark models, while being a fraction of the size of the next best.
获得自动驾驶汽车周围环境的有意义的、可解释的、紧凑的表示对于有效运行和安全至关重要。本文提出了一种解决方案,通过自上而下、以自我为中心的鸟瞰图来表示语义上重要的对象。这项工作的新颖之处在于将这个问题表述为一个对抗性学习任务,分配一个生成器模型来生成鸟瞰图表示,这些表示似乎足够可信,可以被误认为是一个基本的真实样本。这是通过使用基于Wasserstein生成对抗网络的模型来实现的,该模型以单目RGB图像和相应的边界框的对象检测为条件。大量的实验表明,与严格监督的基准模型相比,我们的模型对新数据的鲁棒性更强,而规模只是次优模型的一小部分。
{"title":"DeepBEV: A Conditional Adversarial Network for Bird's Eye View Generation","authors":"Helmi Fraser, Sen Wang","doi":"10.1109/ICPR48806.2021.9412516","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412516","url":null,"abstract":"Obtaining a meaningful, interpretable yet compact representation of the immediate surroundings of an autonomous vehicle is paramount for effective operation as well as safety. This paper proposes a solution to this by representing semantically important objects from a top-down, ego-centric bird's eye view. The novelty in this work is from formulating this problem as an adversarial learning task, tasking a generator model to produce bird's eye view representations which are plausible enough to be mistaken as a ground truth sample. This is achieved by using a Wasserstein Generative Adversarial Network based model conditioned on object detections from monocular RGB images and the corresponding bounding boxes. Extensive experiments show our model is more robust to novel data compared to strictly supervised benchmark models, while being a fraction of the size of the next best.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"1 1","pages":"5581-5586"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82513165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Rethinking ReID: Multi-Feature Fusion Person Re-identification Based on Orientation Constraints 基于取向约束的多特征融合人物再识别
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9413212
M. Ai, Guozhi Shan, Bo Liu, Tianyan Liu
Person re-identification (ReID) aims to identify the specific pedestrian in a series of images or videos. Recently, ReID is receiving more and more attention in the fields of computer vision research and application like intelligent security. One major issue downgrading the ReID model performance lies in that various subjects in the same body orientations look too similar to distinguish by the model, while the same subject viewed in different orientations looks rather different. However, most of the current studies do not particularly differentiate pedestrians in orientation when designing the network, so we rethink this problem particularly from the perspective of person orientation and propose a new network structure by including two branches: one handling samples with the same body orientations and the other handling samples with different body orientations. Correspondingly, we also propose an orientation classifier that can accurately distinguish the orientation of each person. At the same time, the three-part loss functions are introduced for orientation constraint and combined to optimize the network simultaneously. Also, we use global and local features int the training stage in order to make use of multi-level information. Therefore, our network can derive its efficacy from orientation constraints and multiple features. Experiments show that our method not only has competitive performance on multiple datasets, but also can let retrieval results aligned with the orientation of the query sample rank higher, which may have great potential in the practical applications.
人物再识别(ReID)的目的是在一系列图像或视频中识别特定的行人。近年来,ReID在智能安防等计算机视觉研究和应用领域受到越来越多的关注。降低ReID模型性能的一个主要问题是,同一身体方向的不同受试者看起来太相似,无法被模型区分,而不同方向的同一受试者看起来却大相径庭。然而,目前大多数研究在设计网络时并没有特别区分行人的方向,因此我们特别从人的方向来思考这个问题,提出了一种新的网络结构,包括两个分支:一个处理具有相同身体方向的样本,另一个处理具有不同身体方向的样本。相应地,我们也提出了一个可以准确区分每个人的取向的定位分类器。同时,引入三分量损失函数作为方向约束,并将其结合起来进行网络同步优化。同时,我们在训练阶段使用了全局特征和局部特征,以充分利用多层次的信息。因此,我们的网络可以从方向约束和多特征中获得其有效性。实验表明,该方法不仅在多个数据集上具有较好的性能,而且可以使检索结果与查询样本的方向一致,排名更高,在实际应用中具有很大的潜力。
{"title":"Rethinking ReID: Multi-Feature Fusion Person Re-identification Based on Orientation Constraints","authors":"M. Ai, Guozhi Shan, Bo Liu, Tianyan Liu","doi":"10.1109/ICPR48806.2021.9413212","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9413212","url":null,"abstract":"Person re-identification (ReID) aims to identify the specific pedestrian in a series of images or videos. Recently, ReID is receiving more and more attention in the fields of computer vision research and application like intelligent security. One major issue downgrading the ReID model performance lies in that various subjects in the same body orientations look too similar to distinguish by the model, while the same subject viewed in different orientations looks rather different. However, most of the current studies do not particularly differentiate pedestrians in orientation when designing the network, so we rethink this problem particularly from the perspective of person orientation and propose a new network structure by including two branches: one handling samples with the same body orientations and the other handling samples with different body orientations. Correspondingly, we also propose an orientation classifier that can accurately distinguish the orientation of each person. At the same time, the three-part loss functions are introduced for orientation constraint and combined to optimize the network simultaneously. Also, we use global and local features int the training stage in order to make use of multi-level information. Therefore, our network can derive its efficacy from orientation constraints and multiple features. Experiments show that our method not only has competitive performance on multiple datasets, but also can let retrieval results aligned with the orientation of the query sample rank higher, which may have great potential in the practical applications.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"89 1","pages":"1904-1911"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82261466","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Recurrent-Convolutional Model for Automated Segmentation of Craniomaxillofacial CT Scans 颅颌面CT扫描自动分割的深度递归卷积模型
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9413084
Francesca Murabito, S. Palazzo, Federica Proietto Salanitri, F. Rundo, Ulas Bagci, D. Giordano, R. Leonardi, C. Spampinato
In this paper we define a deep learning architecture for automated segmentation of anatomical structures in Craniomaxillofacial (CMF) CT scans that leverages the recent success of encoder-decoder models for semantic segmentation of natural images. In particular, we propose a fully convolutional deep network that combines the advantages of recent fully convolutional models, such as Tiramisu, with squeeze-and-excitation blocks for feature recalibration, integrated with convolutional LSTMs to model spatio-temporal correlations between consecutive slices. The proposed segmentation network shows superior performance and generalization capabilities (to different structures and imaging modalities) than state of the art methods on automated segmentation of CMF structures (e.g., mandibles and airways) in several standard benchmarks (e.g., MICCAI datasets) and on new datasets proposed herein, effectively facing shape variability.
在本文中,我们定义了一个深度学习架构,用于颅颌面(CMF) CT扫描中解剖结构的自动分割,该架构利用了最近成功的编码器-解码器模型对自然图像进行语义分割。特别是,我们提出了一个全卷积深度网络,它结合了最近的全卷积模型(如Tiramisu)的优势,与用于特征重新校准的挤压和激励块相结合,与卷积lstm相结合,以模拟连续切片之间的时空相关性。在几个标准基准(例如MICCAI数据集)和本文提出的新数据集上,所提出的分割网络在CMF结构(例如,下颌骨和气道)的自动分割方面,比目前最先进的方法表现出更好的性能和泛化能力(针对不同的结构和成像模式),有效地面对形状变化。
{"title":"Deep Recurrent-Convolutional Model for Automated Segmentation of Craniomaxillofacial CT Scans","authors":"Francesca Murabito, S. Palazzo, Federica Proietto Salanitri, F. Rundo, Ulas Bagci, D. Giordano, R. Leonardi, C. Spampinato","doi":"10.1109/ICPR48806.2021.9413084","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9413084","url":null,"abstract":"In this paper we define a deep learning architecture for automated segmentation of anatomical structures in Craniomaxillofacial (CMF) CT scans that leverages the recent success of encoder-decoder models for semantic segmentation of natural images. In particular, we propose a fully convolutional deep network that combines the advantages of recent fully convolutional models, such as Tiramisu, with squeeze-and-excitation blocks for feature recalibration, integrated with convolutional LSTMs to model spatio-temporal correlations between consecutive slices. The proposed segmentation network shows superior performance and generalization capabilities (to different structures and imaging modalities) than state of the art methods on automated segmentation of CMF structures (e.g., mandibles and airways) in several standard benchmarks (e.g., MICCAI datasets) and on new datasets proposed herein, effectively facing shape variability.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"7 1","pages":"9062-9067"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82359493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Attention Based Coupled Framework for Road and Pothole Segmentation 基于注意力的道路和坑洞分割耦合框架
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412368
Shaik Masihullah, Ritu Garg, Prerana Mukherjee, Anupama Ray
In this paper, we propose a novel attention based coupled framework for road and pothole segmentation. In many developing countries as well as in rural areas, the drivable areas are neither well-defined, nor well-maintained. Under such circumstances, an Advance Driver Assistant System (ADAS) is needed to assess the drivable area and alert about the potholes ahead to ensure vehicle safety. Moreover, this information can also be used in structured environments for assessment and maintenance of road health. We demonstrate few-shot learning approach for pothole detection to leverage accuracy even with fewer training samples. We report the exhaustive experimental results for road segmentation on KITTI and IDD datasets. We also present pothole segmentation on IDD.
在本文中,我们提出了一种新的基于注意力的道路和坑洞分割耦合框架。在许多发展中国家以及农村地区,可驾驶区域既没有明确界定,也没有得到良好的维护。在这种情况下,需要高级驾驶辅助系统(ADAS)来评估可行驶区域,并提醒前方的坑洼,以确保车辆安全。此外,这些信息还可用于结构化环境,以评估和维护道路健康。我们展示了少量的学习方法,以坑检测利用精度,即使在更少的训练样本。我们报告了在KITTI和IDD数据集上进行道路分割的详尽实验结果。我们还提出了IDD的凹坑分割。
{"title":"Attention Based Coupled Framework for Road and Pothole Segmentation","authors":"Shaik Masihullah, Ritu Garg, Prerana Mukherjee, Anupama Ray","doi":"10.1109/ICPR48806.2021.9412368","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412368","url":null,"abstract":"In this paper, we propose a novel attention based coupled framework for road and pothole segmentation. In many developing countries as well as in rural areas, the drivable areas are neither well-defined, nor well-maintained. Under such circumstances, an Advance Driver Assistant System (ADAS) is needed to assess the drivable area and alert about the potholes ahead to ensure vehicle safety. Moreover, this information can also be used in structured environments for assessment and maintenance of road health. We demonstrate few-shot learning approach for pothole detection to leverage accuracy even with fewer training samples. We report the exhaustive experimental results for road segmentation on KITTI and IDD datasets. We also present pothole segmentation on IDD.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"491 1","pages":"5812-5819"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78835217","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
GAP: Quantifying the Generative Adversarial Set and Class Feature Applicability of Deep Neural Networks 量化深度神经网络的生成对抗集和类特征适用性
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412665
Edward Collier, S. Mukhopadhyay
Recent work in deep neural networks has sought to characterize the nature in which a network learns features and how applicable learnt features are to various problem sets. Deep neural network applicability can be split into three sub-problems; set applicability, class applicability, and instance applicability. In this work we seek to quantify the applicability of features learned during adversarial training, focusing specifically on set and class applicability. We apply techniques for measuring applicability to both generators and discriminators trained on various data sets to quantify applicability.
最近在深度神经网络方面的工作试图描述网络学习特征的性质,以及学习到的特征如何适用于各种问题集。深度神经网络的适用性可分为三个子问题;设置适用性、类适用性和实例适用性。在这项工作中,我们试图量化在对抗训练中学习到的特征的适用性,特别关注集合和类的适用性。我们将测量适用性的技术应用于在各种数据集上训练的生成器和鉴别器,以量化适用性。
{"title":"GAP: Quantifying the Generative Adversarial Set and Class Feature Applicability of Deep Neural Networks","authors":"Edward Collier, S. Mukhopadhyay","doi":"10.1109/ICPR48806.2021.9412665","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412665","url":null,"abstract":"Recent work in deep neural networks has sought to characterize the nature in which a network learns features and how applicable learnt features are to various problem sets. Deep neural network applicability can be split into three sub-problems; set applicability, class applicability, and instance applicability. In this work we seek to quantify the applicability of features learned during adversarial training, focusing specifically on set and class applicability. We apply techniques for measuring applicability to both generators and discriminators trained on various data sets to quantify applicability.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"37 1","pages":"8384-8391"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78946401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Feature Extraction and Selection via Robust Discriminant Analysis and Class Sparsity 基于鲁棒判别分析和类稀疏性的特征提取与选择
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412683
A. Khoder, F. Dornaika
The main goal of discriminant embedding is to extract features that can be compact and informative representations of the original set of features. This paper introduces a hybrid scheme for linear feature extraction for supervised multiclass classification. We introduce a unifying criterion that is able to retain the advantages of robust sparse LDA and Interclass sparsity. Thus, the estimated transformation includes two types of discrimination which are the inter-class sparsity and robust Linear Discriminant Analysis with feature selection. In order to optimize the proposed objective function, we deploy an iterative alternating minimization scheme for estimating the linear transformation and the orthogonal matrix. The introduced scheme is generic in the sense that it can be used for combining and tuning many other linear embedding methods. In the lights of the experiments conducted on six image datasets including faces, objects, and digits, the proposed scheme was able to outperform competing methods in most of the cases.
判别嵌入的主要目标是提取特征,这些特征可以是原始特征集的紧凑且信息丰富的表示。介绍了一种用于有监督多类分类的线性特征提取的混合方案。我们引入了一个统一的准则,能够保留鲁棒稀疏LDA和类间稀疏性的优点。因此,估计变换包括两种类型的判别:类间稀疏性和带特征选择的鲁棒线性判别分析。为了优化所提出的目标函数,我们采用迭代交替最小化方案来估计线性变换和正交矩阵。所介绍的方案具有通用性,可用于组合和调整许多其他线性嵌入方法。在包括人脸、物体和数字在内的6个图像数据集上进行的实验表明,该方案在大多数情况下都能够优于竞争方法。
{"title":"Feature Extraction and Selection via Robust Discriminant Analysis and Class Sparsity","authors":"A. Khoder, F. Dornaika","doi":"10.1109/ICPR48806.2021.9412683","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412683","url":null,"abstract":"The main goal of discriminant embedding is to extract features that can be compact and informative representations of the original set of features. This paper introduces a hybrid scheme for linear feature extraction for supervised multiclass classification. We introduce a unifying criterion that is able to retain the advantages of robust sparse LDA and Interclass sparsity. Thus, the estimated transformation includes two types of discrimination which are the inter-class sparsity and robust Linear Discriminant Analysis with feature selection. In order to optimize the proposed objective function, we deploy an iterative alternating minimization scheme for estimating the linear transformation and the orthogonal matrix. The introduced scheme is generic in the sense that it can be used for combining and tuning many other linear embedding methods. In the lights of the experiments conducted on six image datasets including faces, objects, and digits, the proposed scheme was able to outperform competing methods in most of the cases.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"16 1","pages":"7258-7264"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76377770","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
One step clustering based on a-contrario framework for detection of alterations in historical violins 基于反向框架的一步聚类检测历史小提琴的变化
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412129
Alireza Rezaei, S. L. Hégarat-Mascle, Emanuel Aldea, Piercarlo Dondi, M. Malagodi
Preventive conservation is an important practice in Cultural Heritage. The constant monitoring of the state of conservation of an artwork helps us reduce the risk of damage and number of necessary interventions. In this work, we propose a probabilistic approach for the detection of alterations on the surface of historical violins based on an a-contrario framework. Our method is a one step NFA clustering solution which considers grey-level and spatial density information in one background model. The proposed method is robust to noise and avoids parameter tuning and any assumption about the quantity of the worn-out areas. We have used as input UV induced fluorescence (UVIFL) images for considering details not perceivable with visible light. Tests were conducted on image sequences included in the “Violins UVIFL imagery” dataset. Results illustrate the ability of the algorithm to distinguish the worn area from the surrounding regions. Comparisons with state-of-the-art clustering methods show improved overall precision and recall.
预防性保护是文化遗产保护的重要实践。对艺术品保存状况的持续监测有助于我们减少损坏的风险和必要干预的数量。在这项工作中,我们提出了一种基于反向框架的概率方法来检测历史小提琴表面的变化。我们的方法是在一个背景模型中考虑灰度和空间密度信息的一步NFA聚类解决方案。该方法对噪声具有较强的鲁棒性,避免了参数调整和对磨损区域数量的任何假设。我们使用了作为输入的紫外线诱导荧光(UVIFL)图像来考虑可见光无法感知的细节。对“小提琴UVIFL图像”数据集中包含的图像序列进行了测试。结果表明,该算法能够将磨损区域与周围区域区分开来。与最先进的聚类方法的比较显示出总体精度和召回率的提高。
{"title":"One step clustering based on a-contrario framework for detection of alterations in historical violins","authors":"Alireza Rezaei, S. L. Hégarat-Mascle, Emanuel Aldea, Piercarlo Dondi, M. Malagodi","doi":"10.1109/ICPR48806.2021.9412129","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412129","url":null,"abstract":"Preventive conservation is an important practice in Cultural Heritage. The constant monitoring of the state of conservation of an artwork helps us reduce the risk of damage and number of necessary interventions. In this work, we propose a probabilistic approach for the detection of alterations on the surface of historical violins based on an a-contrario framework. Our method is a one step NFA clustering solution which considers grey-level and spatial density information in one background model. The proposed method is robust to noise and avoids parameter tuning and any assumption about the quantity of the worn-out areas. We have used as input UV induced fluorescence (UVIFL) images for considering details not perceivable with visible light. Tests were conducted on image sequences included in the “Violins UVIFL imagery” dataset. Results illustrate the ability of the algorithm to distinguish the worn area from the surrounding regions. Comparisons with state-of-the-art clustering methods show improved overall precision and recall.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"69 1","pages":"9348-9355"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76531316","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Hyperspectral Imaging for Analysis and Classification of Plastic Waste 塑料垃圾的高光谱成像分析与分类
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412737
Jakub Kraśniewski, Lukasz Dabala, M. Lewandowski
Environmental protection is one of the main challenges facing society nowadays. Even with constantly growing awareness, not all of the sorting can be done by people themselves - the differences between materials are not visible to the human eye. For that reason, we present the use of a hyperspectral camera as a capture device, which allows us to obtain the full spectrum of the material. In this work we propose a method for efficient recognition of the substance of an item. We conducted several experiments and analysis of the spectra of different materials in different conditions on a special measuring stand. That enabled identification of the best features, which can later be used during classification, which was confirmed during the extensive testing procedure.
环境保护是当今社会面临的主要挑战之一。尽管人们的意识在不断增强,但并不是所有的分类都可以由人们自己完成——材料之间的差异是肉眼看不见的。因此,我们提出使用高光谱相机作为捕获设备,这使我们能够获得材料的全光谱。在这项工作中,我们提出了一种有效识别项目实质的方法。我们在专门的测量台上对不同材料在不同条件下的光谱进行了多次实验和分析。这样可以识别出最佳特征,这些特征可以在以后的分类过程中使用,这在广泛的测试过程中得到了证实。
{"title":"Hyperspectral Imaging for Analysis and Classification of Plastic Waste","authors":"Jakub Kraśniewski, Lukasz Dabala, M. Lewandowski","doi":"10.1109/ICPR48806.2021.9412737","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412737","url":null,"abstract":"Environmental protection is one of the main challenges facing society nowadays. Even with constantly growing awareness, not all of the sorting can be done by people themselves - the differences between materials are not visible to the human eye. For that reason, we present the use of a hyperspectral camera as a capture device, which allows us to obtain the full spectrum of the material. In this work we propose a method for efficient recognition of the substance of an item. We conducted several experiments and analysis of the spectra of different materials in different conditions on a special measuring stand. That enabled identification of the best features, which can later be used during classification, which was confirmed during the extensive testing procedure.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"24 1","pages":"4805-4812"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76173581","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Classification of spatially enriched pixel time series with convolutional neural networks 基于卷积神经网络的空间丰富像素时间序列分类
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412892
Mohamed Chelali, Camille Kurtz, A. Puissant, N. Vincent
Satellite Image Time Series (SITS), MRI sequences, and more generally image time series, constitute $2D+t$ data providing spatial and temporal information about an observed scene. Given a pattern recognition task such as image classification, considering jointly such rich information is crucial during the decision process. Nevertheless, due to the complex representation of the data-cube, spatio-temporal features extraction from $2D+t$ data remains difficult to handle. We present in this article an approach to learn such features from this data, and then to proceed to their classification. Our strategy consists in enriching pixel time series with spatial information. It is based on Random Walk to build a novel segment-based representation of the data, passing from a $2D+t$ dimension to a $2D$ one, without loosing too much spatial information. Such new representation is then involved in an end-to-end learning process with a classical 2D Convolutional Neural Network (CNN) in order to learn spatiotemporal features for the classification of image time series. Our approach is evaluated on a remote sensing application for the mapping of agricultural crops. Thanks to a visual attention mechanism, the proposed $2D$ spatio-temporal representation makes also easier the interpretation of a SITS to understand spatiotemporal phenomenons related to soil management practices.
卫星图像时间序列(sit)、MRI序列和更普遍的图像时间序列构成了2D+t数据,提供了观测场景的空间和时间信息。对于像图像分类这样的模式识别任务,综合考虑这些丰富的信息在决策过程中至关重要。然而,由于数据立方体的复杂表示,从$2D+ $ t数据中提取时空特征仍然难以处理。在本文中,我们提出了一种从这些数据中学习这些特征的方法,然后对它们进行分类。我们的策略是用空间信息丰富像素时间序列。它基于随机漫步来构建一种新的基于片段的数据表示,从$2D+ $ t维度传递到$2D维度,而不会丢失太多的空间信息。然后使用经典的2D卷积神经网络(CNN)进行端到端学习过程,以学习用于图像时间序列分类的时空特征。我们的方法在农业作物测绘的遥感应用中进行了评估。由于视觉注意机制,所提出的2D时空表征也使解释sit更容易理解与土壤管理实践相关的时空现象。
{"title":"Classification of spatially enriched pixel time series with convolutional neural networks","authors":"Mohamed Chelali, Camille Kurtz, A. Puissant, N. Vincent","doi":"10.1109/ICPR48806.2021.9412892","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412892","url":null,"abstract":"Satellite Image Time Series (SITS), MRI sequences, and more generally image time series, constitute $2D+t$ data providing spatial and temporal information about an observed scene. Given a pattern recognition task such as image classification, considering jointly such rich information is crucial during the decision process. Nevertheless, due to the complex representation of the data-cube, spatio-temporal features extraction from $2D+t$ data remains difficult to handle. We present in this article an approach to learn such features from this data, and then to proceed to their classification. Our strategy consists in enriching pixel time series with spatial information. It is based on Random Walk to build a novel segment-based representation of the data, passing from a $2D+t$ dimension to a $2D$ one, without loosing too much spatial information. Such new representation is then involved in an end-to-end learning process with a classical 2D Convolutional Neural Network (CNN) in order to learn spatiotemporal features for the classification of image time series. Our approach is evaluated on a remote sensing application for the mapping of agricultural crops. Thanks to a visual attention mechanism, the proposed $2D$ spatio-temporal representation makes also easier the interpretation of a SITS to understand spatiotemporal phenomenons related to soil management practices.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"4 1","pages":"5310-5317"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87487170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2020 25th International Conference on Pattern Recognition (ICPR)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1