首页 > 最新文献

2020 25th International Conference on Pattern Recognition (ICPR)最新文献

英文 中文
Learning Semantic Representations via Joint 3D Face Reconstruction and Facial Attribute Estimation 基于联合三维人脸重建和人脸属性估计的语义表征学习
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412426
Zichun Weng, Youjun Xiang, Xianfeng Li, Juntao Liang, W. Huo, Yuli Fu
We propose a novel joint framework for 3D face reconstruction (3DFR) that integrates facial attribute estimation (FAE) as an auxiliary task. One of the essential problems of 3DFR is to extract semantic facial features (e.g., Big Nose, High Cheekbones, and Asian) from in-the-wild 2D images, which is inherently involved with FAE. These two tasks, though heterogeneous, are highly relevant to each other. To achieve this, we leverage a Convolutional Neural Network to extract shared facial representations for both shape decoder and attribute classifier. We further develop an in-batch hybrid-task training scheme that enables our model to learn from heterogeneous facial datasets jointly within a mini-batch. Thanks to the joint loss that provides supervision from both 3DFR and FAE domains, our model learns the correlations between 3D shapes and facial attributes, which benefit both feature extraction and shape inference. Quantitative evaluation and qualitative visualization results confirm the effectiveness and robustness of our joint framework.
本文提出了一种将人脸属性估计(FAE)作为辅助任务的三维人脸重建联合框架。3DFR的基本问题之一是从野外2D图像中提取语义面部特征(如大鼻子、高颧骨和亚洲人),这本身就涉及到FAE。这两项任务虽然不同,但彼此高度相关。为了实现这一点,我们利用卷积神经网络为形状解码器和属性分类器提取共享的面部表示。我们进一步开发了一种批内混合任务训练方案,使我们的模型能够在一个小批内共同从异构面部数据集中学习。由于联合损失提供了来自3DFR和FAE域的监督,我们的模型学习了3D形状和面部属性之间的相关性,这有利于特征提取和形状推理。定量评价和定性可视化结果证实了我们联合框架的有效性和鲁棒性。
{"title":"Learning Semantic Representations via Joint 3D Face Reconstruction and Facial Attribute Estimation","authors":"Zichun Weng, Youjun Xiang, Xianfeng Li, Juntao Liang, W. Huo, Yuli Fu","doi":"10.1109/ICPR48806.2021.9412426","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412426","url":null,"abstract":"We propose a novel joint framework for 3D face reconstruction (3DFR) that integrates facial attribute estimation (FAE) as an auxiliary task. One of the essential problems of 3DFR is to extract semantic facial features (e.g., Big Nose, High Cheekbones, and Asian) from in-the-wild 2D images, which is inherently involved with FAE. These two tasks, though heterogeneous, are highly relevant to each other. To achieve this, we leverage a Convolutional Neural Network to extract shared facial representations for both shape decoder and attribute classifier. We further develop an in-batch hybrid-task training scheme that enables our model to learn from heterogeneous facial datasets jointly within a mini-batch. Thanks to the joint loss that provides supervision from both 3DFR and FAE domains, our model learns the correlations between 3D shapes and facial attributes, which benefit both feature extraction and shape inference. Quantitative evaluation and qualitative visualization results confirm the effectiveness and robustness of our joint framework.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"41 1","pages":"9696-9702"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73532162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RobusterNet: Improving Copy-Move Forgery Detection with Volterra-based Convolutions RobusterNet:利用基于volterra的卷积改进Copy-Move伪造检测
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412587
Efthimia Kafali, N. Vretos, T. Semertzidis, P. Daras
Convolutional Neural Networks (CNNs) have recently been introduced for addressing copy-move forgery detection (CMFD). However, current CMFD CNN-based approaches have insufficient performance commitment regarding the localization of the positive class. In this paper, this issue is explored by considering both linear and nonlinear interactions between pixels. A nonlinear Inception module based on second-order Volterra kernels is proposed, in order to ameliorate the results of a state-of-the-art CMFD architecture. The outcome of this work shows that a combination of linear and nonlinear convolution kernels can make the input foreground and background pixels more separable. The proposed approach is evaluated on CASIA and CoMoFoD, two publicly available CMFD datasets, and results to an improved positive class localization performance. Moreover, the findings of the proposed method imply that the nonlinear Inception module stimulates immense robustness against miscellaneous post processing attacks.
卷积神经网络(cnn)最近被引入到解决复制-移动伪造检测(CMFD)。然而,目前基于CMFD cnn的方法在正类定位方面没有足够的性能承诺。本文通过考虑像素之间的线性和非线性相互作用来探讨这个问题。提出了一种基于二阶Volterra核的非线性启始模块,以改善最先进的CMFD体系结构的结果。这项工作的结果表明,线性和非线性卷积核的组合可以使输入前景和背景像素更加可分。在CASIA和CoMoFoD这两个公开可用的CMFD数据集上对所提出的方法进行了评估,结果表明该方法提高了正向类定位性能。此外,所提出的方法的研究结果表明,非线性Inception模块对各种后处理攻击具有巨大的鲁棒性。
{"title":"RobusterNet: Improving Copy-Move Forgery Detection with Volterra-based Convolutions","authors":"Efthimia Kafali, N. Vretos, T. Semertzidis, P. Daras","doi":"10.1109/ICPR48806.2021.9412587","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412587","url":null,"abstract":"Convolutional Neural Networks (CNNs) have recently been introduced for addressing copy-move forgery detection (CMFD). However, current CMFD CNN-based approaches have insufficient performance commitment regarding the localization of the positive class. In this paper, this issue is explored by considering both linear and nonlinear interactions between pixels. A nonlinear Inception module based on second-order Volterra kernels is proposed, in order to ameliorate the results of a state-of-the-art CMFD architecture. The outcome of this work shows that a combination of linear and nonlinear convolution kernels can make the input foreground and background pixels more separable. The proposed approach is evaluated on CASIA and CoMoFoD, two publicly available CMFD datasets, and results to an improved positive class localization performance. Moreover, the findings of the proposed method imply that the nonlinear Inception module stimulates immense robustness against miscellaneous post processing attacks.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"34 1","pages":"1160-1165"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73561109","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Coherence and Identity Learning for Arbitrary-length Face Video Generation 任意长度人脸视频生成的一致性和同一性学习
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412380
Shuquan Ye, Chu Han, Jiaying Lin, Guoqiang Han, Shengfeng He
Face synthesis is an interesting yet challenging task in computer vision. It is even much harder to generate a portrait video than a single image. In this paper, we propose a novel video generation framework for synthesizing arbitrary-length face videos without any face exemplar or landmark. To overcome the synthesis ambiguity of face video, we propose a divide-and-conquer strategy to separately address the video face synthesis problem from two aspects, face identity synthesis and rearrangement. To this end, we design a cascaded network which contains three components, Identity-aware GAN (IA-GAN), Face Coherence Network, and Interpolation Network. IA-GAN is proposed to synthesize photorealistic faces with the same identity from a set of noises. Face Coherence Network is designed to re-arrange the faces generated by IA-GAN while keeping the inter-frame coherence. Interpolation Network is introduced to eliminate the discontinuity between two adjacent frames and improve the smoothness of the face video. Experimental results demonstrate that our proposed network is able to generate face video with high visual quality while preserving the identity. Statistics show that our method outperforms state-of-the-art unconditional face video generative models in multiple challenging datasets.
人脸合成是计算机视觉领域一个有趣而又具有挑战性的课题。生成人像视频比生成单个图像要困难得多。在本文中,我们提出了一种新的视频生成框架,用于合成任意长度的人脸视频,而不需要任何人脸样本或地标。为了克服人脸视频的合成歧义,我们提出了分而治之的策略,分别从人脸身份合成和重排两个方面解决视频人脸合成问题。为此,我们设计了一个级联网络,该网络包含三个组件,身份感知GAN (IA-GAN),人脸相干网络和插值网络。提出了一种从一组噪声中合成具有相同身份的逼真人脸的方法。人脸相干网络的目的是在保持帧间相干性的同时,对IA-GAN生成的人脸进行重新排列。为了消除相邻两帧之间的不连续,提高人脸视频的平滑度,引入了插值网络。实验结果表明,我们提出的网络能够在保持身份的前提下生成高视觉质量的人脸视频。统计数据表明,我们的方法在多个具有挑战性的数据集中优于最先进的无条件人脸视频生成模型。
{"title":"Coherence and Identity Learning for Arbitrary-length Face Video Generation","authors":"Shuquan Ye, Chu Han, Jiaying Lin, Guoqiang Han, Shengfeng He","doi":"10.1109/ICPR48806.2021.9412380","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412380","url":null,"abstract":"Face synthesis is an interesting yet challenging task in computer vision. It is even much harder to generate a portrait video than a single image. In this paper, we propose a novel video generation framework for synthesizing arbitrary-length face videos without any face exemplar or landmark. To overcome the synthesis ambiguity of face video, we propose a divide-and-conquer strategy to separately address the video face synthesis problem from two aspects, face identity synthesis and rearrangement. To this end, we design a cascaded network which contains three components, Identity-aware GAN (IA-GAN), Face Coherence Network, and Interpolation Network. IA-GAN is proposed to synthesize photorealistic faces with the same identity from a set of noises. Face Coherence Network is designed to re-arrange the faces generated by IA-GAN while keeping the inter-frame coherence. Interpolation Network is introduced to eliminate the discontinuity between two adjacent frames and improve the smoothness of the face video. Experimental results demonstrate that our proposed network is able to generate face video with high visual quality while preserving the identity. Statistics show that our method outperforms state-of-the-art unconditional face video generative models in multiple challenging datasets.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"31 1","pages":"915-922"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73736795","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Handwritten Signature and Text based User Verification using Smartwatch 使用智能手表的手写签名和基于文本的用户验证
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412048
Raghavendra Ramachandra, S. Venkatesh, K. B. Raja, C. Busch
Wrist-wearable devices such as smartwatch hardware have gained popularity as they provide quick access to various information and easy access to multiple applications. Among the numerous smartwatch applications, user verification based on the handwriting is gaining momentum by considering its reliability and user-friendliness. In this paper, we present a novel technique for user verification using a smartwatch based writing pattern or style. The proposed approach leverages accelerometer data captured from the smartwatch that is further represented using 2D Continuous Wavelet Transform (CWT) and deep features extracted using the pre-trained ResNet50. These features are classified using an ensemble of classifiers to make the final decision on user verification. Extensive experiments are carried out on a newly captured dataset using two different smartwatches with three different writing scenarios (or activities). Experimental results provide critical insights and analysis of the results in such a verification scenario.
智能手表等腕部可穿戴设备由于能够快速访问各种信息和方便地访问多个应用程序而受到欢迎。在众多的智能手表应用中,以手写为基础的用户验证考虑到其可靠性和易用性,正在成为热门。在本文中,我们提出了一种使用基于智能手表的书写模式或风格进行用户验证的新技术。所提出的方法利用从智能手表捕获的加速度计数据,使用二维连续小波变换(CWT)进一步表示,并使用预训练的ResNet50提取深度特征。使用分类器集合对这些特征进行分类,以对用户验证做出最终决定。在新捕获的数据集上进行了广泛的实验,使用两种不同的智能手表和三种不同的写作场景(或活动)。在这样的验证场景中,实验结果提供了对结果的关键见解和分析。
{"title":"Handwritten Signature and Text based User Verification using Smartwatch","authors":"Raghavendra Ramachandra, S. Venkatesh, K. B. Raja, C. Busch","doi":"10.1109/ICPR48806.2021.9412048","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412048","url":null,"abstract":"Wrist-wearable devices such as smartwatch hardware have gained popularity as they provide quick access to various information and easy access to multiple applications. Among the numerous smartwatch applications, user verification based on the handwriting is gaining momentum by considering its reliability and user-friendliness. In this paper, we present a novel technique for user verification using a smartwatch based writing pattern or style. The proposed approach leverages accelerometer data captured from the smartwatch that is further represented using 2D Continuous Wavelet Transform (CWT) and deep features extracted using the pre-trained ResNet50. These features are classified using an ensemble of classifiers to make the final decision on user verification. Extensive experiments are carried out on a newly captured dataset using two different smartwatches with three different writing scenarios (or activities). Experimental results provide critical insights and analysis of the results in such a verification scenario.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"19 1","pages":"5099-5106"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73757112","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Dimensionality Reduction for Data Visualization and Linear Classification, and the Trade-off between Robustness and Classification Accuracy 数据可视化和线性分类的降维,以及鲁棒性和分类精度之间的权衡
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412865
Martin Becker, J. Lippel, Thomas Zielke
This paper has three intertwined goals. The first is to introduce a new similarity measure for scatter plots. It uses Delaunay triangulations to compare two scatter plots regarding their relative positioning of clusters. The second is to apply this measure for the robustness assessment of a recent deep neural network (DNN) approach to dimensionality reduction (DR) for data visualization. It uses a nonlinear generalization of Fisher's linear discriminant analysis (LDA) as the encoder network of a deep autoencoder (DAE). The DAE's decoder network acts as a regularizer. The third goal is to look at different variants of the DNN: ones that promise robustness and ones that promise high classification accuracies. This is to study the trade-off between these two objectives – our results support the recent claim that robustness may be at odds with accuracy; however, results that are balanced regarding both objectives are achievable. We see a restricted Boltzmann machine (RBM) pretraining and the DAE based regularization as important building blocks for achieving balanced results. As a means of assessing the robustness of DR methods, we propose a measure that is based on our similarity measure for scatter plots. The robustness measure comes with a superimposition view of Delaunay triangulations that enables a fast comparison of results from multiple DR methods.
本文有三个相互交织的目标。首先是引入一种新的散点图相似度度量。它使用Delaunay三角测量来比较两个散点图关于它们集群的相对定位。第二个是将该度量应用于最近的深度神经网络(DNN)方法的鲁棒性评估,该方法用于数据可视化的降维(DR)。它采用Fisher线性判别分析(LDA)的非线性推广作为深度自编码器(DAE)的编码器网络。DAE的解码器网络充当正则化器。第三个目标是研究DNN的不同变体:那些承诺鲁棒性的和那些承诺高分类精度的。这是为了研究这两个目标之间的权衡——我们的结果支持最近的说法,即稳健性可能与准确性不一致;然而,平衡两个目标的结果是可以实现的。我们将受限玻尔兹曼机(RBM)预训练和基于DAE的正则化视为实现平衡结果的重要构建块。作为评估DR方法鲁棒性的一种手段,我们提出了一种基于散点图相似性度量的度量。鲁棒性测量带有Delaunay三角测量的叠加视图,可以快速比较多种DR方法的结果。
{"title":"Dimensionality Reduction for Data Visualization and Linear Classification, and the Trade-off between Robustness and Classification Accuracy","authors":"Martin Becker, J. Lippel, Thomas Zielke","doi":"10.1109/ICPR48806.2021.9412865","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412865","url":null,"abstract":"This paper has three intertwined goals. The first is to introduce a new similarity measure for scatter plots. It uses Delaunay triangulations to compare two scatter plots regarding their relative positioning of clusters. The second is to apply this measure for the robustness assessment of a recent deep neural network (DNN) approach to dimensionality reduction (DR) for data visualization. It uses a nonlinear generalization of Fisher's linear discriminant analysis (LDA) as the encoder network of a deep autoencoder (DAE). The DAE's decoder network acts as a regularizer. The third goal is to look at different variants of the DNN: ones that promise robustness and ones that promise high classification accuracies. This is to study the trade-off between these two objectives – our results support the recent claim that robustness may be at odds with accuracy; however, results that are balanced regarding both objectives are achievable. We see a restricted Boltzmann machine (RBM) pretraining and the DAE based regularization as important building blocks for achieving balanced results. As a means of assessing the robustness of DR methods, we propose a measure that is based on our similarity measure for scatter plots. The robustness measure comes with a superimposition view of Delaunay triangulations that enables a fast comparison of results from multiple DR methods.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"9 1","pages":"6478-6485"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74271567","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LiNet: A Lightweight Network for Image Super Resolution LiNet:用于图像超分辨率的轻量级网络
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412823
Armin Mehri, P. B. Ardakani, A. Sappa
This paper proposes a new lightweight network, LiNet, that enhancing technical efficiency in lightweight super resolution and operating approximately like very large and costly networks in terms of number of network parameters and operations. The proposed architecture allows the network to learn more abstract properties by avoiding low-level information via multiple links. LiNet introduces a Compact Dense Module, which contains set of inner and outer blocks, to efficiently extract meaningful information, to better leverage multi-level representations before upsampling stage, and to allow an efficient information and gradient flow within the network. Experiments on benchmark datasets show that the proposed LiNet achieves favorable performance against lightweight state-of-the-art methods.
本文提出了一种新的轻量级网络LiNet,它提高了轻量级超分辨率的技术效率,并且在网络参数数量和操作方面近似于非常大型和昂贵的网络。所提出的体系结构允许网络通过多个链接避免低级信息来学习更抽象的属性。LiNet引入了一个紧凑的密集模块,它包含一组内部和外部块,以有效地提取有意义的信息,在上采样阶段之前更好地利用多层次表示,并允许网络内有效的信息和梯度流。在基准数据集上的实验表明,所提出的LiNet相对于最先进的轻量级方法取得了良好的性能。
{"title":"LiNet: A Lightweight Network for Image Super Resolution","authors":"Armin Mehri, P. B. Ardakani, A. Sappa","doi":"10.1109/ICPR48806.2021.9412823","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412823","url":null,"abstract":"This paper proposes a new lightweight network, LiNet, that enhancing technical efficiency in lightweight super resolution and operating approximately like very large and costly networks in terms of number of network parameters and operations. The proposed architecture allows the network to learn more abstract properties by avoiding low-level information via multiple links. LiNet introduces a Compact Dense Module, which contains set of inner and outer blocks, to efficiently extract meaningful information, to better leverage multi-level representations before upsampling stage, and to allow an efficient information and gradient flow within the network. Experiments on benchmark datasets show that the proposed LiNet achieves favorable performance against lightweight state-of-the-art methods.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"39 1","pages":"7196-7202"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74320347","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Loop-closure detection by LiDAR scan re-identification 闭环检测通过激光雷达扫描重新识别
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412843
Jukka Peltomäki, Xingyang Ni, Jussi Puura, J. Kämäräinen, H. Huttunen
In this work, loop-closure detection from LiDAR scans is defined as an image re-identification problem. Reidentification is performed by computing Euclidean distances of a query scan to a gallery set of previous scans. The distances are computed in a feature embedding space where the scans are mapped by a convolutional neural network (CNN). The network is trained using the triplet loss training strategy. In our experiments we compare different backbone networks, variants of the triplet loss and generic and LiDAR specific data augmentation techniques. With a realistic indoor dataset the best architecture obtains the mean average precision (mAP) above 0.94.
在这项工作中,激光雷达扫描的闭环检测被定义为图像重新识别问题。通过计算查询扫描到先前扫描的库集的欧氏距离来执行重新识别。在特征嵌入空间中计算距离,其中扫描由卷积神经网络(CNN)映射。该网络采用三重损失训练策略进行训练。在我们的实验中,我们比较了不同的主干网、三元丢失的变体以及通用和激光雷达特定的数据增强技术。在真实的室内数据集上,最佳结构的平均精度(mAP)在0.94以上。
{"title":"Loop-closure detection by LiDAR scan re-identification","authors":"Jukka Peltomäki, Xingyang Ni, Jussi Puura, J. Kämäräinen, H. Huttunen","doi":"10.1109/ICPR48806.2021.9412843","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412843","url":null,"abstract":"In this work, loop-closure detection from LiDAR scans is defined as an image re-identification problem. Reidentification is performed by computing Euclidean distances of a query scan to a gallery set of previous scans. The distances are computed in a feature embedding space where the scans are mapped by a convolutional neural network (CNN). The network is trained using the triplet loss training strategy. In our experiments we compare different backbone networks, variants of the triplet loss and generic and LiDAR specific data augmentation techniques. With a realistic indoor dataset the best architecture obtains the mean average precision (mAP) above 0.94.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"134 1","pages":"9107-9114"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75288174","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Unified Framework for Distance-Aware Domain Adaptation 距离感知域自适应的统一框架
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412752
Fei Wang, Youdong Ding, Huan Liang, Yuzhen Gao, W. Che
Unsupervised domain adaptation has achieved significant results by leveraging knowledge from a source domain to learn a related but unlabeled target domain. Previous methods are insufficient to model domain discrepancy and class discrepancy, which may lead to misalignment and poor adaptation performance. To address this problem, in this paper, we propose a unified framework, called distance-aware domain adaptation, which is fully aware of both cross-domain distance and class-discriminative distance. In addition, second-order statistics distance and manifold alignment are also exploited to extract more information from data. In this manner, the generalization error of the target domain in classification problems can be reduced substantially. To validate the proposed method, we conducted experiments on five public datasets and an ablation study. The results demonstrate the good performance of our proposed method.
无监督域自适应通过利用源域的知识来学习相关但未标记的目标域,取得了显著的效果。以往的方法对领域差异和类差异建模不足,可能导致不匹配和自适应性能差。为了解决这一问题,本文提出了一个统一的框架,称为距离感知领域自适应,该框架充分意识到跨领域距离和类别区分距离。此外,还利用二阶统计量距离和流形对齐来从数据中提取更多的信息。这样可以大大降低分类问题中目标域的泛化误差。为了验证所提出的方法,我们在五个公共数据集和消融研究上进行了实验。结果表明,该方法具有良好的性能。
{"title":"A Unified Framework for Distance-Aware Domain Adaptation","authors":"Fei Wang, Youdong Ding, Huan Liang, Yuzhen Gao, W. Che","doi":"10.1109/ICPR48806.2021.9412752","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412752","url":null,"abstract":"Unsupervised domain adaptation has achieved significant results by leveraging knowledge from a source domain to learn a related but unlabeled target domain. Previous methods are insufficient to model domain discrepancy and class discrepancy, which may lead to misalignment and poor adaptation performance. To address this problem, in this paper, we propose a unified framework, called distance-aware domain adaptation, which is fully aware of both cross-domain distance and class-discriminative distance. In addition, second-order statistics distance and manifold alignment are also exploited to extract more information from data. In this manner, the generalization error of the target domain in classification problems can be reduced substantially. To validate the proposed method, we conducted experiments on five public datasets and an ablation study. The results demonstrate the good performance of our proposed method.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"211 1","pages":"1796-1803"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74436090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Incorporating a graph-matching algorithm into a muscle mechanics model 将图形匹配算法纳入肌肉力学模型
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9412767
Pep Santacruz, F. Serratosa
Differential models for the simulation of the muscle mechanics are based on iteratively updating a mesh grid and deducing its new state through a finite element model. Models usually assume that the mesh grid is almost regular, and this makes a degradation of the simulation accuracy in long simulation sequences, since the mesh tends to be less regular when the number of iterations increases. We present a model that has the aim of reducing this accuracy degradation. It is based on recomputing the mesh grid returned by the model in each iteration through the concept of graph matching. The new model is currently in use to analyse the dynamics of the human heart when some pressure is applied to it. The final goal of the project (which is not shown in this paper) is to deduce the optimal position and strength pressure applied to the heart that increases the chance of reviving it with the minimum tissue damage. Experimental validation shows that our model returns a higher accuracy of the muscle position through some iterations than classical differential models with an insignificant increase of runtime. Thus, it is worth recomputing the mesh grid since the simulation accuracy drastically increases at the expense of a low runtime increase.
肌肉力学模拟的微分模型是基于网格网格的迭代更新,并通过有限元模型推导其新状态。模型通常假设网格网格几乎是规则的,这使得长仿真序列的仿真精度下降,因为随着迭代次数的增加,网格往往不那么规则。我们提出了一个模型,其目的是减少这种精度下降。它是基于通过图匹配的概念,在每次迭代中重新计算模型返回的网格。这个新模型目前被用于分析施加压力时人类心脏的动态。该项目的最终目标(在本文中没有显示)是推断出施加在心脏上的最佳位置和强度压力,以最小的组织损伤增加恢复心脏的机会。实验验证表明,通过一些迭代,我们的模型返回的肌肉位置精度高于经典微分模型,而运行时间的增加并不显著。因此,重新计算网格是值得的,因为模拟精度会以较低的运行时间增加为代价大幅提高。
{"title":"Incorporating a graph-matching algorithm into a muscle mechanics model","authors":"Pep Santacruz, F. Serratosa","doi":"10.1109/ICPR48806.2021.9412767","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9412767","url":null,"abstract":"Differential models for the simulation of the muscle mechanics are based on iteratively updating a mesh grid and deducing its new state through a finite element model. Models usually assume that the mesh grid is almost regular, and this makes a degradation of the simulation accuracy in long simulation sequences, since the mesh tends to be less regular when the number of iterations increases. We present a model that has the aim of reducing this accuracy degradation. It is based on recomputing the mesh grid returned by the model in each iteration through the concept of graph matching. The new model is currently in use to analyse the dynamics of the human heart when some pressure is applied to it. The final goal of the project (which is not shown in this paper) is to deduce the optimal position and strength pressure applied to the heart that increases the chance of reviving it with the minimum tissue damage. Experimental validation shows that our model returns a higher accuracy of the muscle position through some iterations than classical differential models with an insignificant increase of runtime. Thus, it is worth recomputing the mesh grid since the simulation accuracy drastically increases at the expense of a low runtime increase.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"110 1","pages":"39-46"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74675079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CQNN: Convolutional Quadratic Neural Networks CQNN:卷积二次神经网络
Pub Date : 2021-01-10 DOI: 10.1109/ICPR48806.2021.9413207
Pranav Mantini, Shishir K. Shah
Image classification is a fundamental task in computer vision. A variety of deep learning models based on the Convolutional Neural Network (CNN) architecture have proven to be an efficient solution. Numerous improvements have been proposed over the years, where broader, deeper, and denser networks have been constructed. However, the atomic operation for these models has remained a linear unit (single neuron). In this work, we pursue an alternative dimension by hypothesizing the atomic operation to be performed by a quadratic unit. We construct convolutional layers using quadratic neurons for feature extraction and subsequently use dense layers for classification. We perform analysis to quantify the implication of replacing linear neurons with quadratic units. Results show a keen improvement in classification accuracy with quadratic neurons over linear neurons.
图像分类是计算机视觉的一项基本任务。基于卷积神经网络(CNN)架构的各种深度学习模型已被证明是一种有效的解决方案。多年来,人们提出了许多改进建议,构建了更广泛、更深、更密集的网络。然而,这些模型的原子操作仍然是线性单位(单个神经元)。在这项工作中,我们通过假设原子操作由二次单元执行来追求另一个维度。我们使用二次神经元构造卷积层进行特征提取,随后使用密集层进行分类。我们进行分析来量化用二次单元代替线性神经元的含义。结果表明,与线性神经元相比,二次神经元的分类精度有明显提高。
{"title":"CQNN: Convolutional Quadratic Neural Networks","authors":"Pranav Mantini, Shishir K. Shah","doi":"10.1109/ICPR48806.2021.9413207","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9413207","url":null,"abstract":"Image classification is a fundamental task in computer vision. A variety of deep learning models based on the Convolutional Neural Network (CNN) architecture have proven to be an efficient solution. Numerous improvements have been proposed over the years, where broader, deeper, and denser networks have been constructed. However, the atomic operation for these models has remained a linear unit (single neuron). In this work, we pursue an alternative dimension by hypothesizing the atomic operation to be performed by a quadratic unit. We construct convolutional layers using quadratic neurons for feature extraction and subsequently use dense layers for classification. We perform analysis to quantify the implication of replacing linear neurons with quadratic units. Results show a keen improvement in classification accuracy with quadratic neurons over linear neurons.","PeriodicalId":6783,"journal":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","volume":"12 1","pages":"9819-9826"},"PeriodicalIF":0.0,"publicationDate":"2021-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74707817","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
期刊
2020 25th International Conference on Pattern Recognition (ICPR)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1