首页 > 最新文献

2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)最新文献

英文 中文
Robust Iris Segmentation Based on Fully Convolutional Networks and Generative Adversarial Networks 基于全卷积网络和生成对抗网络的鲁棒虹膜分割
Pub Date : 2018-09-04 DOI: 10.1109/SIBGRAPI.2018.00043
Cides S. Bezerra, Rayson Laroca, D. Lucio, E. Severo, L. F. Oliveira, A. Britto, D. Menotti
The iris can be considered as one of the most important biometric traits due to its high degree of uniqueness. Iris-based biometrics applications depend mainly on the iris segmentation whose suitability is not robust for different environments such as near-infrared (NIR) and visible (VIS) ones. In this paper, two approaches for robust iris segmentation based on Fully Convolutional Networks (FCNs) and Generative Adversarial Networks (GANs) are described. Similar to a common convolutional network, but without the fully connected layers (i.e., the classification layers), an FCN employs at its end combination of pooling layers from different convolutional layers. Based on the game theory, a GAN is designed as two networks competing with each other to generate the best segmentation. The proposed segmentation networks achieved promising results in all evaluated datasets (i.e., BioSec, CasiaI3, CasiaT4, IITD-1) of NIR images and (NICE.I, CrEye-Iris and MICHE-I) of VIS images in both non-cooperative and cooperative domains, outperforming the baselines techniques which are the best ones found so far in the literature, i.e., a new state of the art for these datasets. Furthermore, we manually labeled 2,431 images from CasiaT4, CrEye-Iris and MICHE-I datasets, making the masks available for research purposes.
虹膜因其高度的独特性而被认为是最重要的生物特征之一。基于虹膜的生物识别应用主要依赖于虹膜分割,其对近红外(NIR)和可见光(VIS)等不同环境的适用性不强。本文介绍了两种基于全卷积网络(fcv)和生成对抗网络(GANs)的稳健虹膜分割方法。与普通的卷积网络类似,但没有完全连接的层(即分类层),FCN在其最终使用来自不同卷积层的池化层的组合。基于博弈论,GAN被设计为两个相互竞争的网络,以产生最佳分割。所提出的分割网络在近红外图像和NICE的所有评估数据集(即BioSec, CasiaI3, CasiaT4, IITD-1)中都取得了令人满意的结果。I, CrEye-Iris和MICHE-I)在非合作和合作领域的VIS图像,优于迄今为止在文献中发现的最好的基线技术,即这些数据集的新技术。此外,我们手动标记了来自CasiaT4, CrEye-Iris和MICHE-I数据集的2,431张图像,使掩码可用于研究目的。
{"title":"Robust Iris Segmentation Based on Fully Convolutional Networks and Generative Adversarial Networks","authors":"Cides S. Bezerra, Rayson Laroca, D. Lucio, E. Severo, L. F. Oliveira, A. Britto, D. Menotti","doi":"10.1109/SIBGRAPI.2018.00043","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00043","url":null,"abstract":"The iris can be considered as one of the most important biometric traits due to its high degree of uniqueness. Iris-based biometrics applications depend mainly on the iris segmentation whose suitability is not robust for different environments such as near-infrared (NIR) and visible (VIS) ones. In this paper, two approaches for robust iris segmentation based on Fully Convolutional Networks (FCNs) and Generative Adversarial Networks (GANs) are described. Similar to a common convolutional network, but without the fully connected layers (i.e., the classification layers), an FCN employs at its end combination of pooling layers from different convolutional layers. Based on the game theory, a GAN is designed as two networks competing with each other to generate the best segmentation. The proposed segmentation networks achieved promising results in all evaluated datasets (i.e., BioSec, CasiaI3, CasiaT4, IITD-1) of NIR images and (NICE.I, CrEye-Iris and MICHE-I) of VIS images in both non-cooperative and cooperative domains, outperforming the baselines techniques which are the best ones found so far in the literature, i.e., a new state of the art for these datasets. Furthermore, we manually labeled 2,431 images from CasiaT4, CrEye-Iris and MICHE-I datasets, making the masks available for research purposes.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128603009","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
The Impact of Preprocessing on Deep Representations for Iris Recognition on Unconstrained Environments 无约束环境下预处理对虹膜识别深度表征的影响
Pub Date : 2018-08-29 DOI: 10.1109/SIBGRAPI.2018.00044
L. A. Zanlorensi, Eduardo José da S. Luz, Rayson Laroca, A. Britto, Luiz Oliveira, D. Menotti
The use of iris as a biometric trait is widely used because of its high level of distinction and uniqueness. Nowadays, one of the major research challenges relies on the recognition of iris images obtained in visible spectrum under unconstrained environments. In this scenario, the acquired iris are affected by capture distance, rotation, blur, motion blur, low contrast and specular reflection, creating noises that disturb the iris recognition systems. Besides delineating the iris region, usually preprocessing techniques such as normalization and segmentation of noisy iris images are employed to minimize these problems. But these techniques inevitably run into some errors. In this context, we propose the use of deep representations, more specifically, architectures based on VGG and ResNet-50 networks, for dealing with the images using (and not) iris segmentation and normalization. We use transfer learning from the face domain and also propose a specific data augmentation technique for iris images. Our results show that the approach using non-normalized and only circle-delimited iris images reaches a new state of the art in the official protocol of the NICE. II competition, a subset of the UBIRIS database, one of the most challenging databases on unconstrained environments, reporting an average Equal Error Rate (EER) of 13.98% which represents an absolute reduction of about 5%.
虹膜作为一种生物特征,因其高度的差异性和唯一性而得到广泛应用。在无约束环境下对可见光谱虹膜图像进行识别是目前虹膜识别研究面临的主要挑战之一。在这种情况下,获得的虹膜受到捕获距离、旋转、模糊、运动模糊、低对比度和镜面反射的影响,产生干扰虹膜识别系统的噪声。除了描绘虹膜区域外,通常采用预处理技术,如对有噪声的虹膜图像进行归一化和分割,以最大限度地减少这些问题。但这些技术不可避免地会遇到一些错误。在这种情况下,我们建议使用深度表示,更具体地说,基于VGG和ResNet-50网络的架构,使用(而不是)虹膜分割和归一化来处理图像。我们使用了人脸域的迁移学习,并提出了一种针对虹膜图像的特定数据增强技术。我们的研究结果表明,在NICE的官方协议中,使用非归一化且仅圆划分的虹膜图像的方法达到了新的水平。II竞争是UBIRIS数据库的一个子集,UBIRIS数据库是无约束环境下最具挑战性的数据库之一,报告平均相等错误率(EER)为13.98%,绝对减少了约5%。
{"title":"The Impact of Preprocessing on Deep Representations for Iris Recognition on Unconstrained Environments","authors":"L. A. Zanlorensi, Eduardo José da S. Luz, Rayson Laroca, A. Britto, Luiz Oliveira, D. Menotti","doi":"10.1109/SIBGRAPI.2018.00044","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00044","url":null,"abstract":"The use of iris as a biometric trait is widely used because of its high level of distinction and uniqueness. Nowadays, one of the major research challenges relies on the recognition of iris images obtained in visible spectrum under unconstrained environments. In this scenario, the acquired iris are affected by capture distance, rotation, blur, motion blur, low contrast and specular reflection, creating noises that disturb the iris recognition systems. Besides delineating the iris region, usually preprocessing techniques such as normalization and segmentation of noisy iris images are employed to minimize these problems. But these techniques inevitably run into some errors. In this context, we propose the use of deep representations, more specifically, architectures based on VGG and ResNet-50 networks, for dealing with the images using (and not) iris segmentation and normalization. We use transfer learning from the face domain and also propose a specific data augmentation technique for iris images. Our results show that the approach using non-normalized and only circle-delimited iris images reaches a new state of the art in the official protocol of the NICE. II competition, a subset of the UBIRIS database, one of the most challenging databases on unconstrained environments, reporting an average Equal Error Rate (EER) of 13.98% which represents an absolute reduction of about 5%.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"495 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133368542","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
On the Learning of Deep Local Features for Robust Face Spoofing Detection 鲁棒人脸欺骗检测中深度局部特征的学习
Pub Date : 2018-06-19 DOI: 10.1109/SIBGRAPI.2018.00040
G. Souza, J. Papa, A. Marana
Biometrics emerged as a robust solution for security systems. However, given the dissemination of biometric applications, criminals are developing techniques to circumvent them by simulating physical or behavioral traits of legal users (spoofing attacks). Despite face being a promising characteristic due to its universality, acceptability and presence of cameras almost everywhere, face recognition systems are extremely vulnerable to such frauds since they can be easily fooled with common printed facial photographs. State-of-the-art approaches, based on Convolutional Neural Networks (CNNs), present good results in face spoofing detection. However, these methods do not consider the importance of learning deep local features from each facial region, even though it is known from face recognition that each facial region presents different visual aspects, which can also be exploited for face spoofing detection. In this work we propose a novel CNN architecture trained in two steps for such task. Initially, each part of the neural network learns features from a given facial region. Afterwards, the whole model is fine-tuned on the whole facial images. Results show that such pre-training step allows the CNN to learn different local spoofing cues, improving the performance and the convergence speed of the final model, outperforming the state-of-the-art approaches.
生物识别技术成为安全系统的一个强大解决方案。然而,鉴于生物识别应用程序的传播,犯罪分子正在开发技术,通过模拟合法用户的身体或行为特征(欺骗攻击)来绕过它们。尽管人脸是一个很有前途的特征,因为它的普遍性、可接受性和几乎无处不在的摄像头,但人脸识别系统极易受到此类欺诈的影响,因为它们很容易被普通的打印面部照片所欺骗。基于卷积神经网络(cnn)的最新方法在人脸欺骗检测中表现出良好的效果。然而,这些方法没有考虑到从每个面部区域学习深度局部特征的重要性,尽管从人脸识别中我们知道每个面部区域呈现不同的视觉方面,这也可以用于人脸欺骗检测。在这项工作中,我们提出了一种新的CNN架构,分为两步训练。最初,神经网络的每个部分从给定的面部区域学习特征。然后,整个模型在整个面部图像上进行微调。结果表明,这种预训练步骤使CNN能够学习不同的局部欺骗线索,提高了最终模型的性能和收敛速度,优于目前最先进的方法。
{"title":"On the Learning of Deep Local Features for Robust Face Spoofing Detection","authors":"G. Souza, J. Papa, A. Marana","doi":"10.1109/SIBGRAPI.2018.00040","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00040","url":null,"abstract":"Biometrics emerged as a robust solution for security systems. However, given the dissemination of biometric applications, criminals are developing techniques to circumvent them by simulating physical or behavioral traits of legal users (spoofing attacks). Despite face being a promising characteristic due to its universality, acceptability and presence of cameras almost everywhere, face recognition systems are extremely vulnerable to such frauds since they can be easily fooled with common printed facial photographs. State-of-the-art approaches, based on Convolutional Neural Networks (CNNs), present good results in face spoofing detection. However, these methods do not consider the importance of learning deep local features from each facial region, even though it is known from face recognition that each facial region presents different visual aspects, which can also be exploited for face spoofing detection. In this work we propose a novel CNN architecture trained in two steps for such task. Initially, each part of the neural network learns features from a given facial region. Afterwards, the whole model is fine-tuned on the whole facial images. Results show that such pre-training step allows the CNN to learn different local spoofing cues, improving the performance and the convergence speed of the final model, outperforming the state-of-the-art approaches.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114856291","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
Bag of Attributes for Video Event Retrieval 视频事件检索的属性包
Pub Date : 2016-07-18 DOI: 10.1109/SIBGRAPI.2018.00064
Leonardo A. Duarte, O. A. B. Penatti, J. Almeida
In this paper, we present the Bag-of-Attributes (BoA) model for video representation aiming at video event retrieval. The BoA model is based on a semantic feature space for representing videos, resulting in high-level video feature vectors. For creating a semantic space, i.e., the attribute space, we can train a classifier using a labeled image dataset, obtaining a classification model that can be understood as a high-level codebook. This model is used to map low-level frame vectors into high-level vectors (e.g., classifier probability scores). Then, we apply pooling operations to the frame vectors to create the final bag of attributes for the video. In the BoA representation, each dimension corresponds to one category (or attribute) of the semantic space. Other interesting properties are: compactness, flexibility regarding the classifier, and ability to encode multiple semantic concepts in a single video representation. Our experiments considered the semantic space created by state-of-the-art convolutional neural networks pre-trained on 1000 object categories of ImageNet. Such deep neural networks were used to classify each video frame and then different coding strategies were used to encode the probability distribution from the softmax layer into a frame vector. Next, different pooling strategies were used to combine frame vectors in the BoA representation for a video. Results using BoA were comparable or superior to the baselines in the task of video event retrieval using the EVVE dataset, with the advantage of providing a much more compact representation.
针对视频事件检索问题,提出了一种基于属性袋(BoA)的视频表示模型。BoA模型基于表示视频的语义特征空间,从而产生高级视频特征向量。为了创建语义空间,即属性空间,我们可以使用标记的图像数据集训练分类器,获得可以理解为高级码本的分类模型。该模型用于将低级帧向量映射到高级向量(例如,分类器概率得分)。然后,我们对帧矢量应用池操作,为视频创建最终的属性包。在BoA表示中,每个维度对应于语义空间的一个类别(或属性)。其他有趣的特性是:紧凑性、关于分类器的灵活性,以及在单个视频表示中编码多个语义概念的能力。我们的实验考虑了由最先进的卷积神经网络在ImageNet的1000个对象类别上预训练产生的语义空间。利用深度神经网络对视频帧进行分类,然后采用不同的编码策略将softmax层的概率分布编码为帧向量。接下来,使用不同的池化策略来组合视频BoA表示中的帧向量。在使用EVVE数据集的视频事件检索任务中,使用BoA的结果与基线相当或优于基线,其优势是提供了更紧凑的表示。
{"title":"Bag of Attributes for Video Event Retrieval","authors":"Leonardo A. Duarte, O. A. B. Penatti, J. Almeida","doi":"10.1109/SIBGRAPI.2018.00064","DOIUrl":"https://doi.org/10.1109/SIBGRAPI.2018.00064","url":null,"abstract":"In this paper, we present the Bag-of-Attributes (BoA) model for video representation aiming at video event retrieval. The BoA model is based on a semantic feature space for representing videos, resulting in high-level video feature vectors. For creating a semantic space, i.e., the attribute space, we can train a classifier using a labeled image dataset, obtaining a classification model that can be understood as a high-level codebook. This model is used to map low-level frame vectors into high-level vectors (e.g., classifier probability scores). Then, we apply pooling operations to the frame vectors to create the final bag of attributes for the video. In the BoA representation, each dimension corresponds to one category (or attribute) of the semantic space. Other interesting properties are: compactness, flexibility regarding the classifier, and ability to encode multiple semantic concepts in a single video representation. Our experiments considered the semantic space created by state-of-the-art convolutional neural networks pre-trained on 1000 object categories of ImageNet. Such deep neural networks were used to classify each video frame and then different coding strategies were used to encode the probability distribution from the softmax layer into a frame vector. Next, different pooling strategies were used to combine frame vectors in the BoA representation for a video. Results using BoA were comparable or superior to the baselines in the task of video event retrieval using the EVVE dataset, with the advantage of providing a much more compact representation.","PeriodicalId":208985,"journal":{"name":"2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131294126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1