首页 > 最新文献

2018 IEEE Winter Applications of Computer Vision Workshops (WACVW)最新文献

英文 中文
Calorific Expenditure Estimation Using Deep Convolutional Network Features 基于深度卷积网络特征的热量消耗估算
Pub Date : 2018-04-26 DOI: 10.1109/WACVW.2018.00014
Baodong Wang, L. Tao, T. Burghardt, M. Mirmehdi
Accurately estimating a person's energy expenditure is an important tool in tracking physical activity levels for healthcare and sports monitoring tasks, amongst other applications. In this paper, we propose a method for deriving calorific expenditure based on deep convolutional neural network features (within a healthcare scenario). Our evaluation shows that the proposed approach gives high accuracy in activity recognition (82.3%) and low normalised root mean square error in calorific expenditure prediction (0.41). It is compared against the current state-ofthe-art calorific expenditure estimation method, based on a classical approach, and exhibits an improvement of 7.8% in the calorific expenditure prediction task. The proposed method is suitable for home monitoring in a controlled environment.
准确估计一个人的能量消耗是跟踪医疗保健和运动监测任务的身体活动水平的重要工具,以及其他应用。在本文中,我们提出了一种基于深度卷积神经网络特征的热量消耗方法(在医疗保健场景中)。我们的评估表明,所提出的方法在活动识别方面具有较高的准确性(82.3%),在热量消耗预测方面具有较低的归一化均方根误差(0.41)。与目前基于经典方法的最先进的热量消耗估算方法相比,该方法在热量消耗预测任务中提高了7.8%。该方法适用于受控环境下的家庭监控。
{"title":"Calorific Expenditure Estimation Using Deep Convolutional Network Features","authors":"Baodong Wang, L. Tao, T. Burghardt, M. Mirmehdi","doi":"10.1109/WACVW.2018.00014","DOIUrl":"https://doi.org/10.1109/WACVW.2018.00014","url":null,"abstract":"Accurately estimating a person's energy expenditure is an important tool in tracking physical activity levels for healthcare and sports monitoring tasks, amongst other applications. In this paper, we propose a method for deriving calorific expenditure based on deep convolutional neural network features (within a healthcare scenario). Our evaluation shows that the proposed approach gives high accuracy in activity recognition (82.3%) and low normalised root mean square error in calorific expenditure prediction (0.41). It is compared against the current state-ofthe-art calorific expenditure estimation method, based on a classical approach, and exhibits an improvement of 7.8% in the calorific expenditure prediction task. The proposed method is suitable for home monitoring in a controlled environment.","PeriodicalId":301220,"journal":{"name":"2018 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130732237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Learning Visual Engagement for Trauma Recovery 学习视觉参与创伤恢复
Pub Date : 2018-03-15 DOI: 10.1109/WACVW.2018.00016
Svati Dhamija, T. Boult
Applications ranging from human emotion understanding to e-health are exploring methods to effectively understand user behavior from self-reported questionnaires. However, little is understood about non-invasive techniques that involve face-based deep-learning models to predict engagement. Current research in visual engagement poses two key questions: 1) how much time do we need to analyze facial behavior for accurate engagement prediction? and 2) which deep learning approach provides the most accurate predictions? In this paper we compare RNN, GRU and LSTM using different length segments of AUs. Our experiments show no significant difference in prediction accuracy when using anywhere between 15 and 90 seconds of data. Moreover, the results reveal that simpler models of recurrent networks are statistically significantly better suited for capturing engagement from AUs.
从人类情感理解到电子健康的应用都在探索从自我报告的问卷中有效理解用户行为的方法。然而,对于涉及基于面部的深度学习模型来预测参与的非侵入性技术,人们知之甚少。目前关于视觉参与的研究提出了两个关键问题:1)我们需要多少时间来分析面部行为以准确预测参与?2)哪种深度学习方法能提供最准确的预测?在本文中,我们比较了RNN, GRU和LSTM使用不同长度的au段。我们的实验表明,当使用15到90秒的数据时,预测精度没有显着差异。此外,结果表明,简单的循环网络模型在统计上明显更适合于从au获取参与度。
{"title":"Learning Visual Engagement for Trauma Recovery","authors":"Svati Dhamija, T. Boult","doi":"10.1109/WACVW.2018.00016","DOIUrl":"https://doi.org/10.1109/WACVW.2018.00016","url":null,"abstract":"Applications ranging from human emotion understanding to e-health are exploring methods to effectively understand user behavior from self-reported questionnaires. However, little is understood about non-invasive techniques that involve face-based deep-learning models to predict engagement. Current research in visual engagement poses two key questions: 1) how much time do we need to analyze facial behavior for accurate engagement prediction? and 2) which deep learning approach provides the most accurate predictions? In this paper we compare RNN, GRU and LSTM using different length segments of AUs. Our experiments show no significant difference in prediction accuracy when using anywhere between 15 and 90 seconds of data. Moreover, the results reveal that simpler models of recurrent networks are statistically significantly better suited for capturing engagement from AUs.","PeriodicalId":301220,"journal":{"name":"2018 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125188750","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Facial Expression Recognition Using a Large Out-of-Context Dataset 基于大型非上下文数据集的面部表情识别
Pub Date : 2018-03-15 DOI: 10.1109/WACVW.2018.00012
Elizabeth Tran, Michael B. Mayhew, Hyojin Kim, P. Karande, A. Kaplan
We develop a method for emotion recognition from facial imagery. This problem is challenging in part because of the subjectivity of ground truth labels and in part because of the relatively small size of existing labeled datasets. We use the FER+ dataset [8], a dataset with multiple emotion labels per image, in order to build an emotion recognition model that encompasses a full range of emotions. Since the amount of data in the FER+ dataset is limited, we explore the use of a much larger face dataset, MS-Celeb-1M [41], in conjunction with the FER+ dataset. Specific layers within an Inception-ResNet-v1 [13, 38] model trained for facial recognition are used for the emotion recognition problem. Thus, we leverage the MS-Celeb-1M dataset in addition to the FER+ dataset and experiment with different architectures to assess the overall performance of neural networks to recognize emotion using facial imagery.
我们开发了一种基于面部图像的情感识别方法。这个问题是具有挑战性的,部分原因是地面真值标签的主观性,部分原因是现有标记数据集的规模相对较小。我们使用FER+数据集[8],每个图像具有多个情绪标签的数据集,以构建包含全范围情绪的情绪识别模型。由于FER+数据集中的数据量有限,我们探索了一个更大的人脸数据集MS-Celeb-1M[41]与FER+数据集结合使用。在Inception-ResNet-v1[13,38]模型中训练用于面部识别的特定层用于情感识别问题。因此,我们利用MS-Celeb-1M数据集和FER+数据集,并使用不同的架构进行实验,以评估神经网络使用面部图像识别情绪的整体性能。
{"title":"Facial Expression Recognition Using a Large Out-of-Context Dataset","authors":"Elizabeth Tran, Michael B. Mayhew, Hyojin Kim, P. Karande, A. Kaplan","doi":"10.1109/WACVW.2018.00012","DOIUrl":"https://doi.org/10.1109/WACVW.2018.00012","url":null,"abstract":"We develop a method for emotion recognition from facial imagery. This problem is challenging in part because of the subjectivity of ground truth labels and in part because of the relatively small size of existing labeled datasets. We use the FER+ dataset [8], a dataset with multiple emotion labels per image, in order to build an emotion recognition model that encompasses a full range of emotions. Since the amount of data in the FER+ dataset is limited, we explore the use of a much larger face dataset, MS-Celeb-1M [41], in conjunction with the FER+ dataset. Specific layers within an Inception-ResNet-v1 [13, 38] model trained for facial recognition are used for the emotion recognition problem. Thus, we leverage the MS-Celeb-1M dataset in addition to the FER+ dataset and experiment with different architectures to assess the overall performance of neural networks to recognize emotion using facial imagery.","PeriodicalId":301220,"journal":{"name":"2018 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"468 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132969943","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Evaluating a Convolutional Neural Network on Short-Wave Infra-Red Images 短波红外图像的卷积神经网络评价
Pub Date : 2018-03-15 DOI: 10.1109/WACVW.2018.00008
M. Bihn, Manuel Günther, Daniel Lemmond, T. Boult
Machine learning algorithms, both traditional and neuralnetwork-based, have been tested against RGB facial images for years, but these algorithms are prone to fail when illumination conditions are insufficient, for example, at night or when images are taken from long distances. Short-Wave Infra-Red (SWIR) illumination provides a much higher intensity and a much more ambient structure than visible light, which makes it better suited for face recognition in different conditions. However, current neural networks require lots of training data, which is not available in the SWIR domain. In this paper, we examine the ability of a convolutional neural network, specifically, the VGG Face network, which was trained on visible spectrum images, to work on SWIR images. Utilizing a dataset containing both RGB and SWIR images, we hypothesize that the VGG Face network will perform well both on facial images taken in RGB and SWIR wavelengths. We expect that the features extracted with VGG Face are independent of the actual wavelengths that the images were taken with. Thus, face recognition with VGG Face is possible between the RGB and SWIR domains. We find that VGG Face performs reasonable on some of the SWIR wavelengths. We can almost reach the same recognition performance when using composite images built from three SWIR wavelengths probing on RGB.
机器学习算法,无论是传统的还是基于神经网络的,已经针对RGB面部图像进行了多年的测试,但是这些算法在照明条件不足的情况下很容易失败,例如,在夜间或从远距离拍摄图像。短波红外线(SWIR)照明提供比可见光更高的强度和更多的环境结构,这使得它更适合在不同条件下进行面部识别。然而,目前的神经网络需要大量的训练数据,这在SWIR领域是不可用的。在本文中,我们研究了卷积神经网络,特别是在可见光谱图像上训练的VGG人脸网络,在SWIR图像上工作的能力。利用包含RGB和SWIR图像的数据集,我们假设VGG Face网络在RGB和SWIR波长的面部图像上都表现良好。我们期望用VGG Face提取的特征与实际拍摄图像的波长无关。因此,在RGB和SWIR域之间使用VGG face进行人脸识别是可能的。我们发现VGG Face在一些SWIR波长上表现合理。在RGB上使用三个SWIR波长探测构建的合成图像几乎可以达到相同的识别性能。
{"title":"Evaluating a Convolutional Neural Network on Short-Wave Infra-Red Images","authors":"M. Bihn, Manuel Günther, Daniel Lemmond, T. Boult","doi":"10.1109/WACVW.2018.00008","DOIUrl":"https://doi.org/10.1109/WACVW.2018.00008","url":null,"abstract":"Machine learning algorithms, both traditional and neuralnetwork-based, have been tested against RGB facial images for years, but these algorithms are prone to fail when illumination conditions are insufficient, for example, at night or when images are taken from long distances. Short-Wave Infra-Red (SWIR) illumination provides a much higher intensity and a much more ambient structure than visible light, which makes it better suited for face recognition in different conditions. However, current neural networks require lots of training data, which is not available in the SWIR domain. In this paper, we examine the ability of a convolutional neural network, specifically, the VGG Face network, which was trained on visible spectrum images, to work on SWIR images. Utilizing a dataset containing both RGB and SWIR images, we hypothesize that the VGG Face network will perform well both on facial images taken in RGB and SWIR wavelengths. We expect that the features extracted with VGG Face are independent of the actual wavelengths that the images were taken with. Thus, face recognition with VGG Face is possible between the RGB and SWIR domains. We find that VGG Face performs reasonable on some of the SWIR wavelengths. We can almost reach the same recognition performance when using composite images built from three SWIR wavelengths probing on RGB.","PeriodicalId":301220,"journal":{"name":"2018 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114803902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Facial Attributes Guided Deep Sketch-to-Photo Synthesis 面部属性引导深度素描到照片合成
Pub Date : 2018-03-15 DOI: 10.1109/WACVW.2018.00006
Hadi Kazemi, S. M. Iranmanesh, Ali Dabouei, Sobhan Soleymani, N. Nasrabadi
Face sketch-photo synthesis is a critical application in law enforcement and digital entertainment industry. Despite the significant improvements in sketch-to-photo synthesis techniques, existing methods have still serious limitations in practice, such as the need for paired data in the training phase or having no control on enforcing facial attributes over the synthesized image. In this work, we present a new framework, which is a conditional version of Cycle-GAN, conditioned on facial attributes. The proposed network forces facial attributes, such as skin and hair color, on the synthesized photo and does not need a set of aligned face-sketch pairs during its training. We evaluate the proposed network by training on two real and synthetic sketch datasets. The hand-sketch images of the FERET dataset and the color face images from the WVU Multi-modal dataset are used as an unpaired input to the proposed conditional CycleGAN with the skin color as the controlled face attribute. For more attribute guided evaluation, a synthetic sketch dataset is created from the CelebA dataset and used to evaluate the performance of the network by forcing several desired facial attributes on the synthesized faces.
人脸素描-照片合成是执法和数字娱乐行业的重要应用。尽管素描到照片合成技术有了很大的进步,但现有的方法在实践中仍然存在严重的局限性,例如在训练阶段需要配对数据,或者无法控制在合成图像上强制执行面部属性。在这项工作中,我们提出了一个新的框架,这是一个循环gan的条件版本,以面部属性为条件。所提出的网络将面部属性(如皮肤和头发颜色)强加到合成照片上,并且在训练过程中不需要一组对齐的面部草图对。我们通过在两个真实和合成的草图数据集上训练来评估所提出的网络。将FERET数据集的手绘图像和WVU多模态数据集的彩色人脸图像作为非配对输入,以肤色作为受控人脸属性,输入到所提出的条件CycleGAN中。对于更多的属性导向评估,从CelebA数据集创建了一个合成草图数据集,并通过在合成的面部上强制几个所需的面部属性来评估网络的性能。
{"title":"Facial Attributes Guided Deep Sketch-to-Photo Synthesis","authors":"Hadi Kazemi, S. M. Iranmanesh, Ali Dabouei, Sobhan Soleymani, N. Nasrabadi","doi":"10.1109/WACVW.2018.00006","DOIUrl":"https://doi.org/10.1109/WACVW.2018.00006","url":null,"abstract":"Face sketch-photo synthesis is a critical application in law enforcement and digital entertainment industry. Despite the significant improvements in sketch-to-photo synthesis techniques, existing methods have still serious limitations in practice, such as the need for paired data in the training phase or having no control on enforcing facial attributes over the synthesized image. In this work, we present a new framework, which is a conditional version of Cycle-GAN, conditioned on facial attributes. The proposed network forces facial attributes, such as skin and hair color, on the synthesized photo and does not need a set of aligned face-sketch pairs during its training. We evaluate the proposed network by training on two real and synthetic sketch datasets. The hand-sketch images of the FERET dataset and the color face images from the WVU Multi-modal dataset are used as an unpaired input to the proposed conditional CycleGAN with the skin color as the controlled face attribute. For more attribute guided evaluation, a synthetic sketch dataset is created from the CelebA dataset and used to evaluate the performance of the network by forcing several desired facial attributes on the synthesized faces.","PeriodicalId":301220,"journal":{"name":"2018 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122887802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
A Multi-task Convolutional Neural Network for Joint Iris Detection and Presentation Attack Detection 一种多任务卷积神经网络联合虹膜检测与表示攻击检测
Pub Date : 2018-03-15 DOI: 10.1109/WACVW.2018.00011
Cunjian Chen, A. Ross
In this work, we propose a multi-task convolutional neural network learning approach that can simultaneously perform iris localization and presentation attack detection (PAD). The proposed multi-task PAD (MT-PAD) is inspired by an object detection method which directly regresses the parameters of the iris bounding box and computes the probability of presentation attack from the input ocular image. Experiments involving both intra-sensor and cross-sensor scenarios suggest that the proposed method can achieve state-of-the-art results on publicly available datasets. To the best of our knowledge, this is the first work that performs iris detection and iris presentation attack detection simultaneously.
在这项工作中,我们提出了一种多任务卷积神经网络学习方法,可以同时执行虹膜定位和呈现攻击检测(PAD)。本文提出的多任务PAD (MT-PAD)的灵感来自于一种目标检测方法,该方法直接回归虹膜边界框的参数,并从输入的眼图像中计算呈现攻击的概率。涉及传感器内和跨传感器场景的实验表明,所提出的方法可以在公开可用的数据集上获得最先进的结果。据我们所知,这是第一个同时进行虹膜检测和虹膜表示攻击检测的工作。
{"title":"A Multi-task Convolutional Neural Network for Joint Iris Detection and Presentation Attack Detection","authors":"Cunjian Chen, A. Ross","doi":"10.1109/WACVW.2018.00011","DOIUrl":"https://doi.org/10.1109/WACVW.2018.00011","url":null,"abstract":"In this work, we propose a multi-task convolutional neural network learning approach that can simultaneously perform iris localization and presentation attack detection (PAD). The proposed multi-task PAD (MT-PAD) is inspired by an object detection method which directly regresses the parameters of the iris bounding box and computes the probability of presentation attack from the input ocular image. Experiments involving both intra-sensor and cross-sensor scenarios suggest that the proposed method can achieve state-of-the-art results on publicly available datasets. To the best of our knowledge, this is the first work that performs iris detection and iris presentation attack detection simultaneously.","PeriodicalId":301220,"journal":{"name":"2018 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"03 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131049945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 64
Generic Object Discrimination for Mobile Assistive Robots Using Projective Light Diffusion 基于投影光扩散的移动辅助机器人一般目标识别
Pub Date : 2018-03-15 DOI: 10.1109/WACVW.2018.00013
P. Papadakis, David Filliat
A number of assistive robot services depend on the classification of objects while dealing with an increased volume of sensory data, scene variability and limited computational resources. We propose using more concise representations via a seamless combination of photometric and geometric features fused by exploiting local photometric/geometric correlation and employing domain transform filtering in order to recover scene structure. This is obtained through a projective light diffusion imaging process (PLDI) which allows capturing surface orientation, image edges and global depth gradients into a single image. Object candidates are finally encoded into a discriminative, wavelet-based descriptor allowing very fast object queries. Experiments with an indoor robot demonstrate improved classification performance compared to alternative methods and an overall superior discriminative power compared to state-of-the-art unsupervised descriptors within ModelNet10 benchmark.
许多辅助机器人服务依赖于物体分类,同时处理越来越多的感官数据、场景变化和有限的计算资源。我们建议使用更简洁的表示,通过利用局部光度/几何相关性融合的光度和几何特征的无缝组合,并采用域变换滤波来恢复场景结构。这是通过投射光扩散成像过程(PLDI)获得的,该过程允许将表面方向,图像边缘和全局深度梯度捕获到单个图像中。候选对象最终被编码成一个判别的、基于小波的描述符,允许非常快速的对象查询。与其他方法相比,室内机器人的实验表明,与ModelNet10基准中最先进的无监督描述符相比,其分类性能有所提高,总体判别能力也有所提高。
{"title":"Generic Object Discrimination for Mobile Assistive Robots Using Projective Light Diffusion","authors":"P. Papadakis, David Filliat","doi":"10.1109/WACVW.2018.00013","DOIUrl":"https://doi.org/10.1109/WACVW.2018.00013","url":null,"abstract":"A number of assistive robot services depend on the classification of objects while dealing with an increased volume of sensory data, scene variability and limited computational resources. We propose using more concise representations via a seamless combination of photometric and geometric features fused by exploiting local photometric/geometric correlation and employing domain transform filtering in order to recover scene structure. This is obtained through a projective light diffusion imaging process (PLDI) which allows capturing surface orientation, image edges and global depth gradients into a single image. Object candidates are finally encoded into a discriminative, wavelet-based descriptor allowing very fast object queries. Experiments with an indoor robot demonstrate improved classification performance compared to alternative methods and an overall superior discriminative power compared to state-of-the-art unsupervised descriptors within ModelNet10 benchmark.","PeriodicalId":301220,"journal":{"name":"2018 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"415 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132337607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
ivisX: An Integrated Video Investigation Suite for Forensic Applications ivisX:集成视频调查套件的法医应用
Pub Date : 2018-03-01 DOI: 10.1109/WACVW.2018.00007
Chengchao Qu, J. Metzler, Eduardo Monari
Video data from surveillance cameras are nowadays an important instrument for investigating crimes and identifying the identity of an offender. The analysis of the mass data acquired from numerous cameras poses enormous challenges to police investigation authorities. Supporting softwares and video management tools currently on the market focus either on elaborate visualization and editing of video data, specific image processing or video content analysis tasks. As a result, such a scattered system landscape further exacerbates the complexity and difficulty of a timely analysis of the available data. This work presents our unified framework ivisX, which is an integrated suite to simplify the entire workflow of video data investigation. The algorithmic backbone of ivisX is built upon an effective content-based search algorithm using region covariance for low-resolution (LR) data and a novel 3D face super-resolution (FSR) approach, which can generate high-resolution (HR) 3D face models to render high-quality facial composites with a single blurred and pixelated face image of the LR domain. Moreover, ivisX has a modular design, which allows for flexible incorporation of various extensions ranging from processing and display of video data from multiple cameras to analysis and documentation of the results into a powerful integrated toolkit to assist forensic investigation.
如今,监控摄像机的视频数据是调查犯罪和识别罪犯身份的重要工具。对从众多摄像机中获取的大量数据进行分析,给警方调查当局带来了巨大的挑战。目前市场上的支持软件和视频管理工具要么专注于视频数据的精细可视化和编辑,要么专注于特定的图像处理或视频内容分析任务。因此,这种分散的系统景观进一步加剧了及时分析可用数据的复杂性和难度。这项工作提出了我们的统一框架ivisX,这是一个集成套件,简化了视频数据调查的整个工作流程。ivisX的算法主干是基于一种有效的基于内容的搜索算法,该算法使用低分辨率(LR)数据的区域协方差和一种新颖的3D人脸超分辨率(FSR)方法,该方法可以生成高分辨率(HR) 3D人脸模型,从而用LR域的单个模糊和像素化人脸图像呈现高质量的面部复合材料。此外,ivisX具有模块化设计,允许灵活地结合各种扩展,从处理和显示来自多个摄像机的视频数据到分析和记录结果,成为一个强大的集成工具包,以协助法医调查。
{"title":"ivisX: An Integrated Video Investigation Suite for Forensic Applications","authors":"Chengchao Qu, J. Metzler, Eduardo Monari","doi":"10.1109/WACVW.2018.00007","DOIUrl":"https://doi.org/10.1109/WACVW.2018.00007","url":null,"abstract":"Video data from surveillance cameras are nowadays an important instrument for investigating crimes and identifying the identity of an offender. The analysis of the mass data acquired from numerous cameras poses enormous challenges to police investigation authorities. Supporting softwares and video management tools currently on the market focus either on elaborate visualization and editing of video data, specific image processing or video content analysis tasks. As a result, such a scattered system landscape further exacerbates the complexity and difficulty of a timely analysis of the available data. This work presents our unified framework ivisX, which is an integrated suite to simplify the entire workflow of video data investigation. The algorithmic backbone of ivisX is built upon an effective content-based search algorithm using region covariance for low-resolution (LR) data and a novel 3D face super-resolution (FSR) approach, which can generate high-resolution (HR) 3D face models to render high-quality facial composites with a single blurred and pixelated face image of the LR domain. Moreover, ivisX has a modular design, which allows for flexible incorporation of various extensions ranging from processing and display of video data from multiple cameras to analysis and documentation of the results into a powerful integrated toolkit to assist forensic investigation.","PeriodicalId":301220,"journal":{"name":"2018 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125853802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Enhancing Optical Cross-Sensor Fingerprint Matching Using Local Textural Features 利用局部纹理特征增强光学指纹匹配
Pub Date : 2018-03-01 DOI: 10.1109/WACVW.2018.00010
Emanuela Marasco, Alex Feldman, Keleigh Rachel Romine
Fingerprint systems have been designed to typically operate on images acquired using the same sensor. Existing fingerprint systems are not able to accurately compare images collected using different sensors. In this paper, we propose a learning-based scheme for enhancing interoperability between optical fingerprint sensors by compensating the output of a traditional commercial matcher. Specifically, cross-sensor differences are captured by incorporating Local Binary Patterns (LBP) and Local Phase Quantization (LPQ), while dimensionality reduction is performed by using Reconstruction Independent Component Analysis (RICA). The evaluation is carried out on rolled fingerprints pertaining to 494 users collected atWest Virginia University and acquired using multiple optical sensors and Ten Print cards. In cross-sensor at False Acceptance Rate of 0.01%, the proposed approach achieves a False Rejection Rate of 4.12%.
指纹系统通常被设计为对使用相同传感器获取的图像进行操作。现有的指纹系统不能准确地比较使用不同传感器收集的图像。在本文中,我们提出了一种基于学习的方案,通过补偿传统商用匹配器的输出来增强光学指纹传感器之间的互操作性。具体来说,通过结合局部二值模式(LBP)和局部相位量化(LPQ)来捕获传感器间的差异,而通过重建独立分量分析(RICA)来进行降维。这项评估是对西弗吉尼亚大学收集的494名用户的卷指纹进行的,这些指纹是通过多个光学传感器和Ten Print卡获得的。在误接受率为0.01%的交叉传感器情况下,该方法的误拒率为4.12%。
{"title":"Enhancing Optical Cross-Sensor Fingerprint Matching Using Local Textural Features","authors":"Emanuela Marasco, Alex Feldman, Keleigh Rachel Romine","doi":"10.1109/WACVW.2018.00010","DOIUrl":"https://doi.org/10.1109/WACVW.2018.00010","url":null,"abstract":"Fingerprint systems have been designed to typically operate on images acquired using the same sensor. Existing fingerprint systems are not able to accurately compare images collected using different sensors. In this paper, we propose a learning-based scheme for enhancing interoperability between optical fingerprint sensors by compensating the output of a traditional commercial matcher. Specifically, cross-sensor differences are captured by incorporating Local Binary Patterns (LBP) and Local Phase Quantization (LPQ), while dimensionality reduction is performed by using Reconstruction Independent Component Analysis (RICA). The evaluation is carried out on rolled fingerprints pertaining to 494 users collected atWest Virginia University and acquired using multiple optical sensors and Ten Print cards. In cross-sensor at False Acceptance Rate of 0.01%, the proposed approach achieves a False Rejection Rate of 4.12%.","PeriodicalId":301220,"journal":{"name":"2018 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116863791","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Automatic Access Control Based on Face and Hand Biometrics in a Non-cooperative Context 非合作环境下基于面部和手部生物特征的自动访问控制
Pub Date : 2018-03-01 DOI: 10.1109/WACVW.2018.00009
M. N. Jahromi, Morten Bojesen Bonderup, Maryam Asadi-Aghbolaghi, Egils Avots, Kamal Nasrollahi, Sergio Escalera, S. Kasaei, T. Moeslund, G. Anbarjafari
Automatic access control systems (ACS) based on the human biometrics or physical tokens are widely employed in public and private areas. Yet these systems, in their conventional forms, are restricted to active interaction from the users. In scenarios where users are not cooperating with the system, these systems are challenged. Failure in cooperation with the biometric systems might be intentional or because the users are incapable of handling the interaction procedure with the biometric system or simply forget to cooperate with it, due to for example, illness like dementia. This work introduces a challenging bimodal database, including face and hand information of the users when they approach a door to open it by its handle in a noncooperative context. We have defined two (an easy and a challenging) protocols on how to use the database. We have reported results on many baseline methods, including deep learning techniques as well as conventional methods on the database. The obtained results show the merit of the proposed database and the challenging nature of access control with non-cooperative users.
基于人体生物特征或物理令牌的自动门禁系统(ACS)广泛应用于公共和私人领域。然而,这些系统,在它们的传统形式中,被限制为来自用户的主动交互。在用户不与系统合作的场景中,这些系统将面临挑战。与生物识别系统的合作失败可能是故意的,也可能是因为用户无法处理与生物识别系统的交互程序,或者只是忘记了与它合作,例如由于痴呆症等疾病。这项工作引入了一个具有挑战性的双峰数据库,包括用户在非合作环境中接近门并通过门把手打开门时的面部和手部信息。我们已经定义了两个关于如何使用数据库的协议(一个简单,一个具有挑战性)。我们已经报告了许多基线方法的结果,包括深度学习技术以及数据库上的传统方法。得到的结果表明了所提出的数据库的优点,以及非合作用户访问控制的挑战性。
{"title":"Automatic Access Control Based on Face and Hand Biometrics in a Non-cooperative Context","authors":"M. N. Jahromi, Morten Bojesen Bonderup, Maryam Asadi-Aghbolaghi, Egils Avots, Kamal Nasrollahi, Sergio Escalera, S. Kasaei, T. Moeslund, G. Anbarjafari","doi":"10.1109/WACVW.2018.00009","DOIUrl":"https://doi.org/10.1109/WACVW.2018.00009","url":null,"abstract":"Automatic access control systems (ACS) based on the human biometrics or physical tokens are widely employed in public and private areas. Yet these systems, in their conventional forms, are restricted to active interaction from the users. In scenarios where users are not cooperating with the system, these systems are challenged. Failure in cooperation with the biometric systems might be intentional or because the users are incapable of handling the interaction procedure with the biometric system or simply forget to cooperate with it, due to for example, illness like dementia. This work introduces a challenging bimodal database, including face and hand information of the users when they approach a door to open it by its handle in a noncooperative context. We have defined two (an easy and a challenging) protocols on how to use the database. We have reported results on many baseline methods, including deep learning techniques as well as conventional methods on the database. The obtained results show the merit of the proposed database and the challenging nature of access control with non-cooperative users.","PeriodicalId":301220,"journal":{"name":"2018 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"376 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115912669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2018 IEEE Winter Applications of Computer Vision Workshops (WACVW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1