首页 > 最新文献

2019 IEEE Winter Applications of Computer Vision Workshops (WACVW)最新文献

英文 中文
Sponsors and Corporate Donors 赞助商及企业捐助者
Pub Date : 2019-01-01 DOI: 10.1109/wacv.2019.00007
{"title":"Sponsors and Corporate Donors","authors":"","doi":"10.1109/wacv.2019.00007","DOIUrl":"https://doi.org/10.1109/wacv.2019.00007","url":null,"abstract":"","PeriodicalId":254512,"journal":{"name":"2019 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121222639","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MFC Datasets: Large-Scale Benchmark Datasets for Media Forensic Challenge Evaluation MFC数据集:媒体取证挑战评估的大规模基准数据集
Pub Date : 2019-01-01 DOI: 10.1109/WACVW.2019.00018
Haiying Guan, Mark Kozak, Eric Robertson, Yooyoung Lee, Amy N. Yates, Andrew Delgado, Daniel Zhou, Timothée Kheyrkhah, Jeff M. Smith, J. Fiscus
We provide a benchmark for digital Media Forensics Challenge (MFC) evaluations. Our comprehensive data comprises over 176,000 high provenance (HP) images and 11,000 HP videos; more than 100,000 manipulated images and 4,000 manipulated videos; 35 million internet images and 300,000 video clips. We have designed and generated a series of development, evaluation, and challenge datasets, and used them to assess the progress and thoroughly analyze the performance of diverse systems on a variety of media forensics tasks in the past two years. In this paper, we first introduce the objectives, challenges, and approaches to building media forensics evaluation datasets. We then discuss our approaches to forensic dataset collection, annotation, and manipulation, and present the design and infrastructure to effectively and efficiently build the evaluation datasets to support various evaluation tasks. Given a specified query, we build an infrastructure that selects the customized evaluation subsets for the targeted analysis report. Finally, we demonstrate the evaluation results in the past evaluations.
我们为数字媒体取证挑战(MFC)评估提供了一个基准。我们的综合数据包括超过176,000张高来源(HP)图像和11,000个HP视频;10万多张篡改图像和4000多段篡改视频;3500万张网络图片和30万段视频。我们设计并生成了一系列的开发、评估和挑战数据集,并使用它们来评估进展,并在过去两年中彻底分析了不同系统在各种媒体取证任务中的性能。在本文中,我们首先介绍了构建媒体取证评估数据集的目标、挑战和方法。然后,我们讨论了取证数据集收集、注释和操作的方法,并提出了有效构建评估数据集的设计和基础设施,以支持各种评估任务。给定一个指定的查询,我们构建一个基础设施,为目标分析报告选择自定义的评估子集。最后,对以往的评价结果进行了论证。
{"title":"MFC Datasets: Large-Scale Benchmark Datasets for Media Forensic Challenge Evaluation","authors":"Haiying Guan, Mark Kozak, Eric Robertson, Yooyoung Lee, Amy N. Yates, Andrew Delgado, Daniel Zhou, Timothée Kheyrkhah, Jeff M. Smith, J. Fiscus","doi":"10.1109/WACVW.2019.00018","DOIUrl":"https://doi.org/10.1109/WACVW.2019.00018","url":null,"abstract":"We provide a benchmark for digital Media Forensics Challenge (MFC) evaluations. Our comprehensive data comprises over 176,000 high provenance (HP) images and 11,000 HP videos; more than 100,000 manipulated images and 4,000 manipulated videos; 35 million internet images and 300,000 video clips. We have designed and generated a series of development, evaluation, and challenge datasets, and used them to assess the progress and thoroughly analyze the performance of diverse systems on a variety of media forensics tasks in the past two years. In this paper, we first introduce the objectives, challenges, and approaches to building media forensics evaluation datasets. We then discuss our approaches to forensic dataset collection, annotation, and manipulation, and present the design and infrastructure to effectively and efficiently build the evaluation datasets to support various evaluation tasks. Given a specified query, we build an infrastructure that selects the customized evaluation subsets for the targeted analysis report. Finally, we demonstrate the evaluation results in the past evaluations.","PeriodicalId":254512,"journal":{"name":"2019 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116876412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 128
Exploring Automatic Face Recognition on Match Performance and Gender Bias for Children 儿童自动人脸识别在匹配表现和性别偏见方面的研究
Pub Date : 2019-01-01 DOI: 10.1109/WACVW.2019.00023
Nisha Srinivas, Matthew Hivner, Kevin Gay, Harleen Atwal, Michael A. King, K. Ricanek
In this work we update the body of knowledge on the performance of child face recognition against a set of commercial-off-the-shelf (COTS) algorithms as well as a set of government sponsored algorithms. In particular, this work examines performance of multiple deep learning face recognition systems (8 distinct solutions) establishing a performance base line for a publicly available child dataset. Furthermore, we examine the phenomenon of gender bias as a function of match performance across the eight (8) systems. This work highlights the continued challenge that exists for child face recognition as a function of aging. Rank-1 accuracy ranges from 0.44 to 0.78 with an average accuracy of 0.63 on a dataset of 745 unique subjects (7,990 total images). Furthermore, when we introduce a distractor set of approximately 10; 000 child faces the rank-1 accuracy decreases across all systems on an average of 10 points. Additionally, the phenomenon of gender bias is exhibited across all systems, although the developers of the face recognition systems claim a near balance of genders was used in the development. The question of gender disparity is elusive, and although co-factors such as makeup, expression, and hair were not explicitly controlled, the dataset does not contain substantial differences across the genders. This work contributes to the body of knowledge in multiple categories, 1. child face recognition, 2. gender bias for face recognition and the notion that females as a sub-population may exhibit Lamb characteristics according to Doddington's Biometric Zoo, and 3. a dataset for child face recognition.
在这项工作中,我们针对一组商用现货(COTS)算法以及一组政府资助的算法更新了儿童面部识别性能的知识体系。特别是,这项工作检查了多个深度学习人脸识别系统(8种不同的解决方案)的性能,为公开可用的儿童数据集建立了性能基线。此外,我们研究了性别偏见现象作为跨八(8)个系统的匹配性能的函数。这项工作强调了儿童面部识别作为年龄函数存在的持续挑战。Rank-1的精度范围从0.44到0.78,在745个不同主题的数据集(总共7990张图像)上平均精度为0.63。进一步,当我们引入一个约为10的分心物集时;在所有系统中,000名儿童面临排名第一的准确率平均下降10分。此外,性别偏见现象在所有系统中都表现出来,尽管人脸识别系统的开发人员声称在开发中使用了近乎平衡的性别。性别差异的问题是难以捉摸的,尽管化妆、表情和头发等辅助因素没有得到明确控制,但数据集并没有包含性别之间的实质性差异。这项工作有助于在多个类别的知识体系,1。2.儿童面部识别;面部识别的性别偏见,以及根据Doddington's Biometric Zoo的观点,女性作为一个亚群体可能会表现出Lamb的特征。儿童人脸识别数据集。
{"title":"Exploring Automatic Face Recognition on Match Performance and Gender Bias for Children","authors":"Nisha Srinivas, Matthew Hivner, Kevin Gay, Harleen Atwal, Michael A. King, K. Ricanek","doi":"10.1109/WACVW.2019.00023","DOIUrl":"https://doi.org/10.1109/WACVW.2019.00023","url":null,"abstract":"In this work we update the body of knowledge on the performance of child face recognition against a set of commercial-off-the-shelf (COTS) algorithms as well as a set of government sponsored algorithms. In particular, this work examines performance of multiple deep learning face recognition systems (8 distinct solutions) establishing a performance base line for a publicly available child dataset. Furthermore, we examine the phenomenon of gender bias as a function of match performance across the eight (8) systems. This work highlights the continued challenge that exists for child face recognition as a function of aging. Rank-1 accuracy ranges from 0.44 to 0.78 with an average accuracy of 0.63 on a dataset of 745 unique subjects (7,990 total images). Furthermore, when we introduce a distractor set of approximately 10; 000 child faces the rank-1 accuracy decreases across all systems on an average of 10 points. Additionally, the phenomenon of gender bias is exhibited across all systems, although the developers of the face recognition systems claim a near balance of genders was used in the development. The question of gender disparity is elusive, and although co-factors such as makeup, expression, and hair were not explicitly controlled, the dataset does not contain substantial differences across the genders. This work contributes to the body of knowledge in multiple categories, 1. child face recognition, 2. gender bias for face recognition and the notion that females as a sub-population may exhibit Lamb characteristics according to Doddington's Biometric Zoo, and 3. a dataset for child face recognition.","PeriodicalId":254512,"journal":{"name":"2019 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116345366","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Exploiting Visual Artifacts to Expose Deepfakes and Face Manipulations 利用视觉伪影暴露深度伪造和面部操纵
Pub Date : 2019-01-01 DOI: 10.1109/WACVW.2019.00020
Falko Matern, C. Riess, M. Stamminger
High quality face editing in videos is a growing concern and spreads distrust in video content. However, upon closer examination, many face editing algorithms exhibit artifacts that resemble classical computer vision issues that stem from face tracking and editing. As a consequence, we wonder how difficult it is to expose artificial faces from current generators? To this end, we review current facial editing methods and several characteristic artifacts from their processin pipelines. We also show that relatively simple visual artifacts can be already quite effective in exposing such manipulations, including Deepfakes and Face2Face. Since the methods are based on visual features, they are easily explicable also to non-technical experts. The methods are easy to implement and offer capabilities for rapid adjustment to new manipulation types with little data available. Despite their simplicity, the methods are able to achieve AUC values of up to 0.866.
视频中高质量的人脸编辑越来越受到关注,并在视频内容中传播不信任。然而,经过仔细研究,许多人脸编辑算法表现出类似于源自人脸跟踪和编辑的经典计算机视觉问题的伪影。因此,我们想知道从电流发生器中暴露人造面孔有多难?为此,我们回顾了当前的面部编辑方法和其处理管道中的几个特征工件。我们还表明,相对简单的视觉伪影已经可以相当有效地暴露这种操纵,包括Deepfakes和Face2Face。由于这些方法是基于视觉特征的,因此对于非技术专家来说也很容易解释。这些方法易于实现,并提供了在可用数据很少的情况下快速调整新的操作类型的能力。尽管方法简单,但AUC值高达0.866。
{"title":"Exploiting Visual Artifacts to Expose Deepfakes and Face Manipulations","authors":"Falko Matern, C. Riess, M. Stamminger","doi":"10.1109/WACVW.2019.00020","DOIUrl":"https://doi.org/10.1109/WACVW.2019.00020","url":null,"abstract":"High quality face editing in videos is a growing concern and spreads distrust in video content. However, upon closer examination, many face editing algorithms exhibit artifacts that resemble classical computer vision issues that stem from face tracking and editing. As a consequence, we wonder how difficult it is to expose artificial faces from current generators? To this end, we review current facial editing methods and several characteristic artifacts from their processin pipelines. We also show that relatively simple visual artifacts can be already quite effective in exposing such manipulations, including Deepfakes and Face2Face. Since the methods are based on visual features, they are easily explicable also to non-technical experts. The methods are easy to implement and offer capabilities for rapid adjustment to new manipulation types with little data available. Despite their simplicity, the methods are able to achieve AUC values of up to 0.866.","PeriodicalId":254512,"journal":{"name":"2019 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127756146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 433
Can Liveness Be Automatically Detected from Latent Fingerprints? 能否从潜在指纹中自动检测出活体?
Pub Date : 2019-01-01 DOI: 10.1109/WACVW.2019.00021
Emanuela Marasco, S. Cando, Larry L Tang
Fingerprint liveness detection has been widely discussed as a solution for addressing the vulnerability of fingerprint recognition systems to presentation attacks. Multiple algorithms have been designed and implemented to operate on images acquired with commercial sensors, but such methodology is not currently available for latent prints. The possibility of wrongful conviction from fake latent evidence is reasonable, since spoof finger marks can be realistically planted at a crime scene. This paper discusses concerns pertaining to spoofing friction ridges with the purpose of leaving fake marks to contaminate the evidence associated with the investigation of a crime. There is no prior literature on liveness detection from latent prints acquired from crime scene. We illustrate the need to address such threat by experimentally evaluating the existing liveness detection approaches on latent fingerprints. This study allow us to gain a deeper understanding of the advantages and disadvantages of the existing methods, and presents a novel research direction focused on investigating the effectiveness of existing countermeasures against the danger of spoofed marks. In particular, we evaluate texture-based detectors initially developed for automatic fingerprint systems and deep convolution neural networks. The experiments are carried out on the NIST SD27 latent fingerprints database.
指纹活动性检测作为一种解决指纹识别系统易受表示攻击的方法,已被广泛讨论。已经设计和实施了多种算法来操作商业传感器获得的图像,但这种方法目前还不能用于潜在印刷品。由于伪造的潜在证据可以在犯罪现场植入伪造的手印,因此有可能被误判。本文讨论了与欺骗摩擦脊有关的问题,其目的是留下假痕迹,污染与犯罪调查有关的证据。目前尚无从犯罪现场获得的潜在指纹进行活体检测的文献。我们说明需要解决这种威胁,通过实验评估现有的活体检测方法对潜在的指纹。本研究使我们能够更深入地了解现有方法的优缺点,并提出了一个新的研究方向,即研究现有对策对假冒商标危险的有效性。特别地,我们评估了最初为自动指纹系统和深度卷积神经网络开发的基于纹理的检测器。实验在NIST SD27潜指纹数据库上进行。
{"title":"Can Liveness Be Automatically Detected from Latent Fingerprints?","authors":"Emanuela Marasco, S. Cando, Larry L Tang","doi":"10.1109/WACVW.2019.00021","DOIUrl":"https://doi.org/10.1109/WACVW.2019.00021","url":null,"abstract":"Fingerprint liveness detection has been widely discussed as a solution for addressing the vulnerability of fingerprint recognition systems to presentation attacks. Multiple algorithms have been designed and implemented to operate on images acquired with commercial sensors, but such methodology is not currently available for latent prints. The possibility of wrongful conviction from fake latent evidence is reasonable, since spoof finger marks can be realistically planted at a crime scene. This paper discusses concerns pertaining to spoofing friction ridges with the purpose of leaving fake marks to contaminate the evidence associated with the investigation of a crime. There is no prior literature on liveness detection from latent prints acquired from crime scene. We illustrate the need to address such threat by experimentally evaluating the existing liveness detection approaches on latent fingerprints. This study allow us to gain a deeper understanding of the advantages and disadvantages of the existing methods, and presents a novel research direction focused on investigating the effectiveness of existing countermeasures against the danger of spoofed marks. In particular, we evaluate texture-based detectors initially developed for automatic fingerprint systems and deep convolution neural networks. The experiments are carried out on the NIST SD27 latent fingerprints database.","PeriodicalId":254512,"journal":{"name":"2019 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"217 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114853113","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Synthesizing Attributes with Unreal Engine for Fine-grained Activity Analysis 用虚幻引擎合成属性进行细粒度活动分析
Pub Date : 2019-01-01 DOI: 10.1109/WACVW.2019.00013
Tae Soo Kim, Michael Peven, Weichao Qiu, A. Yuille, Gregory Hager
We examine the problem of activity recognition in video using simulated data for training. In contrast to the expensive task of obtaining accurate labels from real data, synthetic data creation is not only fast and scalable, but provides ground-truth labels for more than just the activities of interest, including segmentation masks, 3D object keypoints, and more. We aim to successfully transfer a model trained on synthetic data to work on video in the real world. In this work, we provide a method of transferring from synthetic to real at intermediate representations of a video. We wish to perform activity recognition from the low-dimensional latent representation of a scene as a collection of visual attributes. As the ground-truth data does not exist in the ActEV dataset for attributes of interest, specifically orientation of cars in the ground-plane with respect to the camera, we synthesize this data. We show how we can successfully transfer a car orientation classifier, and use its predictions in our defined set of visual attributes to classify actions in video.
我们使用模拟训练数据来研究视频中的活动识别问题。与从真实数据中获得准确标签的昂贵任务相比,合成数据创建不仅快速且可扩展,而且不仅为感兴趣的活动提供了真实的标签,包括分割掩码、3D对象关键点等。我们的目标是成功地将一个经过合成数据训练的模型转移到现实世界的视频上。在这项工作中,我们提供了一种在视频的中间表示中从合成到真实的转换方法。我们希望将场景的低维潜在表示作为视觉属性的集合进行活动识别。由于ActEV数据集中不存在感兴趣属性的地面真实数据,特别是汽车在地平面上相对于相机的方向,因此我们综合了这些数据。我们展示了如何成功地转移汽车方向分类器,并在我们定义的视觉属性集中使用其预测来对视频中的动作进行分类。
{"title":"Synthesizing Attributes with Unreal Engine for Fine-grained Activity Analysis","authors":"Tae Soo Kim, Michael Peven, Weichao Qiu, A. Yuille, Gregory Hager","doi":"10.1109/WACVW.2019.00013","DOIUrl":"https://doi.org/10.1109/WACVW.2019.00013","url":null,"abstract":"We examine the problem of activity recognition in video using simulated data for training. In contrast to the expensive task of obtaining accurate labels from real data, synthetic data creation is not only fast and scalable, but provides ground-truth labels for more than just the activities of interest, including segmentation masks, 3D object keypoints, and more. We aim to successfully transfer a model trained on synthetic data to work on video in the real world. In this work, we provide a method of transferring from synthetic to real at intermediate representations of a video. We wish to perform activity recognition from the low-dimensional latent representation of a scene as a collection of visual attributes. As the ground-truth data does not exist in the ActEV dataset for attributes of interest, specifically orientation of cars in the ground-plane with respect to the camera, we synthesize this data. We show how we can successfully transfer a car orientation classifier, and use its predictions in our defined set of visual attributes to classify actions in video.","PeriodicalId":254512,"journal":{"name":"2019 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"15 27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124255103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
ActEV18: Human Activity Detection Evaluation for Extended Videos ActEV18:扩展视频的人类活动检测评估
Pub Date : 2019-01-01 DOI: 10.1109/WACVW.2019.00008
Yooyoung Lee, J. Fiscus, A. Godil, David Joy, Andrew Delgado, Jim Golden
Video analytic technologies that are able to detect and classify activity are crucial for applications in many domains, such as transportation and public safety. In spite of many data collection efforts and benchmark studies in the computer vision community, there has been a lack of system development that meets practical needs for such specific domain applications. In this paper, we introduce the Activities in Extended Video (ActEV) challenge to facilitate development of video analytic technologies that can automatically detect target activities, and identify and track objects associated with each activity. To benchmark the performance of currently available algorithms, we initiated the ActEV’18 activity-level evaluation along with reference segmentation and leaderboard evaluations. In this paper, we present a summary of results and findings from these evaluations. Fifteen teams from academia and industry participated in the ActEV18 evaluations using 19 activities from the VIRAT V1 dataset.
能够检测和分类活动的视频分析技术对于许多领域的应用至关重要,例如交通和公共安全。尽管在计算机视觉领域进行了大量的数据收集工作和基准研究,但仍缺乏满足此类特定领域应用实际需求的系统开发。在本文中,我们引入了扩展视频中的活动(ActEV)挑战,以促进视频分析技术的发展,这些技术可以自动检测目标活动,并识别和跟踪与每个活动相关的对象。为了对当前可用算法的性能进行基准测试,我们启动了ActEV ' 18活动级别评估以及参考分割和排行榜评估。在本文中,我们对这些评估的结果和发现进行了总结。来自学术界和工业界的15个团队使用来自VIRAT V1数据集的19个活动参与了ActEV18评估。
{"title":"ActEV18: Human Activity Detection Evaluation for Extended Videos","authors":"Yooyoung Lee, J. Fiscus, A. Godil, David Joy, Andrew Delgado, Jim Golden","doi":"10.1109/WACVW.2019.00008","DOIUrl":"https://doi.org/10.1109/WACVW.2019.00008","url":null,"abstract":"Video analytic technologies that are able to detect and classify activity are crucial for applications in many domains, such as transportation and public safety. In spite of many data collection efforts and benchmark studies in the computer vision community, there has been a lack of system development that meets practical needs for such specific domain applications. In this paper, we introduce the Activities in Extended Video (ActEV) challenge to facilitate development of video analytic technologies that can automatically detect target activities, and identify and track objects associated with each activity. To benchmark the performance of currently available algorithms, we initiated the ActEV’18 activity-level evaluation along with reference segmentation and leaderboard evaluations. In this paper, we present a summary of results and findings from these evaluations. Fifteen teams from academia and industry participated in the ActEV18 evaluations using 19 activities from the VIRAT V1 dataset.","PeriodicalId":254512,"journal":{"name":"2019 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"123 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121348905","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Novel Activities Detection Algorithm in Extended Videos 扩展视频中新的活动检测算法
Pub Date : 2019-01-01 DOI: 10.1109/WACVW.2019.00009
L. Yao, Ying Qian
Due to participation in TRECVID ActEV[1] competition, we conduct research on temporal activity recognition. In this paper, we propose a system for activity detection and localize detected activities temporally in extended videos. Our system firstly detects objects in video frames. Secondly, we use position information of detected object, as input to the object tracking model, which can obtain motion information of multiple objects in consecutive frames. Lastly, we input consecutive video frames containing only detected objects into 3D Convolutional Neural Network to achieve features, and 3D CNN is followed by a recurrent neural network for accurately localizing the detected activity.
由于参加了TRECVID ActEV[1]竞赛,我们对时间活动识别进行了研究。在本文中,我们提出了一个活动检测系统,并在扩展视频中暂时定位检测到的活动。我们的系统首先检测视频帧中的物体。其次,将检测到的目标位置信息作为目标跟踪模型的输入,获得连续帧内多个目标的运动信息;最后,我们将只包含检测到的目标的连续视频帧输入到3D卷积神经网络中进行特征提取,然后在3D CNN之后进行循环神经网络精确定位检测到的活动。
{"title":"Novel Activities Detection Algorithm in Extended Videos","authors":"L. Yao, Ying Qian","doi":"10.1109/WACVW.2019.00009","DOIUrl":"https://doi.org/10.1109/WACVW.2019.00009","url":null,"abstract":"Due to participation in TRECVID ActEV[1] competition, we conduct research on temporal activity recognition. In this paper, we propose a system for activity detection and localize detected activities temporally in extended videos. Our system firstly detects objects in video frames. Secondly, we use position information of detected object, as input to the object tracking model, which can obtain motion information of multiple objects in consecutive frames. Lastly, we input consecutive video frames containing only detected objects into 3D Convolutional Neural Network to achieve features, and 3D CNN is followed by a recurrent neural network for accurately localizing the detected activity.","PeriodicalId":254512,"journal":{"name":"2019 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"111 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114861208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Predicting Soft Biometric Attributes from 30 Pixels: A Case Study in NIR Ocular Images 从30像素预测软生物特征属性:近红外眼部图像的案例研究
Pub Date : 2019-01-01 DOI: 10.1109/WACVW.2019.00024
Denton Bobeldyk, A. Ross
In this work, we investigate the possibility of extracting soft biometric attributes, viz., gender, race and eye color, from down-sampled near-infrared ocular images. In particular, we evaluate the possibility of deducing gender, race and eye color from ocular images as small as 56 pixels. Our preliminary analysis yields the surprising result that gender, race and eye color cues are still available in such low-resolution near-infrared images. This research bolsters the previously made assertion in the literature that certain soft biometric attributes can be deduced from poor quality biometric data.
在这项工作中,我们研究了从下采样的近红外眼部图像中提取软生物特征属性(即性别、种族和眼睛颜色)的可能性。特别是,我们评估了从小到56像素的眼部图像推断性别、种族和眼睛颜色的可能性。我们的初步分析得出了一个令人惊讶的结果:在这种低分辨率的近红外图像中,仍然可以找到性别、种族和眼睛颜色的线索。这项研究支持了先前在文献中提出的断言,即某些软生物特征属性可以从质量较差的生物特征数据中推断出来。
{"title":"Predicting Soft Biometric Attributes from 30 Pixels: A Case Study in NIR Ocular Images","authors":"Denton Bobeldyk, A. Ross","doi":"10.1109/WACVW.2019.00024","DOIUrl":"https://doi.org/10.1109/WACVW.2019.00024","url":null,"abstract":"In this work, we investigate the possibility of extracting soft biometric attributes, viz., gender, race and eye color, from down-sampled near-infrared ocular images. In particular, we evaluate the possibility of deducing gender, race and eye color from ocular images as small as 56 pixels. Our preliminary analysis yields the surprising result that gender, race and eye color cues are still available in such low-resolution near-infrared images. This research bolsters the previously made assertion in the literature that certain soft biometric attributes can be deduced from poor quality biometric data.","PeriodicalId":254512,"journal":{"name":"2019 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"102 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134081592","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Considering Race a Problem of Transfer Learning 考虑种族问题的迁移学习
Pub Date : 2018-12-12 DOI: 10.1109/WACVW.2019.00022
Akbir Khan, M. Mahmoud
As biometric applications are fielded to serve large population groups, issues of performance differences between individual sub-groups are becoming increasingly important. In this paper we examine cases where we believe race is one such factor. We look in particular at two forms of problem; facial classification and image synthesis. We take the novel approach of considering race as a boundary for transfer learning in both the task (facial classification) and the domain (synthesis over distinct datasets). We demonstrate a series of techniques to improve transfer learning of facial classification; outperforming similar models trained in the target's own domain. We conduct a study to evaluate the performance drop of Generative Adversarial Networks trained to conduct image synthesis, in this process, we produce a new annotation for the Celeb-A dataset by race. These networks are trained solely on one race and tested on another - demonstrating the subsets of the CelebA to be distinct domains for this task.
随着生物识别应用的应用范围越来越广,各个子群体之间的性能差异问题变得越来越重要。在本文中,我们研究了一些我们认为种族是其中一个因素的案例。我们特别关注两种形式的问题;人脸分类与图像合成。我们采用了一种新颖的方法,将种族作为任务(面部分类)和领域(不同数据集的综合)迁移学习的边界。我们展示了一系列改进面部分类迁移学习的技术;优于在目标领域训练的类似模型。我们进行了一项研究,以评估生成对抗网络训练进行图像合成的性能下降,在这个过程中,我们根据种族为名人- a数据集生成了一个新的注释。这些网络仅在一个种族上进行训练,并在另一个种族上进行测试-证明CelebA的子集是该任务的不同域。
{"title":"Considering Race a Problem of Transfer Learning","authors":"Akbir Khan, M. Mahmoud","doi":"10.1109/WACVW.2019.00022","DOIUrl":"https://doi.org/10.1109/WACVW.2019.00022","url":null,"abstract":"As biometric applications are fielded to serve large population groups, issues of performance differences between individual sub-groups are becoming increasingly important. In this paper we examine cases where we believe race is one such factor. We look in particular at two forms of problem; facial classification and image synthesis. We take the novel approach of considering race as a boundary for transfer learning in both the task (facial classification) and the domain (synthesis over distinct datasets). We demonstrate a series of techniques to improve transfer learning of facial classification; outperforming similar models trained in the target's own domain. We conduct a study to evaluate the performance drop of Generative Adversarial Networks trained to conduct image synthesis, in this process, we produce a new annotation for the Celeb-A dataset by race. These networks are trained solely on one race and tested on another - demonstrating the subsets of the CelebA to be distinct domains for this task.","PeriodicalId":254512,"journal":{"name":"2019 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"247 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115187462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
期刊
2019 IEEE Winter Applications of Computer Vision Workshops (WACVW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1