首页 > 最新文献

2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)最新文献

英文 中文
PaletteNet: Image Recolorization with Given Color Palette 调色板:图像重新着色与给定的调色板
Junho Cho, Sangdoo Yun, Kyoung-Ok Lee, J. Choi
Image recolorization enhances the visual perception of an image for design and artistic purposes. In this work, we present a deep neural network, referred to as PaletteNet, which recolors an image according to a given target color palette that is useful to express the color concept of an image. PaletteNet takes two inputs: a source image to be recolored and a target palette. PaletteNet is then designed to change the color concept of a source image so that the palette of the output image is close to the target palette. To train PaletteNet, the proposed multi-task loss is composed of Euclidean loss and adversarial loss. The experimental results show that the proposed method outperforms the existing recolorization methods. Human experts with a commercial software take on average 18 minutes to recolor an image, while PaletteNet automatically recolors plausible results in less than a second.
为了设计和艺术目的,图像再着色增强了图像的视觉感知。在这项工作中,我们提出了一个深度神经网络,称为PaletteNet,它根据给定的目标调色板重新为图像上色,这有助于表达图像的颜色概念。PaletteNet接受两个输入:要重新着色的源图像和目标调色板。然后,PaletteNet被设计为更改源图像的颜色概念,以便输出图像的调色板接近目标调色板。为了训练PaletteNet,提出的多任务损失由欧几里得损失和对抗损失组成。实验结果表明,该方法优于现有的再着色方法。使用商业软件的人类专家平均需要18分钟来重新为图像上色,而PaletteNet在不到一秒钟的时间内自动重新为可信的结果上色。
{"title":"PaletteNet: Image Recolorization with Given Color Palette","authors":"Junho Cho, Sangdoo Yun, Kyoung-Ok Lee, J. Choi","doi":"10.1109/CVPRW.2017.143","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.143","url":null,"abstract":"Image recolorization enhances the visual perception of an image for design and artistic purposes. In this work, we present a deep neural network, referred to as PaletteNet, which recolors an image according to a given target color palette that is useful to express the color concept of an image. PaletteNet takes two inputs: a source image to be recolored and a target palette. PaletteNet is then designed to change the color concept of a source image so that the palette of the output image is close to the target palette. To train PaletteNet, the proposed multi-task loss is composed of Euclidean loss and adversarial loss. The experimental results show that the proposed method outperforms the existing recolorization methods. Human experts with a commercial software take on average 18 minutes to recolor an image, while PaletteNet automatically recolors plausible results in less than a second.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"23 1","pages":"1058-1066"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78541979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
NTIRE 2017 Challenge on Single Image Super-Resolution: Methods and Results 2017年全图像超分辨率挑战:方法和结果
R. Timofte, E. Agustsson, L. Gool, Ming-Hsuan Yang, Lei Zhang, Bee Lim, Sanghyun Son, Heewon Kim, Seungjun Nah, Kyoung Mu Lee, Xintao Wang, Yapeng Tian, K. Yu, Yulun Zhang, Shixiang Wu, Chao Dong, Liang Lin, Y. Qiao, Chen Change Loy, Woong Bae, J. Yoo, Yoseob Han, J. C. Ye, Jae-Seok Choi, Munchurl Kim, Yuchen Fan, Jiahui Yu, Wei Han, Ding Liu, Haichao Yu, Zhangyang Wang, Humphrey Shi, Xinchao Wang, Thomas S. Huang, Yunjin Chen, K. Zhang, W. Zuo, Zhimin Tang, Linkai Luo, Shaohui Li, Min Fu, Lei Cao, Wen Heng, Giang Bui, Truc Le, Y. Duan, D. Tao, Ruxin Wang, Xu Lin, Jianxin Pang, Jinchang Xu, Yu Zhao, Xiangyu Xu, Jin-shan Pan, Deqing Sun, Yujin Zhang, Xibin Song, Yuchao Dai, Xueying Qin, X. Huynh, Tiantong Guo, Hojjat Seyed Mousavi, T. Vu, V. Monga, Cristóvão Cruz, K. Egiazarian, V. Katkovnik, Rakesh Mehta, A. Jain, Abhinav Agarwalla, C. Praveen, Ruofan Zhou, Hongdiao Wen, Chen Zhu, Zhiqiang Xia, Zhengtao Wang, Qi Guo
This paper reviews the first challenge on single image super-resolution (restoration of rich details in an low resolution image) with focus on proposed solutions and results. A new DIVerse 2K resolution image dataset (DIV2K) was employed. The challenge had 6 competitions divided into 2 tracks with 3 magnification factors each. Track 1 employed the standard bicubic downscaling setup, while Track 2 had unknown downscaling operators (blur kernel and decimation) but learnable through low and high res train images. Each competition had ∽100 registered participants and 20 teams competed in the final testing phase. They gauge the state-of-the-art in single image super-resolution.
本文综述了单幅图像超分辨率(在低分辨率图像中恢复丰富的细节)面临的第一个挑战,重点介绍了解决方案和结果。采用了一种新的多元2K分辨率图像数据集(DIV2K)。这项挑战有6个比赛,分为2个赛道,每个赛道有3个放大系数。轨道1采用标准的双三次降尺度设置,而轨道2有未知的降尺度操作符(模糊内核和抽取),但可以通过低分辨率和高分辨率火车图像学习。∽每项比赛共有100名注册参与者和20支队伍参加最后的测试阶段。他们用最先进的单图像超分辨率来衡量。
{"title":"NTIRE 2017 Challenge on Single Image Super-Resolution: Methods and Results","authors":"R. Timofte, E. Agustsson, L. Gool, Ming-Hsuan Yang, Lei Zhang, Bee Lim, Sanghyun Son, Heewon Kim, Seungjun Nah, Kyoung Mu Lee, Xintao Wang, Yapeng Tian, K. Yu, Yulun Zhang, Shixiang Wu, Chao Dong, Liang Lin, Y. Qiao, Chen Change Loy, Woong Bae, J. Yoo, Yoseob Han, J. C. Ye, Jae-Seok Choi, Munchurl Kim, Yuchen Fan, Jiahui Yu, Wei Han, Ding Liu, Haichao Yu, Zhangyang Wang, Humphrey Shi, Xinchao Wang, Thomas S. Huang, Yunjin Chen, K. Zhang, W. Zuo, Zhimin Tang, Linkai Luo, Shaohui Li, Min Fu, Lei Cao, Wen Heng, Giang Bui, Truc Le, Y. Duan, D. Tao, Ruxin Wang, Xu Lin, Jianxin Pang, Jinchang Xu, Yu Zhao, Xiangyu Xu, Jin-shan Pan, Deqing Sun, Yujin Zhang, Xibin Song, Yuchao Dai, Xueying Qin, X. Huynh, Tiantong Guo, Hojjat Seyed Mousavi, T. Vu, V. Monga, Cristóvão Cruz, K. Egiazarian, V. Katkovnik, Rakesh Mehta, A. Jain, Abhinav Agarwalla, C. Praveen, Ruofan Zhou, Hongdiao Wen, Chen Zhu, Zhiqiang Xia, Zhengtao Wang, Qi Guo","doi":"10.1109/CVPRW.2017.149","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.149","url":null,"abstract":"This paper reviews the first challenge on single image super-resolution (restoration of rich details in an low resolution image) with focus on proposed solutions and results. A new DIVerse 2K resolution image dataset (DIV2K) was employed. The challenge had 6 competitions divided into 2 tracks with 3 magnification factors each. Track 1 employed the standard bicubic downscaling setup, while Track 2 had unknown downscaling operators (blur kernel and decimation) but learnable through low and high res train images. Each competition had ∽100 registered participants and 20 teams competed in the final testing phase. They gauge the state-of-the-art in single image super-resolution.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"4 1","pages":"1110-1121"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87866449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1199
Slot Cars: 3D Modelling for Improved Visual Traffic Analytics 槽车:改进视觉交通分析的3D建模
Eduardo R. Corral-Soto, J. Elder
A major challenge in visual highway traffic analytics is to disaggregate individual vehicles from clusters formed in dense traffic conditions. Here we introduce a data driven 3D generative reasoning method to tackle this segmentation problem. The method is comprised of offline (learning) and online (inference) stages. In the offline stage, we fit a mixture model for the prior distribution of vehicle dimensions to labelled data. Given camera intrinsic parameters and height, we use a parallelism method to estimate highway lane structure and camera tilt to project 3D models to the image. In the online stage, foreground vehicle cluster segments are extracted using motion and background subtraction. For each segment, we use a data-driven MCMC method to estimate the vehicles configuration and dimensions that provide the most likely account of the observed foreground pixels. We evaluate the method on two highway datasets and demonstrate a substantial improvement on the state of the art.
视觉公路交通分析的一个主要挑战是如何从密集交通条件下形成的集群中分解出单个车辆。在这里,我们引入了一种数据驱动的三维生成推理方法来解决这个分割问题。该方法由离线(学习)和在线(推理)两个阶段组成。在离线阶段,我们拟合了车辆尺寸先验分布的混合模型。给定相机的固有参数和高度,我们使用平行度方法来估计高速公路车道结构和相机倾斜,以将三维模型投影到图像中。在在线阶段,使用运动和背景减法提取前景车辆簇段。对于每个片段,我们使用数据驱动的MCMC方法来估计车辆的配置和尺寸,这些配置和尺寸提供了最可能的观测前景像素。我们在两个高速公路数据集上评估了该方法,并展示了对最新技术的实质性改进。
{"title":"Slot Cars: 3D Modelling for Improved Visual Traffic Analytics","authors":"Eduardo R. Corral-Soto, J. Elder","doi":"10.1109/CVPRW.2017.123","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.123","url":null,"abstract":"A major challenge in visual highway traffic analytics is to disaggregate individual vehicles from clusters formed in dense traffic conditions. Here we introduce a data driven 3D generative reasoning method to tackle this segmentation problem. The method is comprised of offline (learning) and online (inference) stages. In the offline stage, we fit a mixture model for the prior distribution of vehicle dimensions to labelled data. Given camera intrinsic parameters and height, we use a parallelism method to estimate highway lane structure and camera tilt to project 3D models to the image. In the online stage, foreground vehicle cluster segments are extracted using motion and background subtraction. For each segment, we use a data-driven MCMC method to estimate the vehicles configuration and dimensions that provide the most likely account of the observed foreground pixels. We evaluate the method on two highway datasets and demonstrate a substantial improvement on the state of the art.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"1 1","pages":"889-897"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87037028","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
DeepXScope: Segmenting Microscopy Images with a Deep Neural Network DeepXScope:用深度神经网络分割显微镜图像
Philip Saponaro, Wayne Treible, Abhishek Kolagunda, Timothy Chaya, J. Caplan, C. Kambhamettu, R. Wisser
High-speed confocal microscopy has shown great promise to yield insights into plant-fungal interactions by allowing for large volumes of leaf tissue to be imaged at high magnification. Currently, segmentation is performed either manually, which is infeasible for large amounts of data, or by developing separate algorithms to extract individual features within the image data. In this work, we propose the use of a single deep convolutional neural network architecture dubbed DeepXScope for automatically segmenting hyphal networks of the fungal pathogen and cell boundaries and stomata of the host plant. DeepXScope is trained on manually annotated images created for each of these structures. We describe experiments that show each individual structure can be accurately extracted automatically using DeepXScope. We anticipate that plant scientists will be able to use this network to automatically extract multiple structures of interest, and we plan to release our tool to the community1.
高速共聚焦显微镜已经显示出巨大的希望,通过允许在高倍率下对大量的叶片组织进行成像,可以深入了解植物与真菌的相互作用。目前,分割要么是手动执行的,这对于大量数据是不可行的,要么是通过开发单独的算法来提取图像数据中的单个特征。在这项工作中,我们提出使用一个称为DeepXScope的单一深度卷积神经网络架构来自动分割真菌病原体的菌丝网络和寄主植物的细胞边界和气孔。DeepXScope是在为这些结构创建的手动注释图像上进行训练的。我们描述的实验表明,使用DeepXScope可以准确地自动提取每个单独的结构。我们预计植物科学家将能够使用该网络自动提取多种感兴趣的结构,并且我们计划向社区发布我们的工具1。
{"title":"DeepXScope: Segmenting Microscopy Images with a Deep Neural Network","authors":"Philip Saponaro, Wayne Treible, Abhishek Kolagunda, Timothy Chaya, J. Caplan, C. Kambhamettu, R. Wisser","doi":"10.1109/CVPRW.2017.117","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.117","url":null,"abstract":"High-speed confocal microscopy has shown great promise to yield insights into plant-fungal interactions by allowing for large volumes of leaf tissue to be imaged at high magnification. Currently, segmentation is performed either manually, which is infeasible for large amounts of data, or by developing separate algorithms to extract individual features within the image data. In this work, we propose the use of a single deep convolutional neural network architecture dubbed DeepXScope for automatically segmenting hyphal networks of the fungal pathogen and cell boundaries and stomata of the host plant. DeepXScope is trained on manually annotated images created for each of these structures. We describe experiments that show each individual structure can be accurately extracted automatically using DeepXScope. We anticipate that plant scientists will be able to use this network to automatically extract multiple structures of interest, and we plan to release our tool to the community1.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"2014 1","pages":"843-850"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88101538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
A Logarithmic X-Ray Imaging Model for Baggage Inspection: Simulation and Object Detection 行李检查的对数x射线成像模型:模拟和目标检测
D. Mery, A. Katsaggelos
In the last years, many computer vision algorithms have been developed for X-ray testing tasks. Some of them deal with baggage inspection, in which the aim is to detect automatically target objects. The progress in automated baggage inspection, however, is modest and very limited compared to what is needed because X-ray screening systems are still being manipulated by human inspectors. In this work, we present an X-ray imaging model that can separate foreground from background in baggage screening. The model can be used in two main tasks: i) Simulation of new X-ray images, where simulated images can be used in training programs for human inspectors, or can be used to enhance datasets for computer vision algorithms. ii) Detection of (threat) objects, where new algorithms can be employed to perform automated baggage inspection or to aid an user in the inspection task showing potential threats. In our model, rather than a multiplication of foreground and background, that is typically used in X-ray imaging, we propose the addition of logarithmic images. This allows the use of linear strategies to superimpose images of threat objects onto X-ray images and the use of sparse representations in order to segment target objects. In our experiments, we simulate new X-ray images of handguns, shuriken and razor blades, in which it is impossible to distinguish simulated and real X-ray images. In addition, we show in our experiments the effective detection of shuriken, razor blades and handguns using the proposed algorithm outperforming some alternative state-of- the-art techniques.
在过去的几年里,许多计算机视觉算法已经开发出来用于x射线测试任务。其中一些用于行李检查,其目的是自动检测目标物体。然而,与需要的相比,自动行李检查方面的进展并不大,而且非常有限,因为x射线检查系统仍由人工检查人员操纵。在这项工作中,我们提出了一种x射线成像模型,可以在行李筛查中分离前景和背景。该模型可用于两个主要任务:i)模拟新的x射线图像,其中模拟图像可用于人类检查员的培训计划,或可用于增强计算机视觉算法的数据集。ii)(威胁)物体的检测,新算法可用于执行自动行李检查或在显示潜在威胁的检查任务中帮助用户。在我们的模型中,我们建议添加对数图像,而不是在x射线成像中通常使用的前景和背景的乘法。这允许使用线性策略将威胁对象的图像叠加到x射线图像上,并使用稀疏表示来分割目标对象。在我们的实验中,我们模拟了手枪、飞刀和剃须刀片的新x射线图像,其中无法区分模拟和真实的x射线图像。此外,我们在实验中表明,使用所提出的算法对飞刀、剃须刀片和手枪的有效检测优于其他一些最先进的技术。
{"title":"A Logarithmic X-Ray Imaging Model for Baggage Inspection: Simulation and Object Detection","authors":"D. Mery, A. Katsaggelos","doi":"10.1109/CVPRW.2017.37","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.37","url":null,"abstract":"In the last years, many computer vision algorithms have been developed for X-ray testing tasks. Some of them deal with baggage inspection, in which the aim is to detect automatically target objects. The progress in automated baggage inspection, however, is modest and very limited compared to what is needed because X-ray screening systems are still being manipulated by human inspectors. In this work, we present an X-ray imaging model that can separate foreground from background in baggage screening. The model can be used in two main tasks: i) Simulation of new X-ray images, where simulated images can be used in training programs for human inspectors, or can be used to enhance datasets for computer vision algorithms. ii) Detection of (threat) objects, where new algorithms can be employed to perform automated baggage inspection or to aid an user in the inspection task showing potential threats. In our model, rather than a multiplication of foreground and background, that is typically used in X-ray imaging, we propose the addition of logarithmic images. This allows the use of linear strategies to superimpose images of threat objects onto X-ray images and the use of sparse representations in order to segment target objects. In our experiments, we simulate new X-ray images of handguns, shuriken and razor blades, in which it is impossible to distinguish simulated and real X-ray images. In addition, we show in our experiments the effective detection of shuriken, razor blades and handguns using the proposed algorithm outperforming some alternative state-of- the-art techniques.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"73 1","pages":"251-259"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83978837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Explaining Distributed Neural Activations via Unsupervised Learning 通过无监督学习解释分布式神经激活
Soheil Kolouri, Charles E. Martin, Heiko Hoffmann
Recent work has demonstrated the emergence of semantic object-part detectors in activation patterns of convolutional neural networks (CNNs), but did not account for the distributed multi-layer neural activations in such networks. In this work, we propose a novel method to extract distributed patterns of activations from a CNN and show that such patterns correspond to high-level visual attributes. We propose an unsupervised learning module that sits above a pre-trained CNN and learns distributed activation patterns of the network. We utilize elastic non-negative matrix factorization to analyze the responses of a pretrained CNN to an input image and extract salient image regions. The corresponding patterns of neural activations for the extracted salient regions are then clustered via unsupervised deep embedding for clustering (DEC) framework. We demonstrate that these distributed activations contain high-level image features that could be explicitly used for image classification.
最近的研究表明,在卷积神经网络(cnn)的激活模式中出现了语义对象部分检测器,但没有考虑到这种网络中的分布式多层神经激活。在这项工作中,我们提出了一种从CNN中提取分布式激活模式的新方法,并表明这些模式对应于高级视觉属性。我们提出了一个无监督学习模块,它位于预训练的CNN之上,并学习网络的分布式激活模式。我们利用弹性非负矩阵分解来分析预训练CNN对输入图像的响应,并提取显著图像区域。然后通过无监督深度嵌入聚类(DEC)框架对提取的显著区域的相应神经激活模式进行聚类。我们证明了这些分布式激活包含可以显式用于图像分类的高级图像特征。
{"title":"Explaining Distributed Neural Activations via Unsupervised Learning","authors":"Soheil Kolouri, Charles E. Martin, Heiko Hoffmann","doi":"10.1109/CVPRW.2017.213","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.213","url":null,"abstract":"Recent work has demonstrated the emergence of semantic object-part detectors in activation patterns of convolutional neural networks (CNNs), but did not account for the distributed multi-layer neural activations in such networks. In this work, we propose a novel method to extract distributed patterns of activations from a CNN and show that such patterns correspond to high-level visual attributes. We propose an unsupervised learning module that sits above a pre-trained CNN and learns distributed activation patterns of the network. We utilize elastic non-negative matrix factorization to analyze the responses of a pretrained CNN to an input image and extract salient image regions. The corresponding patterns of neural activations for the extracted salient regions are then clustered via unsupervised deep embedding for clustering (DEC) framework. We demonstrate that these distributed activations contain high-level image features that could be explicitly used for image classification.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"58 1","pages":"1670-1678"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80624811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Detection of Metadata Tampering Through Discrepancy Between Image Content and Metadata Using Multi-task Deep Learning 基于多任务深度学习的图像内容与元数据差异的元数据篡改检测
Bor-Chun Chen, P. Ghosh, Vlad I. Morariu, L. Davis
Image content or metadata editing software availability and ease of use has resulted in a high demand for automatic image tamper detection algorithms. Most previous work has focused on detection of tampered image content, whereas we develop techniques to detect metadata tampering in outdoor images using sun altitude angle and other meteorological information like temperature, humidity and weather, which can be observed in most outdoor image scenes. To train and evaluate our technique, we create a large dataset of outdoor images labeled with sun altitude angle and other meteorological data (AMOS+M2), which to our knowledge, is the largest publicly available dataset of its kind. Using this dataset, we train separate regression models for sun altitude angle, temperature and humidity and a classification model for weather to detect any discrepancy between image content and its metadata. Finally, a joint multi-task network for these four features shows a relative improvement of 15.5% compared to each of them individually. We include a detailed analysis for using these networks to detect various types of modification to location and time information in image metadata.
图像内容或元数据编辑软件的可用性和易用性导致了对自动图像篡改检测算法的高需求。以前的大部分工作都集中在检测篡改图像内容上,而我们开发了利用太阳高度角和其他气象信息(如温度、湿度和天气)检测室外图像元数据篡改的技术,这些信息可以在大多数室外图像场景中观察到。为了训练和评估我们的技术,我们创建了一个大型户外图像数据集,其中标记了太阳高度角和其他气象数据(AMOS+M2),据我们所知,这是同类数据集中最大的公开数据集。使用该数据集,我们训练了太阳高度角、温度和湿度的独立回归模型,以及天气的分类模型,以检测图像内容与其元数据之间的任何差异。最后,对于这四个特征的联合多任务网络,与单独使用它们相比,其相对改进率为15.5%。我们详细分析了如何使用这些网络来检测图像元数据中位置和时间信息的各种类型的修改。
{"title":"Detection of Metadata Tampering Through Discrepancy Between Image Content and Metadata Using Multi-task Deep Learning","authors":"Bor-Chun Chen, P. Ghosh, Vlad I. Morariu, L. Davis","doi":"10.1109/CVPRW.2017.234","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.234","url":null,"abstract":"Image content or metadata editing software availability and ease of use has resulted in a high demand for automatic image tamper detection algorithms. Most previous work has focused on detection of tampered image content, whereas we develop techniques to detect metadata tampering in outdoor images using sun altitude angle and other meteorological information like temperature, humidity and weather, which can be observed in most outdoor image scenes. To train and evaluate our technique, we create a large dataset of outdoor images labeled with sun altitude angle and other meteorological data (AMOS+M2), which to our knowledge, is the largest publicly available dataset of its kind. Using this dataset, we train separate regression models for sun altitude angle, temperature and humidity and a classification model for weather to detect any discrepancy between image content and its metadata. Finally, a joint multi-task network for these four features shows a relative improvement of 15.5% compared to each of them individually. We include a detailed analysis for using these networks to detect various types of modification to location and time information in image metadata.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"17 1","pages":"1872-1880"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91141718","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Pruning ConvNets Online for Efficient Specialist Models 高效专家模型的在线修剪卷积神经网络
Jia Guo, M. Potkonjak
Convolutional neural networks (CNNs) excel in various computer vision related tasks but are extremely computationally intensive and power hungry to run on mobile and embedded devices. Recent pruning techniques can reduce the computation and memory requirements of CNNs, but a costly retraining step is needed to restore the classification accuracy of the pruned model. In this paper, we present evidence that when only a subset of the classes need to be classified, we could prune a model and achieve reasonable classification accuracy without retraining. The resulting specialist model will require less energy and time to run than the original full model. To compensate for the pruning, we take advantage of the redundancy among filters and class-specific features. We show that even simple methods such as replacing channels with mean or with the most correlated channel can boost the accuracy of the pruned model to reasonable levels.
卷积神经网络(cnn)在各种与计算机视觉相关的任务中表现出色,但在移动和嵌入式设备上运行时,其计算密集型和耗电量极大。最近的剪枝技术可以减少cnn的计算量和内存需求,但需要一个昂贵的再训练步骤来恢复剪枝模型的分类精度。在本文中,我们提供的证据表明,当只需要分类类的一个子集时,我们可以在不重新训练的情况下修剪模型并获得合理的分类精度。由此产生的专家模型将比原始的完整模型需要更少的精力和时间来运行。为了补偿这种修剪,我们利用了过滤器和类特定特征之间的冗余。我们表明,即使是简单的方法,如用均值或最相关的通道替换通道,也可以将修剪模型的精度提高到合理的水平。
{"title":"Pruning ConvNets Online for Efficient Specialist Models","authors":"Jia Guo, M. Potkonjak","doi":"10.1109/CVPRW.2017.58","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.58","url":null,"abstract":"Convolutional neural networks (CNNs) excel in various computer vision related tasks but are extremely computationally intensive and power hungry to run on mobile and embedded devices. Recent pruning techniques can reduce the computation and memory requirements of CNNs, but a costly retraining step is needed to restore the classification accuracy of the pruned model. In this paper, we present evidence that when only a subset of the classes need to be classified, we could prune a model and achieve reasonable classification accuracy without retraining. The resulting specialist model will require less energy and time to run than the original full model. To compensate for the pruning, we take advantage of the redundancy among filters and class-specific features. We show that even simple methods such as replacing channels with mean or with the most correlated channel can boost the accuracy of the pruned model to reasonable levels.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"17 2 1","pages":"430-437"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91206930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Component Biologically Inspired Features with Moving Segmentation for Age Estimation 基于移动分割的年龄估计组件生物学启发特征
G. Hsu, Yi-Tseng Cheng, Choon-Ching Ng, Moi Hoon Yap
We propose the Component Bio-Inspired Feature (CBIF) with a moving segmentation scheme for age estimation. The CBIF defines a superset for the commonly used Bio-Inspired Feature (BIF) with more parameters and flexibility in settings, resulting in features with abundant characteristics. An in-depth study is performed for the determination of the parameters good for capturing age-related traits. The moving segmentation is proposed to better determine the age boundaries good for age grouping, and improve the overall performance. The proposed approach is evaluated on two common benchmarks, FG-NET and MORPH databases, and compared with contemporary approaches to demonstrate its efficacy.
我们提出了一种带有运动分割方案的成分生物启发特征(cif)用于年龄估计。cif为常用的生物启发特征(biif)定义了一个具有更多参数和设置灵活性的超集,从而产生具有丰富特征的特征。进行了深入的研究,以确定有利于捕获年龄相关性状的参数。为了更好地确定适合年龄分组的年龄边界,提高整体性能,提出了移动分割方法。该方法在FG-NET和MORPH数据库这两个常用基准上进行了评估,并与当代方法进行了比较,以证明其有效性。
{"title":"Component Biologically Inspired Features with Moving Segmentation for Age Estimation","authors":"G. Hsu, Yi-Tseng Cheng, Choon-Ching Ng, Moi Hoon Yap","doi":"10.1109/CVPRW.2017.81","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.81","url":null,"abstract":"We propose the Component Bio-Inspired Feature (CBIF) with a moving segmentation scheme for age estimation. The CBIF defines a superset for the commonly used Bio-Inspired Feature (BIF) with more parameters and flexibility in settings, resulting in features with abundant characteristics. An in-depth study is performed for the determination of the parameters good for capturing age-related traits. The moving segmentation is proposed to better determine the age boundaries good for age grouping, and improve the overall performance. The proposed approach is evaluated on two common benchmarks, FG-NET and MORPH databases, and compared with contemporary approaches to demonstrate its efficacy.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"36 1","pages":"540-547"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76609977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Transferable Deep-CNN Features for Detecting Digital and Print-Scanned Morphed Face Images 用于检测数字和打印扫描变形面部图像的可转移深度cnn特征
Ramachandra Raghavendra, K. Raja, S. Venkatesh, C. Busch
Face biometrics is widely used in various applications including border control and facilitating the verification of travellers' identity claim with respect to his electronic passport (ePass). As in most countries, passports are issued to a citizen based on the submitted photo which allows the applicant to provide a morphed face photo to conceal his identity during the application process. In this work, we propose a novel approach leveraging the transferable features from a pre-trained Deep Convolutional Neural Networks (D-CNN) to detect both digital and print-scanned morphed face image. Thus, the proposed approach is based on the feature level fusion of the first fully connected layers of two D-CNN (VGG19 and AlexNet) that are specifically fine-tuned using the morphed face image database. The proposed method is extensively evaluated on the newly constructed database with both digital and print-scanned morphed face images corresponding to bona fide and morphed data reflecting a real-life scenario. The obtained results consistently demonstrate improved detection performance of the proposed scheme over previously proposed methods on both the digital and the print-scanned morphed face image database.
面部生物识别技术广泛应用于各种应用,包括边境管制和方便核实旅客电子护照的身份要求。与大多数国家一样,护照是根据申请人提交的照片发放的,申请人可以在申请过程中提供一张变形的脸部照片来隐藏自己的身份。在这项工作中,我们提出了一种新的方法,利用预训练的深度卷积神经网络(D-CNN)的可转移特征来检测数字和打印扫描的变形人脸图像。因此,所提出的方法是基于两个D-CNN (VGG19和AlexNet)的第一个完全连接层的特征级融合,这些层是使用变形的人脸图像数据库进行特别微调的。该方法在新建立的数据库上进行了广泛的评估,该数据库包含数字和打印扫描的变形人脸图像,这些图像对应于真实的和反映现实场景的变形数据。所得结果一致表明,该方法在数字和打印扫描变形人脸图像数据库上的检测性能优于先前提出的方法。
{"title":"Transferable Deep-CNN Features for Detecting Digital and Print-Scanned Morphed Face Images","authors":"Ramachandra Raghavendra, K. Raja, S. Venkatesh, C. Busch","doi":"10.1109/CVPRW.2017.228","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.228","url":null,"abstract":"Face biometrics is widely used in various applications including border control and facilitating the verification of travellers' identity claim with respect to his electronic passport (ePass). As in most countries, passports are issued to a citizen based on the submitted photo which allows the applicant to provide a morphed face photo to conceal his identity during the application process. In this work, we propose a novel approach leveraging the transferable features from a pre-trained Deep Convolutional Neural Networks (D-CNN) to detect both digital and print-scanned morphed face image. Thus, the proposed approach is based on the feature level fusion of the first fully connected layers of two D-CNN (VGG19 and AlexNet) that are specifically fine-tuned using the morphed face image database. The proposed method is extensively evaluated on the newly constructed database with both digital and print-scanned morphed face images corresponding to bona fide and morphed data reflecting a real-life scenario. The obtained results consistently demonstrate improved detection performance of the proposed scheme over previously proposed methods on both the digital and the print-scanned morphed face image database.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"146 1","pages":"1822-1830"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76611081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 158
期刊
2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1