首页 > 最新文献

2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)最新文献

英文 中文
Enhanced Deep Residual Networks for Single Image Super-Resolution 单幅图像超分辨率增强深度残差网络
Bee Lim, Sanghyun Son, Heewon Kim, Seungjun Nah, Kyoung Mu Lee
Recent research on super-resolution has progressed with the development of deep convolutional neural networks (DCNN). In particular, residual learning techniques exhibit improved performance. In this paper, we develop an enhanced deep super-resolution network (EDSR) with performance exceeding those of current state-of-the-art SR methods. The significant performance improvement of our model is due to optimization by removing unnecessary modules in conventional residual networks. The performance is further improved by expanding the model size while we stabilize the training procedure. We also propose a new multi-scale deep super-resolution system (MDSR) and training method, which can reconstruct high-resolution images of different upscaling factors in a single model. The proposed methods show superior performance over the state-of-the-art methods on benchmark datasets and prove its excellence by winning the NTIRE2017 Super-Resolution Challenge[26].
随着深度卷积神经网络(DCNN)的发展,近年来对超分辨率的研究取得了进展。特别是,残差学习技术表现出更好的性能。在本文中,我们开发了一种增强的深度超分辨率网络(EDSR),其性能超过了目前最先进的深度超分辨率网络方法。我们的模型的显著性能改进是由于通过去除传统残余网络中不必要的模块进行优化。在稳定训练过程的同时,通过扩大模型尺寸进一步提高了性能。我们还提出了一种新的多尺度深度超分辨率系统(MDSR)和训练方法,可以在单个模型中重建不同上尺度因子的高分辨率图像。所提出的方法在基准数据集上表现出优于最先进方法的性能,并通过赢得NTIRE2017超分辨率挑战赛证明了其卓越性[26]。
{"title":"Enhanced Deep Residual Networks for Single Image Super-Resolution","authors":"Bee Lim, Sanghyun Son, Heewon Kim, Seungjun Nah, Kyoung Mu Lee","doi":"10.1109/CVPRW.2017.151","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.151","url":null,"abstract":"Recent research on super-resolution has progressed with the development of deep convolutional neural networks (DCNN). In particular, residual learning techniques exhibit improved performance. In this paper, we develop an enhanced deep super-resolution network (EDSR) with performance exceeding those of current state-of-the-art SR methods. The significant performance improvement of our model is due to optimization by removing unnecessary modules in conventional residual networks. The performance is further improved by expanding the model size while we stabilize the training procedure. We also propose a new multi-scale deep super-resolution system (MDSR) and training method, which can reconstruct high-resolution images of different upscaling factors in a single model. The proposed methods show superior performance over the state-of-the-art methods on benchmark datasets and prove its excellence by winning the NTIRE2017 Super-Resolution Challenge[26].","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"2 1","pages":"1132-1140"},"PeriodicalIF":0.0,"publicationDate":"2017-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79424971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4440
Delineation of Skin Strata in Reflectance Confocal Microscopy Images with Recurrent Convolutional Networks 用循环卷积网络描述反射共聚焦显微镜图像中的皮肤层
A. Bozkurt, Trevor Gale, Kivanç Köse, C. Alessi-Fox, D. Brooks, M. Rajadhyaksha, Jennifer G. Dy
Reflectance confocal microscopy (RCM) is an effective, non-invasive pre-screening tool for cancer diagnosis. However, acquiring and reading RCM images requires extensive training and experience, and novice clinicians exhibit high variance in diagnostic accuracy. Consequently, there is a compelling need for quantitative tools to standardize image acquisition and analysis. In this study, we use deep recurrent convolutional neural networks to delineate skin strata in stacks of RCM images collected at consecutive depths. To perform diagnostic analysis, clinicians collect RCM images at 4-5 specific layers in the tissue. Our model automates this process by discriminating between RCM images of different layers. Testing our model on an expert labeled dataset of 504 RCM stacks, we achieve 87.97% classification accuracy, and a 9-fold reduction in the number of anatomically impossible errors compared to the previous state-of-the-art.
反射共聚焦显微镜(RCM)是一种有效的、非侵入性的癌症诊断预筛查工具。然而,获取和读取RCM图像需要广泛的培训和经验,新手临床医生在诊断准确性方面表现出很大的差异。因此,迫切需要定量工具来标准化图像采集和分析。在这项研究中,我们使用深度递归卷积神经网络来描绘在连续深度收集的RCM图像堆栈中的皮肤层。为了进行诊断分析,临床医生收集组织中4-5个特定层的RCM图像。我们的模型通过区分不同层的RCM图像来自动化这一过程。在504个RCM堆叠的专家标记数据集上测试我们的模型,我们实现了87.97%的分类准确率,与以前的最先进技术相比,解剖学上不可能的错误数量减少了9倍。
{"title":"Delineation of Skin Strata in Reflectance Confocal Microscopy Images with Recurrent Convolutional Networks","authors":"A. Bozkurt, Trevor Gale, Kivanç Köse, C. Alessi-Fox, D. Brooks, M. Rajadhyaksha, Jennifer G. Dy","doi":"10.1109/CVPRW.2017.108","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.108","url":null,"abstract":"Reflectance confocal microscopy (RCM) is an effective, non-invasive pre-screening tool for cancer diagnosis. However, acquiring and reading RCM images requires extensive training and experience, and novice clinicians exhibit high variance in diagnostic accuracy. Consequently, there is a compelling need for quantitative tools to standardize image acquisition and analysis. In this study, we use deep recurrent convolutional neural networks to delineate skin strata in stacks of RCM images collected at consecutive depths. To perform diagnostic analysis, clinicians collect RCM images at 4-5 specific layers in the tissue. Our model automates this process by discriminating between RCM images of different layers. Testing our model on an expert labeled dataset of 504 RCM stacks, we achieve 87.97% classification accuracy, and a 9-fold reduction in the number of anatomically impossible errors compared to the previous state-of-the-art.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"34 1","pages":"777-785"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75421725","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Manifold Guided Label Transfer for Deep Domain Adaptation 深度域自适应的流形引导标签转移
Breton L. Minnehan, A. Savakis
We propose a novel domain adaptation method for deep learning that combines adaptive batch normalization to produce a common feature-space between domains and label transfer with subspace alignment on deep features. The first step of our method automatically conditions the features from the source/target domain to have similar statistical distributions by normalizing the activations in each layer of our network using adaptive batch normalization. We then examine the clustering properties of the normalized features on a manifold to determine if the target features are well suited for the second of our algorithm, label-transfer. The second step of our method performs subspace alignment and k-means clustering on the feature manifold to transfer labels from the closest source cluster to each target cluster. The proposed manifold guided label transfer methods produce state of the art results for deep adaptation on several standard digit recognition datasets.
我们提出了一种新的深度学习领域自适应方法,该方法结合了自适应批归一化来产生域之间的公共特征空间和深度特征上具有子空间对齐的标签转移。我们方法的第一步是通过使用自适应批归一化对网络每层中的激活进行归一化,自动地将源/目标域的特征条件化,使其具有相似的统计分布。然后,我们检查流形上归一化特征的聚类属性,以确定目标特征是否非常适合我们的算法的第二步,标签转移。该方法的第二步在特征流形上执行子空间对齐和k-means聚类,将标签从最近的源聚类转移到每个目标聚类。所提出的歧管引导标签转移方法产生了对几个标准数字识别数据集进行深度适应的最新结果。
{"title":"Manifold Guided Label Transfer for Deep Domain Adaptation","authors":"Breton L. Minnehan, A. Savakis","doi":"10.1109/CVPRW.2017.104","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.104","url":null,"abstract":"We propose a novel domain adaptation method for deep learning that combines adaptive batch normalization to produce a common feature-space between domains and label transfer with subspace alignment on deep features. The first step of our method automatically conditions the features from the source/target domain to have similar statistical distributions by normalizing the activations in each layer of our network using adaptive batch normalization. We then examine the clustering properties of the normalized features on a manifold to determine if the target features are well suited for the second of our algorithm, label-transfer. The second step of our method performs subspace alignment and k-means clustering on the feature manifold to transfer labels from the closest source cluster to each target cluster. The proposed manifold guided label transfer methods produce state of the art results for deep adaptation on several standard digit recognition datasets.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"6 1","pages":"744-752"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78430167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Temporally Steered Gaussian Attention for Video Understanding 用于视频理解的时间导向高斯注意
Shagan Sah, Thang Nguyen, Miguel Domínguez, F. Such, R. Ptucha
Recent advances in video understanding are enabling incredible developments in video search, summarization, automatic captioning and human computer interaction. Attention mechanisms are a powerful way to steer focus onto different sections of the video. Existing mechanisms are driven by prior training probabilities and require input instances of identical temporal duration. We introduce an intuitive video understanding framework which combines continuous attention mechanisms over a family of Gaussian distributions with a hierarchical based video representation. The hierarchical framework enables efficient abstract temporal representations of video. Video attributes steer the attention mechanism intelligently independent of video length. Our fully learnable end-to-end approach helps predict salient temporal regions of action/objects in the video. We demonstrate state-of-the-art captioning results on the popular MSVD, MSR-VTT and M-VAD video datasets.
视频理解的最新进展使视频搜索、摘要、自动字幕和人机交互取得了令人难以置信的发展。注意机制是一种强大的方法,可以将注意力引导到视频的不同部分。现有的机制是由先验训练概率驱动的,并且需要相同时间持续时间的输入实例。我们引入了一个直观的视频理解框架,它结合了高斯分布家族上的连续关注机制和基于层次的视频表示。分层框架实现了视频的高效抽象时间表示。视频属性独立于视频长度,智能地引导注意力机制。我们完全可学习的端到端方法有助于预测视频中动作/对象的显著时间区域。我们在流行的MSVD, MSR-VTT和M-VAD视频数据集上展示了最先进的字幕结果。
{"title":"Temporally Steered Gaussian Attention for Video Understanding","authors":"Shagan Sah, Thang Nguyen, Miguel Domínguez, F. Such, R. Ptucha","doi":"10.1109/CVPRW.2017.274","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.274","url":null,"abstract":"Recent advances in video understanding are enabling incredible developments in video search, summarization, automatic captioning and human computer interaction. Attention mechanisms are a powerful way to steer focus onto different sections of the video. Existing mechanisms are driven by prior training probabilities and require input instances of identical temporal duration. We introduce an intuitive video understanding framework which combines continuous attention mechanisms over a family of Gaussian distributions with a hierarchical based video representation. The hierarchical framework enables efficient abstract temporal representations of video. Video attributes steer the attention mechanism intelligently independent of video length. Our fully learnable end-to-end approach helps predict salient temporal regions of action/objects in the video. We demonstrate state-of-the-art captioning results on the popular MSVD, MSR-VTT and M-VAD video datasets.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"64 1","pages":"2208-2216"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76098011","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
SRHRF+: Self-Example Enhanced Single Image Super-Resolution Using Hierarchical Random Forests SRHRF+:使用分层随机森林增强单幅图像超分辨率的自我示例
Jun-Jie Huang, Tian-Rui Liu, P. Dragotti, T. Stathaki
Example-based single image super-resolution (SISR) methods use external training datasets and have recently attracted a lot of interest. Self-example based SISR methods exploit redundant non-local self-similar patterns in natural images and because of that are more able to adapt to the image at hand to generate high quality super-resolved images. In this paper, we propose to combine the advantages of example-based SISR and self-example based SISR. A novel hierarchical random forests based super-resolution (SRHRF) method is proposed to learn statistical priors from external training images. Each layer of random forests reduce the estimation error due to variance by aggregating prediction models from multiple decision trees. The hierarchical structure further boosts the performance by pushing the estimation error due to bias towards zero. In order to further adaptively improve the super-resolved image, a self-example random forests (SERF) is learned from an image pyramid pair constructed from the down-sampled SRHRF generated result. Extensive numerical results show that the SRHRF method enhanced using SERF (SRHRF+) achieves the state-of-the-art performance on natural images and yields substantially superior performance for image with rich self-similar patterns.
基于示例的单幅图像超分辨率(SISR)方法使用外部训练数据集,最近引起了人们的广泛关注。基于自示例的SISR方法利用了自然图像中冗余的非局部自相似模式,因此更能适应手头的图像,从而生成高质量的超分辨率图像。在本文中,我们提出将基于示例的SISR和基于自示例的SISR的优势结合起来。提出了一种基于分层随机森林的超分辨率(SRHRF)方法,从外部训练图像中学习统计先验。随机森林的每一层通过聚合来自多个决策树的预测模型来减少由于方差引起的估计误差。分层结构通过将由于偏差引起的估计误差推向零进一步提高了性能。为了进一步自适应改进超分辨图像,从下采样SRHRF生成的图像金字塔对中学习自例随机森林(SERF)。大量的数值结果表明,使用SERF (SRHRF+)增强的SRHRF方法在自然图像上达到了最先进的性能,并且在具有丰富自相似模式的图像上产生了显著的优异性能。
{"title":"SRHRF+: Self-Example Enhanced Single Image Super-Resolution Using Hierarchical Random Forests","authors":"Jun-Jie Huang, Tian-Rui Liu, P. Dragotti, T. Stathaki","doi":"10.1109/CVPRW.2017.144","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.144","url":null,"abstract":"Example-based single image super-resolution (SISR) methods use external training datasets and have recently attracted a lot of interest. Self-example based SISR methods exploit redundant non-local self-similar patterns in natural images and because of that are more able to adapt to the image at hand to generate high quality super-resolved images. In this paper, we propose to combine the advantages of example-based SISR and self-example based SISR. A novel hierarchical random forests based super-resolution (SRHRF) method is proposed to learn statistical priors from external training images. Each layer of random forests reduce the estimation error due to variance by aggregating prediction models from multiple decision trees. The hierarchical structure further boosts the performance by pushing the estimation error due to bias towards zero. In order to further adaptively improve the super-resolved image, a self-example random forests (SERF) is learned from an image pyramid pair constructed from the down-sampled SRHRF generated result. Extensive numerical results show that the SRHRF method enhanced using SERF (SRHRF+) achieves the state-of-the-art performance on natural images and yields substantially superior performance for image with rich self-similar patterns.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"12 1","pages":"1067-1075"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75088054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Hand Movement Prediction Based Collision-Free Human-Robot Interaction 基于无碰撞人机交互的手部运动预测
Yiwei Wang, Xin Ye, Yezhou Yang, Wenlong Zhang
We present a framework from vision based hand movement prediction in a real-world human-robot collaborative scenario for safety guarantee. We first propose a perception submodule that takes in visual data solely and predicts human collaborator's hand movement. Then a robot trajectory adaptive planning submodule is developed that takes the noisy movement prediction signal into consideration for optimization. We first collect a new human manipulation dataset that can supplement the previous publicly available dataset with motion capture data to serve as the ground truth of hand location. We then integrate the algorithm with a robot manipulator that can collaborate with human workers on a set of trained manipulation actions, and it is shown that such a robot system outperforms the one without movement prediction in terms of collision avoidance.
我们提出了一个基于视觉的手部运动预测框架,用于现实世界人机协作场景的安全保障。我们首先提出了一个感知子模块,它只接受视觉数据并预测人类合作者的手部运动。然后开发了考虑运动预测信号噪声的机器人轨迹自适应规划子模块进行优化。我们首先收集了一个新的人类操作数据集,该数据集可以用动作捕捉数据补充以前公开可用的数据集,作为手部位置的基础真相。然后,我们将该算法与一个可以与人类工人在一组训练过的操作动作上协作的机器人操作器集成,结果表明,这样的机器人系统在避免碰撞方面优于没有运动预测的机器人系统。
{"title":"Hand Movement Prediction Based Collision-Free Human-Robot Interaction","authors":"Yiwei Wang, Xin Ye, Yezhou Yang, Wenlong Zhang","doi":"10.1109/CVPRW.2017.72","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.72","url":null,"abstract":"We present a framework from vision based hand movement prediction in a real-world human-robot collaborative scenario for safety guarantee. We first propose a perception submodule that takes in visual data solely and predicts human collaborator's hand movement. Then a robot trajectory adaptive planning submodule is developed that takes the noisy movement prediction signal into consideration for optimization. We first collect a new human manipulation dataset that can supplement the previous publicly available dataset with motion capture data to serve as the ground truth of hand location. We then integrate the algorithm with a robot manipulator that can collaborate with human workers on a set of trained manipulation actions, and it is shown that such a robot system outperforms the one without movement prediction in terms of collision avoidance.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"43 1","pages":"492-493"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72657074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Exploiting Reflectional and Rotational Invariance in Single Image Superresolution 利用单幅图像超分辨率的反射和旋转不变性
S. Donné, Laurens Meeus, H. Luong, B. Goossens, W. Philips
Stationarity of reconstruction problems is the crux to enabling convolutional neural networks for many image processing tasks: the output estimate for a pixel is generally not dependent on its location within the image but only on its immediate neighbourhood. We expect other invariances, too. For most pixel-processing tasks, rigid transformations should commute with the processing: a rigid transformation of the input should result in that same transformation of the output. In existing literature this is taken into account indirectly by augmenting the training set: reflected and rotated versions of the inputs are also fed to the network when optimizing the network weights. In contrast, we enforce this invariance through the network design. Because of the encompassing nature of the proposed architecture, it can directly enhance existing CNN-based algorithms. We show how it can be applied to SRCNN and FSRCNN both, speeding up convergence in the initial training phase, and improving performance both for pretrained weights and after finetuning.
重建问题的平稳性是卷积神经网络实现许多图像处理任务的关键:一个像素的输出估计通常不依赖于它在图像中的位置,而只依赖于它的近邻。我们还期待其他的不变性。对于大多数像素处理任务,严格的转换应该与处理同步进行:输入的严格转换应该导致输出的相同转换。在现有文献中,这是通过增加训练集来间接考虑的:在优化网络权重时,也将输入的反射和旋转版本馈给网络。相反,我们通过网络设计来加强这种不变性。由于所提出的架构的包涵性,它可以直接增强现有的基于cnn的算法。我们展示了如何将其应用于SRCNN和FSRCNN,加速初始训练阶段的收敛,并提高预训练权值和微调后的性能。
{"title":"Exploiting Reflectional and Rotational Invariance in Single Image Superresolution","authors":"S. Donné, Laurens Meeus, H. Luong, B. Goossens, W. Philips","doi":"10.1109/CVPRW.2017.141","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.141","url":null,"abstract":"Stationarity of reconstruction problems is the crux to enabling convolutional neural networks for many image processing tasks: the output estimate for a pixel is generally not dependent on its location within the image but only on its immediate neighbourhood. We expect other invariances, too. For most pixel-processing tasks, rigid transformations should commute with the processing: a rigid transformation of the input should result in that same transformation of the output. In existing literature this is taken into account indirectly by augmenting the training set: reflected and rotated versions of the inputs are also fed to the network when optimizing the network weights. In contrast, we enforce this invariance through the network design. Because of the encompassing nature of the proposed architecture, it can directly enhance existing CNN-based algorithms. We show how it can be applied to SRCNN and FSRCNN both, speeding up convergence in the initial training phase, and improving performance both for pretrained weights and after finetuning.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"22 1","pages":"1043-1049"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75124230","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Tampering Detection and Localization Through Clustering of Camera-Based CNN Features 基于摄像机CNN特征聚类的篡改检测与定位
L. Bondi, S. Lameri, David Guera, Paolo Bestagini, E. Delp, S. Tubaro
Due to the rapid proliferation of image capturing devices and user-friendly editing software suites, image manipulation is at everyone's hand. For this reason, the forensic community has developed a series of techniques to determine image authenticity. In this paper, we propose an algorithm for image tampering detection and localization, leveraging characteristic footprints left on images by different camera models. The rationale behind our algorithm is that all pixels of pristine images should be detected as being shot with a single device. Conversely, if a picture is obtained through image composition, traces of multiple devices can be detected. The proposed algorithm exploits a convolutional neural network (CNN) to extract characteristic camera model features from image patches. These features are then analyzed by means of iterative clustering techniques in order to detect whether an image has been forged, and localize the alien region.
由于图像捕捉设备和用户友好的编辑软件套件的快速扩散,图像处理是在每个人的手中。为此,法医学界开发了一系列技术来确定图像的真实性。在本文中,我们提出了一种图像篡改检测和定位算法,利用不同相机型号在图像上留下的特征足迹。我们的算法背后的基本原理是,原始图像的所有像素都应该被检测为使用单个设备拍摄的。相反,如果通过图像合成获得图像,则可以检测到多个设备的痕迹。该算法利用卷积神经网络(CNN)从图像patch中提取相机模型特征。然后通过迭代聚类技术对这些特征进行分析,以检测图像是否被伪造,并定位外来区域。
{"title":"Tampering Detection and Localization Through Clustering of Camera-Based CNN Features","authors":"L. Bondi, S. Lameri, David Guera, Paolo Bestagini, E. Delp, S. Tubaro","doi":"10.1109/CVPRW.2017.232","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.232","url":null,"abstract":"Due to the rapid proliferation of image capturing devices and user-friendly editing software suites, image manipulation is at everyone's hand. For this reason, the forensic community has developed a series of techniques to determine image authenticity. In this paper, we propose an algorithm for image tampering detection and localization, leveraging characteristic footprints left on images by different camera models. The rationale behind our algorithm is that all pixels of pristine images should be detected as being shot with a single device. Conversely, if a picture is obtained through image composition, traces of multiple devices can be detected. The proposed algorithm exploits a convolutional neural network (CNN) to extract characteristic camera model features from image patches. These features are then analyzed by means of iterative clustering techniques in order to detect whether an image has been forged, and localize the alien region.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"3 1","pages":"1855-1864"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77348230","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 157
A Level Set Method for Gland Segmentation 腺体分割的水平集方法
Chen Wang, H. Bu, J. Bao, Chunming Li
Histopathology plays a role as the gold standard in clinic for disease diagnosis. The identification and segmentation of histological structures are the prerequisite to disease diagnosis. With the advent of digital pathology, researchers' attention is attracted by the analysis of digital pathology images. In order to relieve the workload on pathologists, a robust segmentation method is needed in clinic for computer-assisted diagnosis. In this paper, we propose a level set framework to achieve gland image segmentation. The input image is divided into two parts, which contain glands with lumens and glands without lumens, respectively. Our experiments are performed on the clinical datasets of West China Hospital, Sichuan University. The experimental results show that our method can deal with glands without lumens, thus can obtain a better performance.
组织病理学是临床疾病诊断的金标准。组织结构的识别和分割是疾病诊断的前提。随着数字病理学的出现,数字病理图像的分析引起了研究人员的关注。为了减轻病理医师的工作量,需要一种鲁棒的分割方法用于临床计算机辅助诊断。在本文中,我们提出了一个水平集框架来实现腺体图像分割。将输入图像分为两部分,分别包含有管腔的腺体和没有管腔的腺体。我们的实验在四川大学华西医院的临床数据集上进行。实验结果表明,该方法可以处理没有腔体的腺体,从而获得更好的性能。
{"title":"A Level Set Method for Gland Segmentation","authors":"Chen Wang, H. Bu, J. Bao, Chunming Li","doi":"10.1109/CVPRW.2017.120","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.120","url":null,"abstract":"Histopathology plays a role as the gold standard in clinic for disease diagnosis. The identification and segmentation of histological structures are the prerequisite to disease diagnosis. With the advent of digital pathology, researchers' attention is attracted by the analysis of digital pathology images. In order to relieve the workload on pathologists, a robust segmentation method is needed in clinic for computer-assisted diagnosis. In this paper, we propose a level set framework to achieve gland image segmentation. The input image is divided into two parts, which contain glands with lumens and glands without lumens, respectively. Our experiments are performed on the clinical datasets of West China Hospital, Sichuan University. The experimental results show that our method can deal with glands without lumens, thus can obtain a better performance.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"659 1","pages":"865-873"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76843794","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Improved Cooperative Stereo Matching for Dynamic Vision Sensors with Ground Truth Evaluation 基于地面真值评估的动态视觉传感器改进协同立体匹配
E. Piatkowska, J. Kogler, A. Belbachir, M. Gelautz
Event-based vision, as realized by bio-inspired Dynamic Vision Sensors (DVS), is gaining more and more popularity due to its advantages of high temporal resolution, wide dynamic range and power efficiency at the same time. Potential applications include surveillance, robotics, and autonomous navigation under uncontrolled environment conditions. In this paper, we deal with event-based vision for 3D reconstruction of dynamic scene content by using two stationary DVS in a stereo configuration. We focus on a cooperative stereo approach and suggest an improvement over a previously published algorithm that reduces the measured mean error by over 50 percent. An available ground truth data set for stereo event data is utilized to analyze the algorithm's sensitivity to parameter variation and for comparison with competing techniques.
生物动态视觉传感器(DVS)实现的基于事件的视觉以其高时间分辨率、宽动态范围和高能效等优点得到越来越广泛的应用。潜在的应用包括监视、机器人和在不受控制的环境条件下的自主导航。在本文中,我们通过在立体配置中使用两个固定的DVS来处理基于事件的视觉用于动态场景内容的3D重建。我们专注于合作立体方法,并建议对先前发表的算法进行改进,该算法可将测量平均误差降低50%以上。利用现有的立体事件地面真值数据集,分析了该算法对参数变化的敏感性,并与竞争技术进行了比较。
{"title":"Improved Cooperative Stereo Matching for Dynamic Vision Sensors with Ground Truth Evaluation","authors":"E. Piatkowska, J. Kogler, A. Belbachir, M. Gelautz","doi":"10.1109/CVPRW.2017.51","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.51","url":null,"abstract":"Event-based vision, as realized by bio-inspired Dynamic Vision Sensors (DVS), is gaining more and more popularity due to its advantages of high temporal resolution, wide dynamic range and power efficiency at the same time. Potential applications include surveillance, robotics, and autonomous navigation under uncontrolled environment conditions. In this paper, we deal with event-based vision for 3D reconstruction of dynamic scene content by using two stationary DVS in a stereo configuration. We focus on a cooperative stereo approach and suggest an improvement over a previously published algorithm that reduces the measured mean error by over 50 percent. An available ground truth data set for stereo event data is utilized to analyze the algorithm's sensitivity to parameter variation and for comparison with competing techniques.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"05 1","pages":"370-377"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85850247","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
期刊
2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1