首页 > 最新文献

2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)最新文献

英文 中文
Co-Saliency Detection via Mask-Guided Fully Convolutional Networks With Multi-Scale Label Smoothing 基于多尺度标签平滑的掩模引导全卷积网络的协显著性检测
Pub Date : 2019-06-01 DOI: 10.1109/CVPR.2019.00321
Kaihua Zhang, Tengpeng Li, Bo Liu, Qingshan Liu
In image co-saliency detection problem, one critical issue is how to model the concurrent pattern of the co-salient parts, which appears both within each image and across all the relevant images. In this paper, we propose a hierarchical image co-saliency detection framework as a coarse to fine strategy to capture this pattern. We first propose a mask-guided fully convolutional network structure to generate the initial co-saliency detection result. The mask is used for background removal and it is learned from the high-level feature response maps of the pre-trained VGG-net output. We next propose a multi-scale label smoothing model to further refine the detection result. The proposed model jointly optimizes the label smoothness of pixels and superpixels. Experiment results on three popular image co-saliency detection benchmark datasets including iCoseg, MSRC and Cosal2015 demonstrate the remarkable performance compared with the state-of-the-art methods.
在图像共显著性检测问题中,一个关键问题是如何对每个图像和所有相关图像中出现的共显著部分的并发模式进行建模。在本文中,我们提出了一种分层图像共显著性检测框架,作为一种从粗到细的策略来捕获这种模式。我们首先提出了一个掩模引导的全卷积网络结构来生成初始的共显著性检测结果。掩码用于背景去除,它是从预训练的VGG-net输出的高级特征响应图中学习的。接下来,我们提出了一个多尺度标签平滑模型来进一步细化检测结果。该模型对像素和超像素的标签平滑度进行了联合优化。在iCoseg、MSRC和Cosal2015三个流行的图像共显著性检测基准数据集上的实验结果表明,与现有方法相比,该方法具有显著的性能。
{"title":"Co-Saliency Detection via Mask-Guided Fully Convolutional Networks With Multi-Scale Label Smoothing","authors":"Kaihua Zhang, Tengpeng Li, Bo Liu, Qingshan Liu","doi":"10.1109/CVPR.2019.00321","DOIUrl":"https://doi.org/10.1109/CVPR.2019.00321","url":null,"abstract":"In image co-saliency detection problem, one critical issue is how to model the concurrent pattern of the co-salient parts, which appears both within each image and across all the relevant images. In this paper, we propose a hierarchical image co-saliency detection framework as a coarse to fine strategy to capture this pattern. We first propose a mask-guided fully convolutional network structure to generate the initial co-saliency detection result. The mask is used for background removal and it is learned from the high-level feature response maps of the pre-trained VGG-net output. We next propose a multi-scale label smoothing model to further refine the detection result. The proposed model jointly optimizes the label smoothness of pixels and superpixels. Experiment results on three popular image co-saliency detection benchmark datasets including iCoseg, MSRC and Cosal2015 demonstrate the remarkable performance compared with the state-of-the-art methods.","PeriodicalId":6711,"journal":{"name":"2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","volume":"100 1","pages":"3090-3099"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77713283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 70
Retrieval-Augmented Convolutional Neural Networks Against Adversarial Examples 对抗实例的检索-增强卷积神经网络
Pub Date : 2019-06-01 DOI: 10.1109/CVPR.2019.01183
Jake Zhao, Kyunghyun Cho
We propose a retrieval-augmented convolutional network (RaCNN) and propose to train it with local mixup, a novel variant of the recently proposed mixup algorithm. The proposed hybrid architecture combining a convolutional network and an off-the-shelf retrieval engine was designed to mitigate the adverse effect of off-manifold adversarial examples, while the proposed local mixup addresses on-manifold ones by explicitly encouraging the classifier to locally behave linearly on the data manifold. Our evaluation of the proposed approach against seven readilyavailable adversarial attacks on three datasets–CIFAR-10, SVHN and ImageNet–demonstrate the improved robustness compared to a vanilla convolutional network, and comparable performance with the state-of-the-art reactive defense approaches.
我们提出了一种检索增强卷积网络(RaCNN),并提出用局部混合训练它,局部混合是最近提出的混合算法的一种新变体。所提出的混合架构结合了卷积网络和现成的检索引擎,旨在减轻非流形对抗示例的不利影响,而所提出的局部混合通过明确鼓励分类器在数据流形上局部线性行为来解决非流形的问题。我们对所提出的方法在三个数据集(cifar -10、SVHN和imagenet)上针对七种现成的对抗性攻击的评估表明,与普通卷积网络相比,该方法的鲁棒性得到了提高,性能与最先进的反应性防御方法相当。
{"title":"Retrieval-Augmented Convolutional Neural Networks Against Adversarial Examples","authors":"Jake Zhao, Kyunghyun Cho","doi":"10.1109/CVPR.2019.01183","DOIUrl":"https://doi.org/10.1109/CVPR.2019.01183","url":null,"abstract":"We propose a retrieval-augmented convolutional network (RaCNN) and propose to train it with local mixup, a novel variant of the recently proposed mixup algorithm. The proposed hybrid architecture combining a convolutional network and an off-the-shelf retrieval engine was designed to mitigate the adverse effect of off-manifold adversarial examples, while the proposed local mixup addresses on-manifold ones by explicitly encouraging the classifier to locally behave linearly on the data manifold. Our evaluation of the proposed approach against seven readilyavailable adversarial attacks on three datasets–CIFAR-10, SVHN and ImageNet–demonstrate the improved robustness compared to a vanilla convolutional network, and comparable performance with the state-of-the-art reactive defense approaches.","PeriodicalId":6711,"journal":{"name":"2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","volume":"30 1","pages":"11555-11563"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77985991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Occlusion-Net: 2D/3D Occluded Keypoint Localization Using Graph Networks Occlusion-Net:使用图网络进行2D/3D遮挡关键点定位
Pub Date : 2019-06-01 DOI: 10.1109/CVPR.2019.00750
Dinesh Reddy Narapureddy, Minh Vo, S. Narasimhan
We present Occlusion-Net, a framework to predict 2D and 3D locations of occluded keypoints for objects, in a largely self-supervised manner. We use an off-the-shelf detector as input (like MaskRCNN) that is trained only on visible key point annotations. This is the only supervision used in this work. A graph encoder network then explicitly classifies invisible edges and a graph decoder network corrects the occluded keypoint locations from the initial detector. Central to this work is a trifocal tensor loss that provides indirect self-supervision for occluded keypoint locations that are visible in other views of the object. The 2D keypoints are then passed into a 3D graph network that estimates the 3D shape and camera pose using the self-supervised re-projection loss. At test time, our approach successfully localizes keypoints in a single view under a diverse set of severe occlusion settings. We demonstrate and evaluate our approach on synthetic CAD data as well as a large image set capturing vehicles at many busy city intersections. As an interesting aside, we compare the accuracy of human labels of invisible keypoints against those obtained from geometric trifocal-tensor loss.
我们提出了Occlusion-Net,这是一个以很大程度上自我监督的方式预测物体遮挡关键点的2D和3D位置的框架。我们使用一个现成的检测器作为输入(如MaskRCNN),它只在可见的关键点注释上进行训练。这是这项工作中使用的唯一监督。然后,图形编码器网络明确地对不可见的边缘进行分类,图形解码器网络从初始检测器中纠正被遮挡的关键点位置。这项工作的核心是三焦张量损失,它为在物体的其他视图中可见的被遮挡的关键点位置提供间接的自我监督。然后将2D关键点传递到3D图形网络中,该网络使用自监督重投影损失来估计3D形状和相机姿态。在测试时,我们的方法在不同的严重遮挡设置下成功地定位了单个视图中的关键点。我们在合成CAD数据以及在许多繁忙的城市十字路口捕获车辆的大型图像集上演示并评估了我们的方法。作为一个有趣的问题,我们比较了不可见关键点的人类标签与几何三焦张量损失获得的标签的准确性。
{"title":"Occlusion-Net: 2D/3D Occluded Keypoint Localization Using Graph Networks","authors":"Dinesh Reddy Narapureddy, Minh Vo, S. Narasimhan","doi":"10.1109/CVPR.2019.00750","DOIUrl":"https://doi.org/10.1109/CVPR.2019.00750","url":null,"abstract":"We present Occlusion-Net, a framework to predict 2D and 3D locations of occluded keypoints for objects, in a largely self-supervised manner. We use an off-the-shelf detector as input (like MaskRCNN) that is trained only on visible key point annotations. This is the only supervision used in this work. A graph encoder network then explicitly classifies invisible edges and a graph decoder network corrects the occluded keypoint locations from the initial detector. Central to this work is a trifocal tensor loss that provides indirect self-supervision for occluded keypoint locations that are visible in other views of the object. The 2D keypoints are then passed into a 3D graph network that estimates the 3D shape and camera pose using the self-supervised re-projection loss. At test time, our approach successfully localizes keypoints in a single view under a diverse set of severe occlusion settings. We demonstrate and evaluate our approach on synthetic CAD data as well as a large image set capturing vehicles at many busy city intersections. As an interesting aside, we compare the accuracy of human labels of invisible keypoints against those obtained from geometric trifocal-tensor loss.","PeriodicalId":6711,"journal":{"name":"2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","volume":"48 1","pages":"7318-7327"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74325584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 42
Dynamic Recursive Neural Network 动态递归神经网络
Pub Date : 2019-06-01 DOI: 10.1109/CVPR.2019.00529
Qiushan Guo, Zhipeng Yu, Yichao Wu, Ding Liang, Haoyu Qin, Junjie Yan
This paper proposes the dynamic recursive neural network (DRNN), which simplifies the duplicated building blocks in deep neural network. Different from forwarding through different blocks sequentially in previous networks, we demonstrate that the DRNN can achieve better performance with fewer blocks by employing block recursively. We further add a gate structure to each block, which can adaptively decide the loop times of recursive blocks to reduce the computational cost. Since the recursive networks are hard to train, we propose the Loopy Variable Batch Normalization (LVBN) to stabilize the volatile gradient. Further, we improve the LVBN to correct statistical bias caused by the gate structure. Experiments show that the DRNN reduces the parameters and computational cost and while outperforms the original model in term of the accuracy consistently on CIFAR-10 and ImageNet-1k. Lastly we visualize and discuss the relation between image saliency and the number of loop time.
本文提出了动态递归神经网络(DRNN),简化了深度神经网络中重复构建块的问题。与以往网络中不同的块顺序转发不同,我们证明了DRNN通过递归地使用块可以在更少的块上获得更好的性能。我们进一步在每个块中加入一个门结构,它可以自适应地决定递归块的循环次数,从而降低计算成本。由于递归网络难以训练,我们提出了环路变量批归一化(LVBN)来稳定波动梯度。进一步,我们改进了LVBN,以纠正由栅极结构引起的统计偏差。实验表明,在CIFAR-10和ImageNet-1k上,DRNN减少了参数和计算成本,同时在准确率方面优于原模型。最后,我们可视化并讨论了图像显著性与循环时间的关系。
{"title":"Dynamic Recursive Neural Network","authors":"Qiushan Guo, Zhipeng Yu, Yichao Wu, Ding Liang, Haoyu Qin, Junjie Yan","doi":"10.1109/CVPR.2019.00529","DOIUrl":"https://doi.org/10.1109/CVPR.2019.00529","url":null,"abstract":"This paper proposes the dynamic recursive neural network (DRNN), which simplifies the duplicated building blocks in deep neural network. Different from forwarding through different blocks sequentially in previous networks, we demonstrate that the DRNN can achieve better performance with fewer blocks by employing block recursively. We further add a gate structure to each block, which can adaptively decide the loop times of recursive blocks to reduce the computational cost. Since the recursive networks are hard to train, we propose the Loopy Variable Batch Normalization (LVBN) to stabilize the volatile gradient. Further, we improve the LVBN to correct statistical bias caused by the gate structure. Experiments show that the DRNN reduces the parameters and computational cost and while outperforms the original model in term of the accuracy consistently on CIFAR-10 and ImageNet-1k. Lastly we visualize and discuss the relation between image saliency and the number of loop time.","PeriodicalId":6711,"journal":{"name":"2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","volume":"1 1","pages":"5142-5151"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74555291","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
Learning Linear Transformations for Fast Image and Video Style Transfer 学习线性变换快速图像和视频风格转移
Pub Date : 2019-06-01 DOI: 10.1109/CVPR.2019.00393
Xueting Li, Sifei Liu, J. Kautz, Ming-Hsuan Yang
Given a random pair of images, a universal style transfer method extracts the feel from a reference image to synthesize an output based on the look of a content image. Recent algorithms based on second-order statistics, however, are either computationally expensive or prone to generate artifacts due to the trade-off between image quality and runtime performance. In this work, we present an approach for universal style transfer that learns the transformation matrix in a data-driven fashion. Our algorithm is efficient yet flexible to transfer different levels of styles with the same auto-encoder network. It also produces stable video style transfer results due to the preservation of the content affinity. In addition, we propose a linear propagation module to enable a feed-forward network for photo-realistic style transfer. We demonstrate the effectiveness of our approach on three tasks: artistic style, photo-realistic and video style transfer, with comparisons to state-of-the-art methods.
给定随机的一对图像,通用风格转换方法从参考图像中提取感觉,以基于内容图像的外观合成输出。然而,基于二阶统计量的最新算法要么计算成本高,要么由于图像质量和运行时性能之间的权衡而容易产生伪影。在这项工作中,我们提出了一种通用风格迁移方法,该方法以数据驱动的方式学习变换矩阵。我们的算法在使用相同的自编码器网络传输不同层次的风格时,效率高且灵活。由于保留了内容亲和力,也产生了稳定的视频风格转移效果。此外,我们提出了一个线性传播模块,以实现前馈网络的真实感风格传递。我们展示了我们的方法在三个任务上的有效性:艺术风格、照片真实感和视频风格转移,并与最先进的方法进行了比较。
{"title":"Learning Linear Transformations for Fast Image and Video Style Transfer","authors":"Xueting Li, Sifei Liu, J. Kautz, Ming-Hsuan Yang","doi":"10.1109/CVPR.2019.00393","DOIUrl":"https://doi.org/10.1109/CVPR.2019.00393","url":null,"abstract":"Given a random pair of images, a universal style transfer method extracts the feel from a reference image to synthesize an output based on the look of a content image. Recent algorithms based on second-order statistics, however, are either computationally expensive or prone to generate artifacts due to the trade-off between image quality and runtime performance. In this work, we present an approach for universal style transfer that learns the transformation matrix in a data-driven fashion. Our algorithm is efficient yet flexible to transfer different levels of styles with the same auto-encoder network. It also produces stable video style transfer results due to the preservation of the content affinity. In addition, we propose a linear propagation module to enable a feed-forward network for photo-realistic style transfer. We demonstrate the effectiveness of our approach on three tasks: artistic style, photo-realistic and video style transfer, with comparisons to state-of-the-art methods.","PeriodicalId":6711,"journal":{"name":"2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","volume":"40 1","pages":"3804-3812"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76156775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 178
Learning Personalized Modular Network Guided by Structured Knowledge 结构化知识引导下的个性化模块化网络学习
Pub Date : 2019-06-01 DOI: 10.1109/CVPR.2019.00915
Xiaodan Liang
The dominant deep learning approaches use a "one-size-fits-all" paradigm with the hope that underlying characteristics of diverse inputs can be captured via a fixed structure. They also overlook the importance of explicitly modeling feature hierarchy. However, complex real-world tasks often require discovering diverse reasoning paths for different inputs to achieve satisfying predictions, especially for challenging large-scale recognition tasks with complex label relations. In this paper, we treat the structured commonsense knowledge (e.g. concept hierarchy) as the guidance of customizing more powerful and explainable network structures for distinct inputs, leading to dynamic and individualized inference paths. Give an off-the-shelf large network configuration, the proposed Personalized Modular Network (PMN) is learned by selectively activating a sequence of network modules where each of them is designated to recognize particular levels of structured knowledge. Learning semantic configurations and activation of modules to align well with structured knowledge can be regarded as a decision-making procedure, which is solved by a new graph-based reinforcement learning algorithm. Experiments on three semantic segmentation tasks and classification tasks show our PMN can achieve superior performance with the reduced number of network modules while discovering personalized and explainable module configurations for each input.
主流的深度学习方法使用“一刀切”的范式,希望通过固定的结构捕获不同输入的潜在特征。他们还忽略了显式建模特征层次结构的重要性。然而,复杂的现实世界任务通常需要为不同的输入发现不同的推理路径来实现令人满意的预测,特别是对于具有复杂标签关系的挑战性大规模识别任务。在本文中,我们将结构化的常识性知识(例如概念层次)作为指导,为不同的输入定制更强大和可解释的网络结构,从而导致动态和个性化的推理路径。给出一个现成的大型网络配置,所提出的个性化模块化网络(PMN)是通过选择性地激活一系列网络模块来学习的,其中每个模块都被指定为识别特定级别的结构化知识。学习语义配置和激活模块以使其与结构化知识很好地对齐可以看作是一个决策过程,该过程由一种新的基于图的强化学习算法来解决。在三个语义分割任务和分类任务上的实验表明,PMN可以在减少网络模块数量的同时,为每个输入发现个性化和可解释的模块配置,从而获得更好的性能。
{"title":"Learning Personalized Modular Network Guided by Structured Knowledge","authors":"Xiaodan Liang","doi":"10.1109/CVPR.2019.00915","DOIUrl":"https://doi.org/10.1109/CVPR.2019.00915","url":null,"abstract":"The dominant deep learning approaches use a \"one-size-fits-all\" paradigm with the hope that underlying characteristics of diverse inputs can be captured via a fixed structure. They also overlook the importance of explicitly modeling feature hierarchy. However, complex real-world tasks often require discovering diverse reasoning paths for different inputs to achieve satisfying predictions, especially for challenging large-scale recognition tasks with complex label relations. In this paper, we treat the structured commonsense knowledge (e.g. concept hierarchy) as the guidance of customizing more powerful and explainable network structures for distinct inputs, leading to dynamic and individualized inference paths. Give an off-the-shelf large network configuration, the proposed Personalized Modular Network (PMN) is learned by selectively activating a sequence of network modules where each of them is designated to recognize particular levels of structured knowledge. Learning semantic configurations and activation of modules to align well with structured knowledge can be regarded as a decision-making procedure, which is solved by a new graph-based reinforcement learning algorithm. Experiments on three semantic segmentation tasks and classification tasks show our PMN can achieve superior performance with the reduced number of network modules while discovering personalized and explainable module configurations for each input.","PeriodicalId":6711,"journal":{"name":"2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","volume":"131 1","pages":"8936-8944"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76377204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Sensitive-Sample Fingerprinting of Deep Neural Networks 深度神经网络的敏感样本指纹识别
Pub Date : 2019-06-01 DOI: 10.1109/CVPR.2019.00486
Zecheng He, Tianwei Zhang, R. Lee
Numerous cloud-based services are provided to help customers develop and deploy deep learning applications. When a customer deploys a deep learning model in the cloud and serves it to end-users, it is important to be able to verify that the deployed model has not been tampered with. In this paper, we propose a novel and practical methodology to verify the integrity of remote deep learning models, with only black-box access to the target models. Specifically, we define Sensitive-Sample fingerprints, which are a small set of human unnoticeable transformed inputs that make the model outputs sensitive to the model's parameters. Even small model changes can be clearly reflected in the model outputs. Experimental results on different types of model integrity attacks show that we proposed approach is both effective and efficient. It can detect model integrity breaches with high accuracy (>99.95%) and guaranteed zero false positives on all evaluated attacks. Meanwhile, it only requires up to 103X fewer model inferences, compared with non-sensitive samples.
提供了许多基于云的服务来帮助客户开发和部署深度学习应用程序。当客户在云中部署深度学习模型并将其提供给最终用户时,能够验证部署的模型没有被篡改是很重要的。在本文中,我们提出了一种新颖实用的方法来验证远程深度学习模型的完整性,只需黑盒访问目标模型。具体来说,我们定义了敏感样本指纹,它是一小组人类不明显的转换输入,使模型输出对模型的参数敏感。即使很小的模型变化也能在模型输出中清楚地反映出来。针对不同类型的模型完整性攻击的实验结果表明,该方法是有效的。它可以以高准确率(>99.95%)检测模型完整性破坏,并保证对所有评估的攻击为零误报。同时,与非敏感样本相比,它只需要少103X的模型推断。
{"title":"Sensitive-Sample Fingerprinting of Deep Neural Networks","authors":"Zecheng He, Tianwei Zhang, R. Lee","doi":"10.1109/CVPR.2019.00486","DOIUrl":"https://doi.org/10.1109/CVPR.2019.00486","url":null,"abstract":"Numerous cloud-based services are provided to help customers develop and deploy deep learning applications. When a customer deploys a deep learning model in the cloud and serves it to end-users, it is important to be able to verify that the deployed model has not been tampered with. In this paper, we propose a novel and practical methodology to verify the integrity of remote deep learning models, with only black-box access to the target models. Specifically, we define Sensitive-Sample fingerprints, which are a small set of human unnoticeable transformed inputs that make the model outputs sensitive to the model's parameters. Even small model changes can be clearly reflected in the model outputs. Experimental results on different types of model integrity attacks show that we proposed approach is both effective and efficient. It can detect model integrity breaches with high accuracy (>99.95%) and guaranteed zero false positives on all evaluated attacks. Meanwhile, it only requires up to 103X fewer model inferences, compared with non-sensitive samples.","PeriodicalId":6711,"journal":{"name":"2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","volume":"51 17 1","pages":"4724-4732"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76117014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
CrDoCo: Pixel-Level Domain Transfer With Cross-Domain Consistency CrDoCo:具有跨域一致性的像素级域传输
Pub Date : 2019-06-01 DOI: 10.1109/CVPR.2019.00189
Yun-Chun Chen, Yen-Yu Lin, Ming-Hsuan Yang, Jia-Bin Huang
Unsupervised domain adaptation algorithms aim to transfer the knowledge learned from one domain to another (e.g., synthetic to real images). The adapted representations often do not capture pixel-level domain shifts that are crucial for dense prediction tasks (e.g., semantic segmentation). In this paper, we present a novel pixel-wise adversarial domain adaptation algorithm. By leveraging image-to-image translation methods for data augmentation, our key insight is that while the translated images between domains may differ in styles, their predictions for the task should be consistent. We exploit this property and introduce a cross-domain consistency loss that enforces our adapted model to produce consistent predictions. Through extensive experimental results, we show that our method compares favorably against the state-of-the-art on a wide variety of unsupervised domain adaptation tasks.
无监督领域自适应算法旨在将从一个领域学习到的知识转移到另一个领域(例如,将合成图像转移到真实图像)。适应的表示通常不能捕获像素级的域转移,这对于密集的预测任务(例如,语义分割)至关重要。本文提出了一种新的逐像素对抗域自适应算法。通过利用图像到图像的翻译方法进行数据增强,我们的关键见解是,虽然域之间的翻译图像在风格上可能不同,但它们对任务的预测应该是一致的。我们利用这一特性并引入跨域一致性损失,以强制我们的适应模型产生一致的预测。通过广泛的实验结果,我们表明我们的方法在各种无监督域自适应任务上优于最先进的方法。
{"title":"CrDoCo: Pixel-Level Domain Transfer With Cross-Domain Consistency","authors":"Yun-Chun Chen, Yen-Yu Lin, Ming-Hsuan Yang, Jia-Bin Huang","doi":"10.1109/CVPR.2019.00189","DOIUrl":"https://doi.org/10.1109/CVPR.2019.00189","url":null,"abstract":"Unsupervised domain adaptation algorithms aim to transfer the knowledge learned from one domain to another (e.g., synthetic to real images). The adapted representations often do not capture pixel-level domain shifts that are crucial for dense prediction tasks (e.g., semantic segmentation). In this paper, we present a novel pixel-wise adversarial domain adaptation algorithm. By leveraging image-to-image translation methods for data augmentation, our key insight is that while the translated images between domains may differ in styles, their predictions for the task should be consistent. We exploit this property and introduce a cross-domain consistency loss that enforces our adapted model to produce consistent predictions. Through extensive experimental results, we show that our method compares favorably against the state-of-the-art on a wide variety of unsupervised domain adaptation tasks.","PeriodicalId":6711,"journal":{"name":"2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","volume":"29 1","pages":"1791-1800"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81995819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 244
Adaptive Transfer Network for Cross-Domain Person Re-Identification 跨域人员再识别的自适应转移网络
Pub Date : 2019-06-01 DOI: 10.1109/CVPR.2019.00737
Jiawei Liu, Zhengjun Zha, Di Chen, Richang Hong, Meng Wang
Recent deep learning based person re-identification approaches have steadily improved the performance for benchmarks, however they often fail to generalize well from one domain to another. In this work, we propose a novel adaptive transfer network (ATNet) for effective cross-domain person re-identification. ATNet looks into the essential causes of domain gap and addresses it following the principle of "divide-and-conquer". It decomposes the complicated cross-domain transfer into a set of factor-wise sub-transfers, each of which concentrates on style transfer with respect to a certain imaging factor, e.g., illumination, resolution and camera view etc. An adaptive ensemble strategy is proposed to fuse factor-wise transfers by perceiving the affect magnitudes of various factors on images. Such "decomposition-and-ensemble" strategy gives ATNet the capability of precise style transfer at factor level and eventually effective transfer across domains. In particular, ATNet consists of a transfer network composed by multiple factor-wise CycleGANs and an ensemble CycleGAN as well as a selection network that infers the affects of different factors on transferring each image. Extensive experimental results on three widely-used datasets, i.e., Market-1501, DukeMTMC-reID and PRID2011 have demonstrated the effectiveness of the proposed ATNet with significant performance improvements over state-of-the-art methods.
最近基于深度学习的人员再识别方法已经稳步提高了基准测试的性能,但是它们往往不能很好地从一个领域推广到另一个领域。在这项工作中,我们提出了一种新的自适应迁移网络(ATNet),用于有效的跨域人员再识别。ATNet研究了产生域名鸿沟的根本原因,并遵循“分而治之”的原则加以解决。它将复杂的跨域迁移分解为一组基于因子的子迁移,每个子迁移集中于相对于特定成像因子的风格迁移,例如照明,分辨率和相机视图等。提出了一种自适应集成策略,通过感知各种因素对图像的影响程度来融合因子迁移。这种“分解-集成”策略使ATNet能够在要素水平上进行精确的风格迁移,并最终实现有效的跨域迁移。其中,ATNet包括由多因子CycleGAN和集成CycleGAN组成的传输网络,以及推断不同因素对传输每张图像的影响的选择网络。在三个广泛使用的数据集(即Market-1501, DukeMTMC-reID和PRID2011)上的大量实验结果证明了所提出的ATNet的有效性,其性能优于最先进的方法。
{"title":"Adaptive Transfer Network for Cross-Domain Person Re-Identification","authors":"Jiawei Liu, Zhengjun Zha, Di Chen, Richang Hong, Meng Wang","doi":"10.1109/CVPR.2019.00737","DOIUrl":"https://doi.org/10.1109/CVPR.2019.00737","url":null,"abstract":"Recent deep learning based person re-identification approaches have steadily improved the performance for benchmarks, however they often fail to generalize well from one domain to another. In this work, we propose a novel adaptive transfer network (ATNet) for effective cross-domain person re-identification. ATNet looks into the essential causes of domain gap and addresses it following the principle of \"divide-and-conquer\". It decomposes the complicated cross-domain transfer into a set of factor-wise sub-transfers, each of which concentrates on style transfer with respect to a certain imaging factor, e.g., illumination, resolution and camera view etc. An adaptive ensemble strategy is proposed to fuse factor-wise transfers by perceiving the affect magnitudes of various factors on images. Such \"decomposition-and-ensemble\" strategy gives ATNet the capability of precise style transfer at factor level and eventually effective transfer across domains. In particular, ATNet consists of a transfer network composed by multiple factor-wise CycleGANs and an ensemble CycleGAN as well as a selection network that infers the affects of different factors on transferring each image. Extensive experimental results on three widely-used datasets, i.e., Market-1501, DukeMTMC-reID and PRID2011 have demonstrated the effectiveness of the proposed ATNet with significant performance improvements over state-of-the-art methods.","PeriodicalId":6711,"journal":{"name":"2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","volume":"57 1","pages":"7195-7204"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82802948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 217
Dichromatic Model Based Temporal Color Constancy for AC Light Sources 基于二色模型的交流光源时间色常数
Pub Date : 2019-06-01 DOI: 10.1109/CVPR.2019.01261
Jun-Sang Yoo, Jong-Ok Kim
Existing dichromatic color constancy approach commonly requires a number of spatial pixels which have high specularity. In this paper, we propose a novel approach to estimate the illuminant chromaticity of AC light source using high-speed camera. We found that the temporal observations of an image pixel at a fixed location distribute on an identical dichromatic plane. Instead of spatial pixels with high specularity, multiple temporal samples of a pixel are exploited to determine AC pixels for dichromatic plane estimation, whose pixel intensity is sinusoidally varying well. A dichromatic plane is calculated per each AC pixel, and illuminant chromaticity is determined by the intersection of dichromatic planes. From multiple dichromatic planes, an optimal illuminant is estimated with a novel MAP framework. It is shown that the proposed method outperforms both existing dichromatic based methods and temporal color constancy methods, irrespective of the amount of specularity.
现有的二色恒常性方法通常需要大量具有高反射性的空间像素。本文提出了一种利用高速摄像机测量交流光源色度的新方法。我们发现在固定位置的图像像素的时间观测分布在相同的二色平面上。利用一个像元的多个时间样本来确定二色平面估计的交流像元,而不是具有高反射率的空间像元,其像元强度具有良好的正弦变化。每个AC像素计算一个二色平面,并且光源色度由二色平面的交集确定。从多个二色平面出发,利用一种新的MAP框架估计出最优光源。结果表明,该方法与现有的基于二色的方法和时间颜色常数的方法相比,无论镜面的多少,都优于现有的方法。
{"title":"Dichromatic Model Based Temporal Color Constancy for AC Light Sources","authors":"Jun-Sang Yoo, Jong-Ok Kim","doi":"10.1109/CVPR.2019.01261","DOIUrl":"https://doi.org/10.1109/CVPR.2019.01261","url":null,"abstract":"Existing dichromatic color constancy approach commonly requires a number of spatial pixels which have high specularity. In this paper, we propose a novel approach to estimate the illuminant chromaticity of AC light source using high-speed camera. We found that the temporal observations of an image pixel at a fixed location distribute on an identical dichromatic plane. Instead of spatial pixels with high specularity, multiple temporal samples of a pixel are exploited to determine AC pixels for dichromatic plane estimation, whose pixel intensity is sinusoidally varying well. A dichromatic plane is calculated per each AC pixel, and illuminant chromaticity is determined by the intersection of dichromatic planes. From multiple dichromatic planes, an optimal illuminant is estimated with a novel MAP framework. It is shown that the proposed method outperforms both existing dichromatic based methods and temporal color constancy methods, irrespective of the amount of specularity.","PeriodicalId":6711,"journal":{"name":"2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","volume":"55 1","pages":"12321-12330"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80195576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
期刊
2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1