首页 > 最新文献

2017 IEEE International Conference on Computer Vision Workshops (ICCVW)最新文献

英文 中文
Computer Vision for the Visually Impaired: the Sound of Vision System 视障人士的计算机视觉:视觉之声系统
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.175
S. Caraiman, A. Morar, Mateusz Owczarek, A. Burlacu, D. Rzeszotarski, N. Botezatu, P. Herghelegiu, F. Moldoveanu, P. Strumiłło, A. Moldoveanu
This paper presents a computer vision based sensory substitution device for the visually impaired. Its main objective is to provide the users with a 3D representation of the environment around them, conveyed by means of the hearing and tactile senses. One of the biggest challenges for this system is to ensure pervasiveness, i.e., to be usable in any indoor or outdoor environments and in any illumination conditions. This work reveals both the hardware (3D acquisition system) and software (3D processing pipeline) used for developing this sensory substitution device and provides insight on its exploitation in various scenarios. Preliminary experiments with blind users revealed good usability results and provided valuable feedback for system improvement.
提出了一种基于计算机视觉的视障感觉替代装置。它的主要目标是为用户提供周围环境的3D表示,通过听觉和触觉来传达。该系统面临的最大挑战之一是确保普及性,即在任何室内或室外环境以及任何照明条件下都可以使用。这项工作揭示了用于开发这种感官替代装置的硬件(3D采集系统)和软件(3D处理管道),并提供了对其在各种场景中的开发的见解。盲人用户的初步实验显示了良好的可用性结果,并为系统改进提供了有价值的反馈。
{"title":"Computer Vision for the Visually Impaired: the Sound of Vision System","authors":"S. Caraiman, A. Morar, Mateusz Owczarek, A. Burlacu, D. Rzeszotarski, N. Botezatu, P. Herghelegiu, F. Moldoveanu, P. Strumiłło, A. Moldoveanu","doi":"10.1109/ICCVW.2017.175","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.175","url":null,"abstract":"This paper presents a computer vision based sensory substitution device for the visually impaired. Its main objective is to provide the users with a 3D representation of the environment around them, conveyed by means of the hearing and tactile senses. One of the biggest challenges for this system is to ensure pervasiveness, i.e., to be usable in any indoor or outdoor environments and in any illumination conditions. This work reveals both the hardware (3D acquisition system) and software (3D processing pipeline) used for developing this sensory substitution device and provides insight on its exploitation in various scenarios. Preliminary experiments with blind users revealed good usability results and provided valuable feedback for system improvement.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133349540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 70
Dynamic Mode Decomposition for Background Modeling 动态模式分解背景建模
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.220
Seth D. Pendergrass, S. Brunton, J. Kutz, N. Benjamin Erichson, T. Askham
The Dynamic Mode Decomposition (DMD) is a spatiotemporal matrix decomposition method capable of background modeling in video streams. DMD is a regression technique that integrates Fourier transforms and singular value decomposition. Innovations in compressed sensing allow for a scalable and rapid decomposition of video streams that scales with the intrinsic rank of the matrix, rather than the size of the actual video. Our results show that the quality of the resulting background model is competitive, quantified by the F-measure, recall and precision. A GPU (graphics processing unit) accelerated implementation is also possible allowing the algorithm to operate efficiently on streaming data. In addition, it is possible to leverage the native compressed format of many data streams, such as HD video and computational physics codes that are represented sparsely in the Fourier domain, to massively reduce data transfer from CPU to GPU and to enable sparse matrix multiplications.
动态模态分解(DMD)是一种能够对视频流进行背景建模的时空矩阵分解方法。DMD是一种集傅里叶变换和奇异值分解于一体的回归技术。压缩感知的创新允许视频流的可扩展和快速分解,该分解随矩阵的固有秩而不是实际视频的大小而缩放。我们的研究结果表明,所得到的背景模型的质量是有竞争力的,通过f度量、召回率和精度来量化。GPU(图形处理单元)加速实现也可能允许算法有效地处理流数据。此外,可以利用许多数据流的本机压缩格式,例如高清视频和在傅里叶域中稀疏表示的计算物理代码,以大规模减少从CPU到GPU的数据传输,并启用稀疏矩阵乘法。
{"title":"Dynamic Mode Decomposition for Background Modeling","authors":"Seth D. Pendergrass, S. Brunton, J. Kutz, N. Benjamin Erichson, T. Askham","doi":"10.1109/ICCVW.2017.220","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.220","url":null,"abstract":"The Dynamic Mode Decomposition (DMD) is a spatiotemporal matrix decomposition method capable of background modeling in video streams. DMD is a regression technique that integrates Fourier transforms and singular value decomposition. Innovations in compressed sensing allow for a scalable and rapid decomposition of video streams that scales with the intrinsic rank of the matrix, rather than the size of the actual video. Our results show that the quality of the resulting background model is competitive, quantified by the F-measure, recall and precision. A GPU (graphics processing unit) accelerated implementation is also possible allowing the algorithm to operate efficiently on streaming data. In addition, it is possible to leverage the native compressed format of many data streams, such as HD video and computational physics codes that are represented sparsely in the Fourier domain, to massively reduce data transfer from CPU to GPU and to enable sparse matrix multiplications.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133647232","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
PVNN: A Neural Network Library for Photometric Vision PVNN:用于光度视觉的神经网络库
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.69
Ye Yu, W. Smith
In this paper we show how a differentiable, physics-based renderer suitable for photometric vision tasks can be implemented as layers in a deep neural network. The layers include geometric operations for representation transformations, reflectance evaluations with arbitrary numbers of light sources and statistical bidirectional reflectance distribution function (BRDF) models. We make an implementation of these layers available as a neural network library (PVNN) for Theano. The layers can be incorporated into any neural network architecture, allowing parts of the photometric image formation process to be explicitly modelled in a network that is trained end to end via backpropagation. As an exemplar application, we show how to train a network with encoder-decoder architecture that learns to estimate BRDF parameters from a single image in an unsupervised manner.
在本文中,我们展示了一个适合光度视觉任务的可微分的、基于物理的渲染器如何在深度神经网络中作为层实现。这些层包括用于表示转换的几何操作、任意数量光源的反射率评估和统计双向反射率分布函数(BRDF)模型。我们将这些层的实现作为Theano的神经网络库(PVNN)。这些层可以被整合到任何神经网络架构中,允许部分光度图像形成过程在通过反向传播端到端训练的网络中明确建模。作为示例应用,我们展示了如何训练具有编码器-解码器架构的网络,该网络学习以无监督的方式从单个图像中估计BRDF参数。
{"title":"PVNN: A Neural Network Library for Photometric Vision","authors":"Ye Yu, W. Smith","doi":"10.1109/ICCVW.2017.69","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.69","url":null,"abstract":"In this paper we show how a differentiable, physics-based renderer suitable for photometric vision tasks can be implemented as layers in a deep neural network. The layers include geometric operations for representation transformations, reflectance evaluations with arbitrary numbers of light sources and statistical bidirectional reflectance distribution function (BRDF) models. We make an implementation of these layers available as a neural network library (PVNN) for Theano. The layers can be incorporated into any neural network architecture, allowing parts of the photometric image formation process to be explicitly modelled in a network that is trained end to end via backpropagation. As an exemplar application, we show how to train a network with encoder-decoder architecture that learns to estimate BRDF parameters from a single image in an unsupervised manner.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133109687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Shape-from-Polarisation: A Nonlinear Least Squares Approach 偏振形状:非线性最小二乘方法
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.350
Ye Yu, Dizhong Zhu, W. Smith
In this paper we present a new type of approach for estimating surface height from polarimetric data, i.e. a sequence of images in which a linear polarising filter is rotated in front of a camera. In contrast to all previous shape-from-polarisation methods, we do not first transform the observed data into a polarisation image. Instead, we minimise the sum of squared residuals between predicted and observed intensities over all pixels and polariser angles. This is a nonlinear least squares optimisation problem in which the unknown is the surface height. The forward prediction is a series of transformations for which we provide analytical derivatives allowing the overall problem to be efficiently optimised using Gauss-Newton type methods with an analytical Jacobian matrix. The method is very general and can incorporate any (differentiable) illumination, reflectance or polarisation model. We also propose a variant of the method which uses image ratios to remove dependence on illumination and albedo. We demonstrate our methods on glossy objects, including with albedo variations, and provide comparison to a state of the art approach.
在本文中,我们提出了一种从偏振数据估计表面高度的新方法,即在相机前旋转线性偏振滤光片的图像序列。与所有以前的偏振形状方法相反,我们不首先将观测数据转换为偏振图像。相反,我们最小化所有像素和偏振镜角度上预测和观测强度之间的平方残差之和。这是一个非线性最小二乘优化问题,其中未知的是表面高度。前向预测是一系列变换,我们为其提供解析导数,允许使用具有解析雅可比矩阵的高斯-牛顿型方法有效地优化整个问题。该方法非常通用,可以包含任何(可微分)照明、反射率或偏振模型。我们还提出了一种使用图像比率来消除对照度和反照率依赖的方法。我们展示了我们在光滑物体上的方法,包括反照率变化,并提供了与最先进方法的比较。
{"title":"Shape-from-Polarisation: A Nonlinear Least Squares Approach","authors":"Ye Yu, Dizhong Zhu, W. Smith","doi":"10.1109/ICCVW.2017.350","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.350","url":null,"abstract":"In this paper we present a new type of approach for estimating surface height from polarimetric data, i.e. a sequence of images in which a linear polarising filter is rotated in front of a camera. In contrast to all previous shape-from-polarisation methods, we do not first transform the observed data into a polarisation image. Instead, we minimise the sum of squared residuals between predicted and observed intensities over all pixels and polariser angles. This is a nonlinear least squares optimisation problem in which the unknown is the surface height. The forward prediction is a series of transformations for which we provide analytical derivatives allowing the overall problem to be efficiently optimised using Gauss-Newton type methods with an analytical Jacobian matrix. The method is very general and can incorporate any (differentiable) illumination, reflectance or polarisation model. We also propose a variant of the method which uses image ratios to remove dependence on illumination and albedo. We demonstrate our methods on glossy objects, including with albedo variations, and provide comparison to a state of the art approach.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"159 5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123080184","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Discrepancy-Based Networks for Unsupervised Domain Adaptation: A Comparative Study 基于差异的无监督域自适应网络的比较研究
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.312
G. Csurka, Fabien Baradel, Boris Chidlovskii, S. Clinchant
Domain Adaptation (DA) exploits labeled data and models from similar domains in order to alleviate the annotation burden when learning a model in a new domain. Our contribution to the field is three-fold. First, we propose a new dataset, LandMarkDA, to study the adaptation between landmark place recognition models trained with different artistic image styles, such as photos, paintings and drawings. The new LandMarkDA proposes new adaptation challenges, where current deep architectures show their limits. Second, we propose an experimental study of recent shallow and deep adaptation networks, based on using Maximum Mean Discrepancy to bridge the domain gap. We study different design choices for these models by varying the network architectures and evaluate them on OFF31 and the new LandMarkDA collections. We show that shallow networks can still be competitive under an appropriate feature extraction. Finally, we also benchmark a new DA method that successfully combines the artistic image style-transfer with deep discrepancy-based networks.
领域适应(Domain Adaptation, DA)利用来自相似领域的标记数据和模型,以减轻在新领域学习模型时的标注负担。我们对这个领域的贡献是三重的。首先,我们提出了一个新的数据集LandMarkDA,研究了不同艺术图像风格(如照片、绘画和素描)训练的地标性地点识别模型之间的自适应。新的LandMarkDA提出了新的适应挑战,当前的深度架构显示出其局限性。其次,我们提出了一种基于最大均值差异来弥补域差距的浅层和深层自适应网络的实验研究。我们通过改变网络架构来研究这些模型的不同设计选择,并在OFF31和新的LandMarkDA集合上对它们进行评估。我们表明,在适当的特征提取下,浅网络仍然可以具有竞争力。最后,我们还测试了一种新的数据处理方法,该方法成功地将艺术图像风格转移与基于深度差异的网络相结合。
{"title":"Discrepancy-Based Networks for Unsupervised Domain Adaptation: A Comparative Study","authors":"G. Csurka, Fabien Baradel, Boris Chidlovskii, S. Clinchant","doi":"10.1109/ICCVW.2017.312","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.312","url":null,"abstract":"Domain Adaptation (DA) exploits labeled data and models from similar domains in order to alleviate the annotation burden when learning a model in a new domain. Our contribution to the field is three-fold. First, we propose a new dataset, LandMarkDA, to study the adaptation between landmark place recognition models trained with different artistic image styles, such as photos, paintings and drawings. The new LandMarkDA proposes new adaptation challenges, where current deep architectures show their limits. Second, we propose an experimental study of recent shallow and deep adaptation networks, based on using Maximum Mean Discrepancy to bridge the domain gap. We study different design choices for these models by varying the network architectures and evaluate them on OFF31 and the new LandMarkDA collections. We show that shallow networks can still be competitive under an appropriate feature extraction. Finally, we also benchmark a new DA method that successfully combines the artistic image style-transfer with deep discrepancy-based networks.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"124 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123312387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Towards a Spatio-Temporal Atlas of 3D Cellular Parameters During Leaf Morphogenesis 叶片形态发生过程中三维细胞参数的时空图谱
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.14
F. Selka, T. Blein, J. Burguet, E. Biot, P. Laufs, P. Andrey
Morphogenesis is a complex process that integrates several mechanisms from the molecular to the organ scales. In plants, division and growth are the two fundamental cellular mechanisms that drive morphogenesis. However, little is known about how these mechanisms are coordinated to establish functional tissue structure. A fundamental bottleneck is the current lack of techniques to systematically quantify the spatio-temporal evolution of 3D cell morphology during organ growth. Using leaf development as a relevant and challenging model to study morphogenesis, we developed a computational framework for cell analysis and quantification from 3D images and for the generation of 3D cell shape atlas. A remarkable feature of leaf morphogenesis being the formation of a laminar-like structure, we propose to automatically separate the cells corresponding to the leaf sides in the segmented leaves, by applying a clustering algorithm. The performance of the proposed pipeline was experimentally assessed on a dataset of 46 leaves in an early developmental state.
形态发生是一个复杂的过程,整合了从分子到器官尺度的多种机制。在植物中,分裂和生长是驱动形态发生的两个基本细胞机制。然而,人们对这些机制如何协调建立功能性组织结构知之甚少。一个基本的瓶颈是目前缺乏技术来系统地量化三维细胞形态在器官生长过程中的时空演变。利用叶片发育作为一个相关且具有挑战性的模型来研究形态发生,我们开发了一个计算框架,用于从3D图像中进行细胞分析和量化,并用于生成3D细胞形状图谱。由于叶片形态发生的显著特征是层状结构的形成,我们提出了采用聚类算法自动分离叶片中对应叶侧的细胞。在46个处于早期发育状态的叶片数据集上,实验评估了所提出的管道的性能。
{"title":"Towards a Spatio-Temporal Atlas of 3D Cellular Parameters During Leaf Morphogenesis","authors":"F. Selka, T. Blein, J. Burguet, E. Biot, P. Laufs, P. Andrey","doi":"10.1109/ICCVW.2017.14","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.14","url":null,"abstract":"Morphogenesis is a complex process that integrates several mechanisms from the molecular to the organ scales. In plants, division and growth are the two fundamental cellular mechanisms that drive morphogenesis. However, little is known about how these mechanisms are coordinated to establish functional tissue structure. A fundamental bottleneck is the current lack of techniques to systematically quantify the spatio-temporal evolution of 3D cell morphology during organ growth. Using leaf development as a relevant and challenging model to study morphogenesis, we developed a computational framework for cell analysis and quantification from 3D images and for the generation of 3D cell shape atlas. A remarkable feature of leaf morphogenesis being the formation of a laminar-like structure, we propose to automatically separate the cells corresponding to the leaf sides in the segmented leaves, by applying a clustering algorithm. The performance of the proposed pipeline was experimentally assessed on a dataset of 46 leaves in an early developmental state.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"146 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123391354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Reliable Isometric Point Correspondence from Depth 可靠的深度等距点对应
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.152
Emel Küpçü, Y. Yemez
We propose a new iterative isometric point correspondence method that relies on diffusion distance to handle challenges posed by commodity depth sensors, which usually provide incomplete and noisy surface data exhibiting holes and gaps. We formulate the correspondence problem as finding an optimal partial mapping between two given point sets, that minimizes deviation from isometry. Our algorithm starts with an initial rough correspondence between keypoints, obtained via a standard descriptor matching technique. This initial correspondence is then pruned and updated by iterating a perfect matching algorithm until convergence to find as many reliable correspondences as possible. For shapes with intrinsic symmetries such as human models, we additionally provide a symmetry aware extension to improve our formulation. The experiments show that our method provides state of the art performance over depth frames exhibiting occlusions, large deformations and topological noise.
我们提出了一种新的迭代等距点对应方法,该方法依赖于扩散距离来处理商品深度传感器带来的挑战,这些传感器通常提供不完整和有噪声的地表数据,显示孔洞和间隙。我们将对应问题表述为寻找两个给定点集之间的最优部分映射,使与等距的偏差最小化。我们的算法从关键点之间的初始粗略对应开始,通过标准描述符匹配技术获得。然后通过迭代完美匹配算法对初始对应进行修剪和更新,直到收敛以找到尽可能多的可靠对应。对于具有内在对称性的形状,如人体模型,我们还提供了一个对称感知扩展来改进我们的公式。实验表明,我们的方法在具有遮挡、大变形和拓扑噪声的深度帧上提供了最先进的性能。
{"title":"Reliable Isometric Point Correspondence from Depth","authors":"Emel Küpçü, Y. Yemez","doi":"10.1109/ICCVW.2017.152","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.152","url":null,"abstract":"We propose a new iterative isometric point correspondence method that relies on diffusion distance to handle challenges posed by commodity depth sensors, which usually provide incomplete and noisy surface data exhibiting holes and gaps. We formulate the correspondence problem as finding an optimal partial mapping between two given point sets, that minimizes deviation from isometry. Our algorithm starts with an initial rough correspondence between keypoints, obtained via a standard descriptor matching technique. This initial correspondence is then pruned and updated by iterating a perfect matching algorithm until convergence to find as many reliable correspondences as possible. For shapes with intrinsic symmetries such as human models, we additionally provide a symmetry aware extension to improve our formulation. The experiments show that our method provides state of the art performance over depth frames exhibiting occlusions, large deformations and topological noise.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123994489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Learning to Identify While Failing to Discriminate 学会认同而不歧视
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.298
Jure Sokolić, M. Rodrigues, Qiang Qiu, G. Sapiro
Privacy and fairness are critical in computer vision applications, in particular when dealing with human identification. Achieving a universally secure, private, and fair systems is practically impossible as the exploitation of additional data can reveal private information in the original one. Faced with this challenge, we propose a new line of research, where the privacy is learned and used in a closed environment. The goal is to ensure that a given entity, trusted to infer certain information with our data, is blocked from inferring protected information from it. We design a system that learns to succeed on the positive task while simultaneously fail at the negative one, and illustrate this with challenging cases where the positive task (face verification) is harder than the negative one (gender classification). The framework opens the door to privacy and fairness in very important closed scenarios, ranging from private data accumulation companies to law-enforcement and hospitals.
隐私和公平在计算机视觉应用中是至关重要的,特别是在处理人类身份识别时。实现普遍安全、私密、公平的系统实际上是不可能的,因为利用额外的数据可能会泄露原始数据中的私人信息。面对这一挑战,我们提出了一个新的研究方向,在一个封闭的环境中学习和使用隐私。目标是确保一个给定的实体(被信任可以从我们的数据推断出某些信息)被阻止从它推断出受保护的信息。我们设计了一个系统,学习在积极任务上取得成功,同时在消极任务上失败,并通过积极任务(面部识别)比消极任务(性别分类)更难的挑战性案例来说明这一点。该框架在非常重要的封闭场景(从私人数据积累公司到执法部门和医院)中为隐私和公平打开了大门。
{"title":"Learning to Identify While Failing to Discriminate","authors":"Jure Sokolić, M. Rodrigues, Qiang Qiu, G. Sapiro","doi":"10.1109/ICCVW.2017.298","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.298","url":null,"abstract":"Privacy and fairness are critical in computer vision applications, in particular when dealing with human identification. Achieving a universally secure, private, and fair systems is practically impossible as the exploitation of additional data can reveal private information in the original one. Faced with this challenge, we propose a new line of research, where the privacy is learned and used in a closed environment. The goal is to ensure that a given entity, trusted to infer certain information with our data, is blocked from inferring protected information from it. We design a system that learns to succeed on the positive task while simultaneously fail at the negative one, and illustrate this with challenging cases where the positive task (face verification) is harder than the negative one (gender classification). The framework opens the door to privacy and fairness in very important closed scenarios, ranging from private data accumulation companies to law-enforcement and hospitals.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125128276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Siamese Networks for Chromosome Classification 染色体分类的暹罗网络
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.17
Swati, Gaurav Gupta, Mohit Yadav, Monika Sharma, L. Vig
Karyotying is the process of pairing and ordering 23 pairs of human chromosomes from cell images on the basis of size, centromere position, and banding pattern. Karyotyping during metaphase is often used by clinical cytogeneticists to analyze human chromosomes for diagnostic purposes. It requires experience, domain expertise and considerable manual effort to efficiently perform karyotyping and diagnosis of various disorders. Therefore, automation or even partial automation is highly desirable to assist technicians and reduce the cognitive load necessary for karyotyping. With these motivations, in this paper, we attempt to develop methods for chromosome classification by borrowing the latest ideas from deep learning. More specifically, we perform straightening on chromosomes and feed them into Siamese Networks to push the embeddings of samples coming from similar labels closer. Further, we propose to perform balanced sampling from the pairwise dataset while selecting dissimilar training pairs for Siamese Networks, and an MLP based prediction on top of the embeddings obtained from the trained Siamese Networks. We perform our experiments on a real world dataset of healthy patients collected from a hospital and exhaustively compare the effect of different straightening techniques, by applying them to chromosome images prior to classification. Results demonstrate that the proposed methods speed up both training and prediction by 83 and 3 folds, respectively; while surpassing the performance of a very competitive baseline created utilizing deep convolutional neural networks.
染色体核合是根据细胞图像中的大小、着丝粒位置和带型对23对人类染色体进行配对和排序的过程。临床细胞遗传学家经常使用中期核型分析人类染色体的诊断目的。它需要经验,领域的专业知识和相当大的人工努力,有效地执行核型和各种疾病的诊断。因此,自动化或甚至部分自动化是非常可取的,以协助技术人员并减少核型所必需的认知负荷。基于这些动机,在本文中,我们尝试通过借鉴深度学习的最新思想来开发染色体分类方法。更具体地说,我们对染色体进行拉直,并将它们输入到暹罗网络中,以使来自相似标签的样本嵌入得更近。此外,我们建议从成对数据集中进行平衡采样,同时为Siamese网络选择不同的训练对,并在从训练的Siamese网络获得的嵌入之上进行基于MLP的预测。我们在从医院收集的健康患者的真实世界数据集上进行实验,并通过在分类之前将它们应用于染色体图像,详尽地比较了不同矫直技术的效果。结果表明,该方法的训练和预测速度分别提高了83倍和3倍;同时超越了利用深度卷积神经网络创建的非常有竞争力的基线的性能。
{"title":"Siamese Networks for Chromosome Classification","authors":"Swati, Gaurav Gupta, Mohit Yadav, Monika Sharma, L. Vig","doi":"10.1109/ICCVW.2017.17","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.17","url":null,"abstract":"Karyotying is the process of pairing and ordering 23 pairs of human chromosomes from cell images on the basis of size, centromere position, and banding pattern. Karyotyping during metaphase is often used by clinical cytogeneticists to analyze human chromosomes for diagnostic purposes. It requires experience, domain expertise and considerable manual effort to efficiently perform karyotyping and diagnosis of various disorders. Therefore, automation or even partial automation is highly desirable to assist technicians and reduce the cognitive load necessary for karyotyping. With these motivations, in this paper, we attempt to develop methods for chromosome classification by borrowing the latest ideas from deep learning. More specifically, we perform straightening on chromosomes and feed them into Siamese Networks to push the embeddings of samples coming from similar labels closer. Further, we propose to perform balanced sampling from the pairwise dataset while selecting dissimilar training pairs for Siamese Networks, and an MLP based prediction on top of the embeddings obtained from the trained Siamese Networks. We perform our experiments on a real world dataset of healthy patients collected from a hospital and exhaustively compare the effect of different straightening techniques, by applying them to chromosome images prior to classification. Results demonstrate that the proposed methods speed up both training and prediction by 83 and 3 folds, respectively; while surpassing the performance of a very competitive baseline created utilizing deep convolutional neural networks.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"31 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125658186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
Detecting Smiles of Young Children via Deep Transfer Learning 通过深度迁移学习检测幼儿的微笑
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.196
Yu Xia, Di Huang, Yunhong Wang
Smile detection is an interesting topic in computer vision and has received increasing attention in recent years. However, the challenge caused by age variations has not been sufficiently focused on before. In this paper, we first highlight the impact of the discrepancy between infants and adults in a quantitative way on a newly collected database. We then formulate this issue as an unsupervised domain adaptation problem and present the solution of deep transfer learning, which applies the state of the art transfer learning methods, namely Deep Adaptation Networks (DAN) and Joint Adaptation Network (JAN), to two baseline deep models, i.e. AlexNet and ResNet. Thanks to DAN and JAN, the knowledge learned by deep models from adults can be transferred to infants, where very limited labeled data are available for training. Cross-dataset experiments are conducted and the results evidently demonstrate the effectiveness of the proposed approach to smile detection across such an age gap.
微笑检测是计算机视觉领域一个有趣的研究课题,近年来受到越来越多的关注。然而,年龄差异带来的挑战以前没有得到足够的关注。在本文中,我们首先以定量的方式对新收集的数据库强调了婴儿和成人之间差异的影响。然后,我们将此问题表述为无监督域适应问题,并提出了深度迁移学习的解决方案,该解决方案将最先进的迁移学习方法,即深度适应网络(DAN)和联合适应网络(JAN)应用于两个基线深度模型,即AlexNet和ResNet。由于DAN和JAN,深度模型从成人那里学到的知识可以转移到婴儿身上,而婴儿可以用于训练的标记数据非常有限。进行了跨数据集的实验,结果明显表明了该方法在跨越这种年龄差距的微笑检测中的有效性。
{"title":"Detecting Smiles of Young Children via Deep Transfer Learning","authors":"Yu Xia, Di Huang, Yunhong Wang","doi":"10.1109/ICCVW.2017.196","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.196","url":null,"abstract":"Smile detection is an interesting topic in computer vision and has received increasing attention in recent years. However, the challenge caused by age variations has not been sufficiently focused on before. In this paper, we first highlight the impact of the discrepancy between infants and adults in a quantitative way on a newly collected database. We then formulate this issue as an unsupervised domain adaptation problem and present the solution of deep transfer learning, which applies the state of the art transfer learning methods, namely Deep Adaptation Networks (DAN) and Joint Adaptation Network (JAN), to two baseline deep models, i.e. AlexNet and ResNet. Thanks to DAN and JAN, the knowledge learned by deep models from adults can be transferred to infants, where very limited labeled data are available for training. Cross-dataset experiments are conducted and the results evidently demonstrate the effectiveness of the proposed approach to smile detection across such an age gap.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128278314","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
期刊
2017 IEEE International Conference on Computer Vision Workshops (ICCVW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1