首页 > 最新文献

2017 IEEE International Conference on Computer Vision Workshops (ICCVW)最新文献

英文 中文
Fusing Geometry and Appearance for Road Segmentation 融合几何和外观的道路分割
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.28
Gong Cheng, Yiming Qian, J. Elder
We propose a novel method for fusing geometric and appearance cues for road surface segmentation. Modeling colour cues using Gaussian mixtures allows the fusion to be performed optimally within a Bayesian framework, avoiding ad hoc weights. Adaptation to different scene conditions is accomplished through nearest-neighbour appearance model selection over a dictionary of mixture models learned from training data, and the thorny problem of selecting the number of components in each mixture is solved through a novel cross-validation approach. Quantitative evaluation reveals that the proposed fusion method significantly improves segmentation accuracy relative to a method that uses geometric cues alone.
我们提出了一种融合几何和外观线索的路面分割新方法。使用高斯混合建模颜色线索允许在贝叶斯框架内最佳地进行融合,避免了特别的权重。通过从训练数据中学习到的混合模型字典中选择最近邻外观模型来实现对不同场景条件的适应,并通过一种新的交叉验证方法解决了每种混合物中选择组件数量的棘手问题。定量评价表明,与单独使用几何线索的方法相比,所提出的融合方法显著提高了分割精度。
{"title":"Fusing Geometry and Appearance for Road Segmentation","authors":"Gong Cheng, Yiming Qian, J. Elder","doi":"10.1109/ICCVW.2017.28","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.28","url":null,"abstract":"We propose a novel method for fusing geometric and appearance cues for road surface segmentation. Modeling colour cues using Gaussian mixtures allows the fusion to be performed optimally within a Bayesian framework, avoiding ad hoc weights. Adaptation to different scene conditions is accomplished through nearest-neighbour appearance model selection over a dictionary of mixture models learned from training data, and the thorny problem of selecting the number of components in each mixture is solved through a novel cross-validation approach. Quantitative evaluation reveals that the proposed fusion method significantly improves segmentation accuracy relative to a method that uses geometric cues alone.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116768027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Discrepancy-Based Networks for Unsupervised Domain Adaptation: A Comparative Study 基于差异的无监督域自适应网络的比较研究
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.312
G. Csurka, Fabien Baradel, Boris Chidlovskii, S. Clinchant
Domain Adaptation (DA) exploits labeled data and models from similar domains in order to alleviate the annotation burden when learning a model in a new domain. Our contribution to the field is three-fold. First, we propose a new dataset, LandMarkDA, to study the adaptation between landmark place recognition models trained with different artistic image styles, such as photos, paintings and drawings. The new LandMarkDA proposes new adaptation challenges, where current deep architectures show their limits. Second, we propose an experimental study of recent shallow and deep adaptation networks, based on using Maximum Mean Discrepancy to bridge the domain gap. We study different design choices for these models by varying the network architectures and evaluate them on OFF31 and the new LandMarkDA collections. We show that shallow networks can still be competitive under an appropriate feature extraction. Finally, we also benchmark a new DA method that successfully combines the artistic image style-transfer with deep discrepancy-based networks.
领域适应(Domain Adaptation, DA)利用来自相似领域的标记数据和模型,以减轻在新领域学习模型时的标注负担。我们对这个领域的贡献是三重的。首先,我们提出了一个新的数据集LandMarkDA,研究了不同艺术图像风格(如照片、绘画和素描)训练的地标性地点识别模型之间的自适应。新的LandMarkDA提出了新的适应挑战,当前的深度架构显示出其局限性。其次,我们提出了一种基于最大均值差异来弥补域差距的浅层和深层自适应网络的实验研究。我们通过改变网络架构来研究这些模型的不同设计选择,并在OFF31和新的LandMarkDA集合上对它们进行评估。我们表明,在适当的特征提取下,浅网络仍然可以具有竞争力。最后,我们还测试了一种新的数据处理方法,该方法成功地将艺术图像风格转移与基于深度差异的网络相结合。
{"title":"Discrepancy-Based Networks for Unsupervised Domain Adaptation: A Comparative Study","authors":"G. Csurka, Fabien Baradel, Boris Chidlovskii, S. Clinchant","doi":"10.1109/ICCVW.2017.312","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.312","url":null,"abstract":"Domain Adaptation (DA) exploits labeled data and models from similar domains in order to alleviate the annotation burden when learning a model in a new domain. Our contribution to the field is three-fold. First, we propose a new dataset, LandMarkDA, to study the adaptation between landmark place recognition models trained with different artistic image styles, such as photos, paintings and drawings. The new LandMarkDA proposes new adaptation challenges, where current deep architectures show their limits. Second, we propose an experimental study of recent shallow and deep adaptation networks, based on using Maximum Mean Discrepancy to bridge the domain gap. We study different design choices for these models by varying the network architectures and evaluate them on OFF31 and the new LandMarkDA collections. We show that shallow networks can still be competitive under an appropriate feature extraction. Finally, we also benchmark a new DA method that successfully combines the artistic image style-transfer with deep discrepancy-based networks.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"124 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123312387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Towards a Spatio-Temporal Atlas of 3D Cellular Parameters During Leaf Morphogenesis 叶片形态发生过程中三维细胞参数的时空图谱
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.14
F. Selka, T. Blein, J. Burguet, E. Biot, P. Laufs, P. Andrey
Morphogenesis is a complex process that integrates several mechanisms from the molecular to the organ scales. In plants, division and growth are the two fundamental cellular mechanisms that drive morphogenesis. However, little is known about how these mechanisms are coordinated to establish functional tissue structure. A fundamental bottleneck is the current lack of techniques to systematically quantify the spatio-temporal evolution of 3D cell morphology during organ growth. Using leaf development as a relevant and challenging model to study morphogenesis, we developed a computational framework for cell analysis and quantification from 3D images and for the generation of 3D cell shape atlas. A remarkable feature of leaf morphogenesis being the formation of a laminar-like structure, we propose to automatically separate the cells corresponding to the leaf sides in the segmented leaves, by applying a clustering algorithm. The performance of the proposed pipeline was experimentally assessed on a dataset of 46 leaves in an early developmental state.
形态发生是一个复杂的过程,整合了从分子到器官尺度的多种机制。在植物中,分裂和生长是驱动形态发生的两个基本细胞机制。然而,人们对这些机制如何协调建立功能性组织结构知之甚少。一个基本的瓶颈是目前缺乏技术来系统地量化三维细胞形态在器官生长过程中的时空演变。利用叶片发育作为一个相关且具有挑战性的模型来研究形态发生,我们开发了一个计算框架,用于从3D图像中进行细胞分析和量化,并用于生成3D细胞形状图谱。由于叶片形态发生的显著特征是层状结构的形成,我们提出了采用聚类算法自动分离叶片中对应叶侧的细胞。在46个处于早期发育状态的叶片数据集上,实验评估了所提出的管道的性能。
{"title":"Towards a Spatio-Temporal Atlas of 3D Cellular Parameters During Leaf Morphogenesis","authors":"F. Selka, T. Blein, J. Burguet, E. Biot, P. Laufs, P. Andrey","doi":"10.1109/ICCVW.2017.14","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.14","url":null,"abstract":"Morphogenesis is a complex process that integrates several mechanisms from the molecular to the organ scales. In plants, division and growth are the two fundamental cellular mechanisms that drive morphogenesis. However, little is known about how these mechanisms are coordinated to establish functional tissue structure. A fundamental bottleneck is the current lack of techniques to systematically quantify the spatio-temporal evolution of 3D cell morphology during organ growth. Using leaf development as a relevant and challenging model to study morphogenesis, we developed a computational framework for cell analysis and quantification from 3D images and for the generation of 3D cell shape atlas. A remarkable feature of leaf morphogenesis being the formation of a laminar-like structure, we propose to automatically separate the cells corresponding to the leaf sides in the segmented leaves, by applying a clustering algorithm. The performance of the proposed pipeline was experimentally assessed on a dataset of 46 leaves in an early developmental state.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"146 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123391354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Learning to Identify While Failing to Discriminate 学会认同而不歧视
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.298
Jure Sokolić, M. Rodrigues, Qiang Qiu, G. Sapiro
Privacy and fairness are critical in computer vision applications, in particular when dealing with human identification. Achieving a universally secure, private, and fair systems is practically impossible as the exploitation of additional data can reveal private information in the original one. Faced with this challenge, we propose a new line of research, where the privacy is learned and used in a closed environment. The goal is to ensure that a given entity, trusted to infer certain information with our data, is blocked from inferring protected information from it. We design a system that learns to succeed on the positive task while simultaneously fail at the negative one, and illustrate this with challenging cases where the positive task (face verification) is harder than the negative one (gender classification). The framework opens the door to privacy and fairness in very important closed scenarios, ranging from private data accumulation companies to law-enforcement and hospitals.
隐私和公平在计算机视觉应用中是至关重要的,特别是在处理人类身份识别时。实现普遍安全、私密、公平的系统实际上是不可能的,因为利用额外的数据可能会泄露原始数据中的私人信息。面对这一挑战,我们提出了一个新的研究方向,在一个封闭的环境中学习和使用隐私。目标是确保一个给定的实体(被信任可以从我们的数据推断出某些信息)被阻止从它推断出受保护的信息。我们设计了一个系统,学习在积极任务上取得成功,同时在消极任务上失败,并通过积极任务(面部识别)比消极任务(性别分类)更难的挑战性案例来说明这一点。该框架在非常重要的封闭场景(从私人数据积累公司到执法部门和医院)中为隐私和公平打开了大门。
{"title":"Learning to Identify While Failing to Discriminate","authors":"Jure Sokolić, M. Rodrigues, Qiang Qiu, G. Sapiro","doi":"10.1109/ICCVW.2017.298","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.298","url":null,"abstract":"Privacy and fairness are critical in computer vision applications, in particular when dealing with human identification. Achieving a universally secure, private, and fair systems is practically impossible as the exploitation of additional data can reveal private information in the original one. Faced with this challenge, we propose a new line of research, where the privacy is learned and used in a closed environment. The goal is to ensure that a given entity, trusted to infer certain information with our data, is blocked from inferring protected information from it. We design a system that learns to succeed on the positive task while simultaneously fail at the negative one, and illustrate this with challenging cases where the positive task (face verification) is harder than the negative one (gender classification). The framework opens the door to privacy and fairness in very important closed scenarios, ranging from private data accumulation companies to law-enforcement and hospitals.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125128276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Reliable Isometric Point Correspondence from Depth 可靠的深度等距点对应
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.152
Emel Küpçü, Y. Yemez
We propose a new iterative isometric point correspondence method that relies on diffusion distance to handle challenges posed by commodity depth sensors, which usually provide incomplete and noisy surface data exhibiting holes and gaps. We formulate the correspondence problem as finding an optimal partial mapping between two given point sets, that minimizes deviation from isometry. Our algorithm starts with an initial rough correspondence between keypoints, obtained via a standard descriptor matching technique. This initial correspondence is then pruned and updated by iterating a perfect matching algorithm until convergence to find as many reliable correspondences as possible. For shapes with intrinsic symmetries such as human models, we additionally provide a symmetry aware extension to improve our formulation. The experiments show that our method provides state of the art performance over depth frames exhibiting occlusions, large deformations and topological noise.
我们提出了一种新的迭代等距点对应方法,该方法依赖于扩散距离来处理商品深度传感器带来的挑战,这些传感器通常提供不完整和有噪声的地表数据,显示孔洞和间隙。我们将对应问题表述为寻找两个给定点集之间的最优部分映射,使与等距的偏差最小化。我们的算法从关键点之间的初始粗略对应开始,通过标准描述符匹配技术获得。然后通过迭代完美匹配算法对初始对应进行修剪和更新,直到收敛以找到尽可能多的可靠对应。对于具有内在对称性的形状,如人体模型,我们还提供了一个对称感知扩展来改进我们的公式。实验表明,我们的方法在具有遮挡、大变形和拓扑噪声的深度帧上提供了最先进的性能。
{"title":"Reliable Isometric Point Correspondence from Depth","authors":"Emel Küpçü, Y. Yemez","doi":"10.1109/ICCVW.2017.152","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.152","url":null,"abstract":"We propose a new iterative isometric point correspondence method that relies on diffusion distance to handle challenges posed by commodity depth sensors, which usually provide incomplete and noisy surface data exhibiting holes and gaps. We formulate the correspondence problem as finding an optimal partial mapping between two given point sets, that minimizes deviation from isometry. Our algorithm starts with an initial rough correspondence between keypoints, obtained via a standard descriptor matching technique. This initial correspondence is then pruned and updated by iterating a perfect matching algorithm until convergence to find as many reliable correspondences as possible. For shapes with intrinsic symmetries such as human models, we additionally provide a symmetry aware extension to improve our formulation. The experiments show that our method provides state of the art performance over depth frames exhibiting occlusions, large deformations and topological noise.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123994489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Local Depth Edge Detection in Humans and Deep Neural Networks 人体局部深度边缘检测与深度神经网络
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.316
Krista A. Ehinger, E. Graf, W. Adams, J. Elder
Distinguishing edges caused by a change in depth from other types of edges is an important problem in early vision. We investigate the performance of humans and computer vision models on this task. We use spherical imagery with ground-truth LiDAR range data to build an objective ground-truth dataset for edge classification. We compare various computational models for classifying depth from non-depth edges in small images patches and achieve the best performance (86%) with a convolutional neural network. We investigate human performance on this task in a behavioral experiment and find that human performance is lower than the CNN. Although human and CNN depth responses are correlated, observers' responses are better predicted by other observers than by the CNN. The responses of CNNs and human observers also show a slightly different pattern of correlation with low-level edge cues, which suggests that CNNs and human observers may weight these features differently for classifying edges.
区分深度变化引起的边缘和其他类型的边缘是早期视觉中的一个重要问题。我们研究了人类和计算机视觉模型在这个任务上的表现。我们利用球面图像和真地激光雷达距离数据建立了一个客观的真地数据集,用于边缘分类。我们比较了各种计算模型在小图像斑块中从非深度边缘分类深度,并使用卷积神经网络获得了最佳性能(86%)。我们在一个行为实验中调查了人类在这个任务上的表现,发现人类的表现低于CNN。虽然人类和CNN的深度响应是相关的,但观察者的反应被其他观察者比CNN更好地预测。cnn和人类观察者的反应也显示出与低水平边缘线索的关联模式略有不同,这表明cnn和人类观察者在对边缘进行分类时可能对这些特征的权重不同。
{"title":"Local Depth Edge Detection in Humans and Deep Neural Networks","authors":"Krista A. Ehinger, E. Graf, W. Adams, J. Elder","doi":"10.1109/ICCVW.2017.316","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.316","url":null,"abstract":"Distinguishing edges caused by a change in depth from other types of edges is an important problem in early vision. We investigate the performance of humans and computer vision models on this task. We use spherical imagery with ground-truth LiDAR range data to build an objective ground-truth dataset for edge classification. We compare various computational models for classifying depth from non-depth edges in small images patches and achieve the best performance (86%) with a convolutional neural network. We investigate human performance on this task in a behavioral experiment and find that human performance is lower than the CNN. Although human and CNN depth responses are correlated, observers' responses are better predicted by other observers than by the CNN. The responses of CNNs and human observers also show a slightly different pattern of correlation with low-level edge cues, which suggests that CNNs and human observers may weight these features differently for classifying edges.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128928736","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Locating Crop Plant Centers from UAV-Based RGB Imagery 从基于无人机的RGB图像定位农作物中心
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.238
Yuhao Chen, Javier Ribera, C. Boomsma, E. Delp
In this paper we propose a method to find the location of crop plants in Unmanned Aerial Vehicle (UAV) imagery. Finding the location of plants is a crucial step to derive and track phenotypic traits for each plant. We describe some initial work in estimating field crop plant locations. We approach the problem by classifying pixels as a plant center or a non plant center. We use Multiple Instance Learning (MIL) to handle the ambiguity of plant center labeling in training data. The classification results are then post-processed to estimate the exact location of the crop plant. Experimental evaluation is conducted to evaluate the method and the result achieved an overall precision and recall of 66% and 64%, respectively.
本文提出了一种在无人机(UAV)图像中寻找农作物位置的方法。寻找植物的位置是推导和追踪每种植物表型性状的关键步骤。我们描述了估算田间作物种植位置的一些初步工作。我们通过将像素分类为植物中心或非植物中心来解决这个问题。我们使用多实例学习(MIL)来处理训练数据中植物中心标注的模糊性。然后对分类结果进行后处理,以估计作物的确切位置。对该方法进行了实验评价,结果表明,该方法的总体查准率和查全率分别达到66%和64%。
{"title":"Locating Crop Plant Centers from UAV-Based RGB Imagery","authors":"Yuhao Chen, Javier Ribera, C. Boomsma, E. Delp","doi":"10.1109/ICCVW.2017.238","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.238","url":null,"abstract":"In this paper we propose a method to find the location of crop plants in Unmanned Aerial Vehicle (UAV) imagery. Finding the location of plants is a crucial step to derive and track phenotypic traits for each plant. We describe some initial work in estimating field crop plant locations. We approach the problem by classifying pixels as a plant center or a non plant center. We use Multiple Instance Learning (MIL) to handle the ambiguity of plant center labeling in training data. The classification results are then post-processed to estimate the exact location of the crop plant. Experimental evaluation is conducted to evaluate the method and the result achieved an overall precision and recall of 66% and 64%, respectively.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130721906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Detecting Smiles of Young Children via Deep Transfer Learning 通过深度迁移学习检测幼儿的微笑
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.196
Yu Xia, Di Huang, Yunhong Wang
Smile detection is an interesting topic in computer vision and has received increasing attention in recent years. However, the challenge caused by age variations has not been sufficiently focused on before. In this paper, we first highlight the impact of the discrepancy between infants and adults in a quantitative way on a newly collected database. We then formulate this issue as an unsupervised domain adaptation problem and present the solution of deep transfer learning, which applies the state of the art transfer learning methods, namely Deep Adaptation Networks (DAN) and Joint Adaptation Network (JAN), to two baseline deep models, i.e. AlexNet and ResNet. Thanks to DAN and JAN, the knowledge learned by deep models from adults can be transferred to infants, where very limited labeled data are available for training. Cross-dataset experiments are conducted and the results evidently demonstrate the effectiveness of the proposed approach to smile detection across such an age gap.
微笑检测是计算机视觉领域一个有趣的研究课题,近年来受到越来越多的关注。然而,年龄差异带来的挑战以前没有得到足够的关注。在本文中,我们首先以定量的方式对新收集的数据库强调了婴儿和成人之间差异的影响。然后,我们将此问题表述为无监督域适应问题,并提出了深度迁移学习的解决方案,该解决方案将最先进的迁移学习方法,即深度适应网络(DAN)和联合适应网络(JAN)应用于两个基线深度模型,即AlexNet和ResNet。由于DAN和JAN,深度模型从成人那里学到的知识可以转移到婴儿身上,而婴儿可以用于训练的标记数据非常有限。进行了跨数据集的实验,结果明显表明了该方法在跨越这种年龄差距的微笑检测中的有效性。
{"title":"Detecting Smiles of Young Children via Deep Transfer Learning","authors":"Yu Xia, Di Huang, Yunhong Wang","doi":"10.1109/ICCVW.2017.196","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.196","url":null,"abstract":"Smile detection is an interesting topic in computer vision and has received increasing attention in recent years. However, the challenge caused by age variations has not been sufficiently focused on before. In this paper, we first highlight the impact of the discrepancy between infants and adults in a quantitative way on a newly collected database. We then formulate this issue as an unsupervised domain adaptation problem and present the solution of deep transfer learning, which applies the state of the art transfer learning methods, namely Deep Adaptation Networks (DAN) and Joint Adaptation Network (JAN), to two baseline deep models, i.e. AlexNet and ResNet. Thanks to DAN and JAN, the knowledge learned by deep models from adults can be transferred to infants, where very limited labeled data are available for training. Cross-dataset experiments are conducted and the results evidently demonstrate the effectiveness of the proposed approach to smile detection across such an age gap.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128278314","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Are They Going to Cross? A Benchmark Dataset and Baseline for Pedestrian Crosswalk Behavior 他们要过马路吗?人行横道行为的基准数据集和基线
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.33
Amir Rasouli, Iuliia Kotseruba, John K. Tsotsos
Designing autonomous vehicles suitable for urban environments remains an unresolved problem. One of the major dilemmas faced by autonomous cars is how to understand the intention of other road users and communicate with them. The existing datasets do not provide the necessary means for such higher level analysis of traffic scenes. With this in mind, we introduce a novel dataset which in addition to providing the bounding box information for pedestrian detection, also includes the behavioral and contextual annotations for the scenes. This allows combining visual and semantic information for better understanding of pedestrians' intentions in various traffic scenarios. We establish baseline approaches for analyzing the data and show that combining visual and contextual information can improve prediction of pedestrian intention at the point of crossing by at least 20%.
设计适合城市环境的自动驾驶汽车仍然是一个未解决的问题。自动驾驶汽车面临的主要难题之一是如何理解其他道路使用者的意图并与他们沟通。现有的数据集无法提供对交通场景进行更高层次分析的必要手段。考虑到这一点,我们引入了一个新的数据集,除了为行人检测提供边界框信息外,还包括场景的行为和上下文注释。这可以将视觉和语义信息结合起来,更好地理解行人在各种交通场景中的意图。我们建立了基线方法来分析数据,并表明结合视觉和上下文信息可以将行人意图的预测提高至少20%。
{"title":"Are They Going to Cross? A Benchmark Dataset and Baseline for Pedestrian Crosswalk Behavior","authors":"Amir Rasouli, Iuliia Kotseruba, John K. Tsotsos","doi":"10.1109/ICCVW.2017.33","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.33","url":null,"abstract":"Designing autonomous vehicles suitable for urban environments remains an unresolved problem. One of the major dilemmas faced by autonomous cars is how to understand the intention of other road users and communicate with them. The existing datasets do not provide the necessary means for such higher level analysis of traffic scenes. With this in mind, we introduce a novel dataset which in addition to providing the bounding box information for pedestrian detection, also includes the behavioral and contextual annotations for the scenes. This allows combining visual and semantic information for better understanding of pedestrians' intentions in various traffic scenarios. We establish baseline approaches for analyzing the data and show that combining visual and contextual information can improve prediction of pedestrian intention at the point of crossing by at least 20%.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128787549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 200
Max-Boost-GAN: Max Operation to Boost Generative Ability of Generative Adversarial Networks Max-Boost- gan:提高生成对抗网络生成能力的最大操作
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.140
Xinhan Di, Pengqian Yu
Generative adversarial networks (GANs) can be used to learn a generation function from a joint probability distribution as an input, and then visual samples with semantic properties can be generated from a marginal probability distribution. In this paper, we propose a novel algorithm named Max-Boost-GAN, which is demonstrated to boost the generative ability of GANs when the error of generation is upper bounded. Moreover, the Max-Boost-GAN can be used to learn the generation functions from two marginal probability distributions as the input, and samples of higher visual quality and variety could be generated from the joint probability distribution. Finally, novel objective functions are proposed for obtaining convergence during training the Max-Boost-GAN. Experiments on the generation of binary digits and RGB human faces show that the Max-Boost-GAN achieves boosted ability of generation as expected.
生成式对抗网络(GANs)可以从联合概率分布中学习生成函数作为输入,然后从边缘概率分布中生成具有语义属性的视觉样本。在本文中,我们提出了一种新的算法Max-Boost-GAN,当生成误差为上界时,该算法可以提高gan的生成能力。此外,Max-Boost-GAN可以从两个边缘概率分布中学习生成函数作为输入,并且可以从联合概率分布中生成更高视觉质量和多样性的样本。最后,提出了在训练Max-Boost-GAN时获得收敛性的新目标函数。对二进制数和RGB人脸的生成实验表明,Max-Boost-GAN达到了预期的增强生成能力。
{"title":"Max-Boost-GAN: Max Operation to Boost Generative Ability of Generative Adversarial Networks","authors":"Xinhan Di, Pengqian Yu","doi":"10.1109/ICCVW.2017.140","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.140","url":null,"abstract":"Generative adversarial networks (GANs) can be used to learn a generation function from a joint probability distribution as an input, and then visual samples with semantic properties can be generated from a marginal probability distribution. In this paper, we propose a novel algorithm named Max-Boost-GAN, which is demonstrated to boost the generative ability of GANs when the error of generation is upper bounded. Moreover, the Max-Boost-GAN can be used to learn the generation functions from two marginal probability distributions as the input, and samples of higher visual quality and variety could be generated from the joint probability distribution. Finally, novel objective functions are proposed for obtaining convergence during training the Max-Boost-GAN. Experiments on the generation of binary digits and RGB human faces show that the Max-Boost-GAN achieves boosted ability of generation as expected.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126629739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2017 IEEE International Conference on Computer Vision Workshops (ICCVW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1