首页 > 最新文献

2014 Canadian Conference on Computer and Robot Vision最新文献

英文 中文
Exploring Underwater Environments with Curiosity 带着好奇心探索水下环境
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.22
Yogesh A. Girdhar, G. Dudek
This paper presents a novel approach to modeling curiosity in a mobile robot, which is useful for monitoring and adaptive data collection tasks. We use ROST, a real time topic modeling framework to build a semantic perception model of the environment, using which, we plan a path through the locations in the world with high semantic information content. We demonstrate the approach using the Aqua robot in a variety of different scenarios, and find the robot be able to do tasks such as coral reef inspection, diver following, and sea floor exploration, without any prior training or preparation.
本文提出了一种新的移动机器人好奇心建模方法,该方法可用于监测和自适应数据收集任务。我们使用实时主题建模框架ROST建立了环境的语义感知模型,利用该模型,我们在世界上具有高语义信息含量的位置规划路径。我们在各种不同的场景中使用Aqua机器人演示了这种方法,并发现机器人能够完成诸如珊瑚礁检查,潜水员跟随和海底勘探等任务,而无需任何事先的培训或准备。
{"title":"Exploring Underwater Environments with Curiosity","authors":"Yogesh A. Girdhar, G. Dudek","doi":"10.1109/CRV.2014.22","DOIUrl":"https://doi.org/10.1109/CRV.2014.22","url":null,"abstract":"This paper presents a novel approach to modeling curiosity in a mobile robot, which is useful for monitoring and adaptive data collection tasks. We use ROST, a real time topic modeling framework to build a semantic perception model of the environment, using which, we plan a path through the locations in the world with high semantic information content. We demonstrate the approach using the Aqua robot in a variety of different scenarios, and find the robot be able to do tasks such as coral reef inspection, diver following, and sea floor exploration, without any prior training or preparation.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123443535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
A New Fitness Based Adaptive Parameter Particle Swarm Optimizer 一种新的基于适应度的自适应参数粒子群优化算法
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.52
S. Akhtar, E. Abdel-Rahman, Abdul-Rahim Ahmad
Particle swarm optimization (PSO) is a stochastic global optimization approach whose search characteristics are controlled by three parameters, inertial weight w, cognitive parameter c1 and social parameter c2. Large parameter values facilitate exploration by searching new horizons of solution space. On the other hand, small parameter values facilitate exploitation by searching the neighborhood. An appropriate value of these parameters provides a balance between exploration and exploitation and results in better performance. An adaptive parameter PSO (AP-PSO) algorithm is proposed in this work where the inertial weight is gradually decreased and values of the cognitive and social parameters depend on the fitness values. Good fitness values support exploitation and poor fitness values support exploration. The proposed algorithm has shown excellent performance on low dimensional system identification problems as well as high dimensional articulated human tracking (AHT) problems.
粒子群算法(PSO)是一种随机全局优化算法,其搜索特性由惯性权重w、认知参数c1和社会参数c2三个参数控制。大的参数值有利于通过寻找解空间的新视野来进行探索。另一方面,较小的参数值便于搜索邻域进行开发。这些参数的适当值可以在勘探和开发之间取得平衡,从而获得更好的性能。本文提出了一种自适应参数粒子群算法(AP-PSO),该算法将惯性权重逐渐减小,认知参数和社会参数的取值取决于适应度值。好的健身价值观支持开发,差的健身价值观支持探索。该算法在低维系统识别问题和高维关节人跟踪(AHT)问题上均表现出优异的性能。
{"title":"A New Fitness Based Adaptive Parameter Particle Swarm Optimizer","authors":"S. Akhtar, E. Abdel-Rahman, Abdul-Rahim Ahmad","doi":"10.1109/CRV.2014.52","DOIUrl":"https://doi.org/10.1109/CRV.2014.52","url":null,"abstract":"Particle swarm optimization (PSO) is a stochastic global optimization approach whose search characteristics are controlled by three parameters, inertial weight w, cognitive parameter c1 and social parameter c2. Large parameter values facilitate exploration by searching new horizons of solution space. On the other hand, small parameter values facilitate exploitation by searching the neighborhood. An appropriate value of these parameters provides a balance between exploration and exploitation and results in better performance. An adaptive parameter PSO (AP-PSO) algorithm is proposed in this work where the inertial weight is gradually decreased and values of the cognitive and social parameters depend on the fitness values. Good fitness values support exploitation and poor fitness values support exploration. The proposed algorithm has shown excellent performance on low dimensional system identification problems as well as high dimensional articulated human tracking (AHT) problems.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126003891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Stix-Fusion: A Probabilistic Stixel Integration Technique Stixel - fusion:一种概率Stixel集成技术
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.11
M. Muffert, Nicolai Schneider, U. Franke
In summer 2013, a Mercedes S-Class drove completely autonomously for about 100 km from Mannheim to Pforzheim, Germany, using only close-to-production sensors. In this project, called Mercedes Benz Intelligent Drive, stereo vision was one of the main sensing components. For the representation of free space and obstacles we relied on the so called Stixel World, a generic 3D intermediate representation which is computed from dense disparity images. In spite of the high performance of the Stixel World in most common traffic scenes, the availability of this technique is limited. For instance under adverse weather, rain or even spray water on the windshield results in erroneous disparity images which generate false Stixel results. This can lead to undesired behavior of autonomous vehicles. Our goal is to use the Stixel World for a robust free space estimation and a reliable obstacle detection even during difficult weather conditions. In this paper, we meet this challenge and fuse the Stixels incrementally into a reference grid map. Our new approach is formulated in a Bayesian manner and is based on existence estimation methods. We evaluate our new technique on a manually labeled database with emphasis on bad weather scenarios. The number of structures which are detected mistakenly within free space areas is reduced by a factor of two whereas the detection rate of obstacles increases at the same time.
2013年夏天,一辆奔驰s级轿车从曼海姆(Mannheim)到德国普福尔茨海姆(Pforzheim),完全自动驾驶了大约100公里,只使用了近距离生产的传感器。在这个名为梅赛德斯-奔驰智能驾驶的项目中,立体视觉是主要的传感组件之一。对于自由空间和障碍物的表示,我们依赖于所谓的Stixel World,这是一种从密集视差图像中计算出来的通用3D中间表示。尽管Stixel World在大多数常见的交通场景中具有高性能,但这种技术的可用性是有限的。例如,在恶劣天气下,雨水甚至在挡风玻璃上喷水会导致错误的视差图像,从而产生错误的Stixel结果。这可能会导致自动驾驶汽车的不良行为。我们的目标是使用Stixel World进行稳健的自由空间估计和可靠的障碍物检测,即使在恶劣的天气条件下。在本文中,我们应对了这一挑战,并将Stixels增量地融合到参考网格地图中。我们的新方法是以贝叶斯方法和存在估计方法为基础的。我们在一个人工标记的数据库上评估了我们的新技术,重点是恶劣天气场景。在自由空间区域内被错误检测的结构数量减少了1 / 2,而障碍物的检测率同时提高。
{"title":"Stix-Fusion: A Probabilistic Stixel Integration Technique","authors":"M. Muffert, Nicolai Schneider, U. Franke","doi":"10.1109/CRV.2014.11","DOIUrl":"https://doi.org/10.1109/CRV.2014.11","url":null,"abstract":"In summer 2013, a Mercedes S-Class drove completely autonomously for about 100 km from Mannheim to Pforzheim, Germany, using only close-to-production sensors. In this project, called Mercedes Benz Intelligent Drive, stereo vision was one of the main sensing components. For the representation of free space and obstacles we relied on the so called Stixel World, a generic 3D intermediate representation which is computed from dense disparity images. In spite of the high performance of the Stixel World in most common traffic scenes, the availability of this technique is limited. For instance under adverse weather, rain or even spray water on the windshield results in erroneous disparity images which generate false Stixel results. This can lead to undesired behavior of autonomous vehicles. Our goal is to use the Stixel World for a robust free space estimation and a reliable obstacle detection even during difficult weather conditions. In this paper, we meet this challenge and fuse the Stixels incrementally into a reference grid map. Our new approach is formulated in a Bayesian manner and is based on existence estimation methods. We evaluate our new technique on a manually labeled database with emphasis on bad weather scenarios. The number of structures which are detected mistakenly within free space areas is reduced by a factor of two whereas the detection rate of obstacles increases at the same time.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128270411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Robust Detection of Paint Defects in Moulded Plastic Parts 模塑件漆面缺陷的鲁棒检测
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.48
Cole Tarry, Michael Stachowsky, M. Moussa
A method for detecting local defects in moulded plastic parts is presented. The method uses deflectometry to produce a contrast enhanced image that is later processed in a novel algorithm. The method operates without the need for accurate mechanical models and is robust to changes in image resolution. Experimental results show that the method can detect subtle defects with over 90% accuracy on most parts.
提出了一种检测模塑件局部缺陷的方法。该方法使用偏转术来产生对比度增强的图像,该图像随后在一种新的算法中进行处理。该方法不需要精确的力学模型,并且对图像分辨率的变化具有鲁棒性。实验结果表明,该方法对大多数零件的细微缺陷检测准确率在90%以上。
{"title":"Robust Detection of Paint Defects in Moulded Plastic Parts","authors":"Cole Tarry, Michael Stachowsky, M. Moussa","doi":"10.1109/CRV.2014.48","DOIUrl":"https://doi.org/10.1109/CRV.2014.48","url":null,"abstract":"A method for detecting local defects in moulded plastic parts is presented. The method uses deflectometry to produce a contrast enhanced image that is later processed in a novel algorithm. The method operates without the need for accurate mechanical models and is robust to changes in image resolution. Experimental results show that the method can detect subtle defects with over 90% accuracy on most parts.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134513715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Sign Language Fingerspelling Classification from Depth and Color Images Using a Deep Belief Network 基于深度信念网络的深度和颜色图像手语拼写分类
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.20
Lucas Rioux-Maldague, P. Giguère
Automatic sign language recognition is an open problem that has received a lot of attention recently, not only because of its usefulness to signers, but also due to the numerous applications a sign classifier can have. In this article, we present a new feature extraction technique for hand pose recognition using depth and intensity images captured from a Microsoft Kinect sensor. We applied our technique to American Sign Language finger spelling classification using a Deep Belief Network, for which our feature extraction technique is tailored. We evaluated our results on a multi-user data set with two scenarios: one with all known users and one with an unseen user. We achieved 99% recall and precision on the first, and 77% recall and 79% precision on the second. Our method is also capable of real-time sign classification and is adaptive to any environment or lightning intensity.
自动手语识别是最近受到广泛关注的一个开放问题,不仅是因为它对手语者的有用性,还因为符号分类器可以有许多应用。在这篇文章中,我们提出了一种新的特征提取技术,用于手部姿势识别,使用从微软Kinect传感器捕获的深度和强度图像。我们将该技术应用于使用深度信念网络的美国手语手指拼写分类,我们的特征提取技术是为此量身定制的。我们用两种场景在一个多用户数据集上评估我们的结果:一种是所有已知用户,另一种是不可见用户。我们在第一次上达到了99%的查全率和查准率,在第二次上达到了77%的查全率和79%的查准率。我们的方法还能够实时标记分类,并适应任何环境或闪电强度。
{"title":"Sign Language Fingerspelling Classification from Depth and Color Images Using a Deep Belief Network","authors":"Lucas Rioux-Maldague, P. Giguère","doi":"10.1109/CRV.2014.20","DOIUrl":"https://doi.org/10.1109/CRV.2014.20","url":null,"abstract":"Automatic sign language recognition is an open problem that has received a lot of attention recently, not only because of its usefulness to signers, but also due to the numerous applications a sign classifier can have. In this article, we present a new feature extraction technique for hand pose recognition using depth and intensity images captured from a Microsoft Kinect sensor. We applied our technique to American Sign Language finger spelling classification using a Deep Belief Network, for which our feature extraction technique is tailored. We evaluated our results on a multi-user data set with two scenarios: one with all known users and one with an unseen user. We achieved 99% recall and precision on the first, and 77% recall and 79% precision on the second. Our method is also capable of real-time sign classification and is adaptive to any environment or lightning intensity.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134265366","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
Image Retrieval Using Landmark Indexing for Indoor Navigation 基于地标索引的室内导航图像检索
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.17
Dwaipayan Sinha, M. Ahmed, M. Greenspan
A novel approach is proposed for real-time retrieval of images from a large database of overlapping images of an indoor environment. The procedure extracts visual features from images using selected computer vision techniques, and processes the extracted features to create a reduced list of features annotated with the frame numbers they appear in. This method is named landmark indexing. Unlike some state-of-the-art approaches, the proposed method does not need to consider large image adjacency graphs because the overlap of the images in the map sufficiently increases information gain, and mapping of similar features to the same landmark reduces the search space to improve search efficiency. Empirical evidence from experiments on real datasets shows better performance and accuracy than other approaches. Experiments are further performed by integrating the image retrieval technique into a 3D real-time navigation system. This system is tested in several indoor environments and all experiments show highly accurate localization results.
提出了一种从室内环境的大型重叠图像数据库中实时检索图像的新方法。该程序使用选定的计算机视觉技术从图像中提取视觉特征,并对提取的特征进行处理,以创建用它们出现的帧号注释的简化特征列表。这种方法被命名为地标索引。与现有的一些方法不同,该方法不需要考虑大型图像邻接图,因为地图中图像的重叠足以增加信息增益,并且将相似特征映射到相同的地标上减少了搜索空间,从而提高了搜索效率。在真实数据集上进行的实验表明,与其他方法相比,该方法具有更好的性能和准确性。将图像检索技术整合到三维实时导航系统中,进一步进行了实验。该系统在多个室内环境下进行了测试,所有实验均显示出高精度的定位结果。
{"title":"Image Retrieval Using Landmark Indexing for Indoor Navigation","authors":"Dwaipayan Sinha, M. Ahmed, M. Greenspan","doi":"10.1109/CRV.2014.17","DOIUrl":"https://doi.org/10.1109/CRV.2014.17","url":null,"abstract":"A novel approach is proposed for real-time retrieval of images from a large database of overlapping images of an indoor environment. The procedure extracts visual features from images using selected computer vision techniques, and processes the extracted features to create a reduced list of features annotated with the frame numbers they appear in. This method is named landmark indexing. Unlike some state-of-the-art approaches, the proposed method does not need to consider large image adjacency graphs because the overlap of the images in the map sufficiently increases information gain, and mapping of similar features to the same landmark reduces the search space to improve search efficiency. Empirical evidence from experiments on real datasets shows better performance and accuracy than other approaches. Experiments are further performed by integrating the image retrieval technique into a 3D real-time navigation system. This system is tested in several indoor environments and all experiments show highly accurate localization results.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125138081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Multi-task Learning of Facial Landmarks and Expression 面部标志和表情的多任务学习
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.21
Terrance Devries, Kumar Biswaranjan, Graham W. Taylor
Recently, deep neural networks have been shown to perform competitively on the task of predicting facial expression from images. Trained by gradient-based methods, these networks are amenable to "multi-task" learning via a multiple term objective. In this paper we demonstrate that learning representations to predict the position and shape of facial landmarks can improve expression recognition from images. We show competitive results on two large-scale datasets, the ICML 2013 Facial Expression Recognition challenge, and the Toronto Face Database.
最近,深度神经网络在从图像中预测面部表情的任务上表现得很有竞争力。通过基于梯度的方法训练,这些网络可以通过多个术语目标进行“多任务”学习。在本文中,我们证明了学习表征来预测面部标志的位置和形状可以提高从图像中识别表情。我们展示了两个大规模数据集的竞争结果,ICML 2013面部表情识别挑战和多伦多面部数据库。
{"title":"Multi-task Learning of Facial Landmarks and Expression","authors":"Terrance Devries, Kumar Biswaranjan, Graham W. Taylor","doi":"10.1109/CRV.2014.21","DOIUrl":"https://doi.org/10.1109/CRV.2014.21","url":null,"abstract":"Recently, deep neural networks have been shown to perform competitively on the task of predicting facial expression from images. Trained by gradient-based methods, these networks are amenable to \"multi-task\" learning via a multiple term objective. In this paper we demonstrate that learning representations to predict the position and shape of facial landmarks can improve expression recognition from images. We show competitive results on two large-scale datasets, the ICML 2013 Facial Expression Recognition challenge, and the Toronto Face Database.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117035602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 81
Direct Matrix Factorization and Alignment Refinement: Application to Defect Detection 直接矩阵分解和对齐精化:在缺陷检测中的应用
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.26
Zhen Qin, P. Beek, Xu Chen
Defect detection approaches based on template differencing require precise alignment of the input and template image, however, such alignment is easily affected by the presence of defects. Often, non-trivial pre/post-processing steps and/or manual parameter tuning are needed to remove false alarms, complicating the system and hampering automation. In this work, we explicitly address alignment and defect extraction jointly, and provide a general iterative algorithm to improve both their performance to pixel-wise accuracy. We achieve this by utilizing and extending the robust rank minimization and alignment method of [12]. We propose an effective and efficient optimization algorithm to decompose a template-guided image matrix into a low-rank part relating to alignment-refined defect-free images and an explicit error component containing the defects of interest. Our algorithm is fully automatic, training-free, only needs trivial pre/post-processing procedures, and has few parameters. The rank minimization formulation only requires a linearly correlated template image, and a template-guided approach relieves the common assumption of small defects, making our system very general. We demonstrate the performance of our novel approach qualitatively and quantitatively on a real-world data-set with defects of varying appearance.
基于模板差分的缺陷检测方法要求输入和模板图像的精确对齐,但这种对齐容易受到缺陷存在的影响。通常,需要重要的预处理/后处理步骤和/或手动参数调整来消除假警报,使系统复杂化并阻碍自动化。在这项工作中,我们明确地解决了对齐和缺陷提取的问题,并提供了一个通用的迭代算法来提高它们的性能到像素精度。我们利用并扩展了[12]的鲁棒秩最小化和对齐方法来实现这一点。我们提出了一种有效的优化算法,将模板引导的图像矩阵分解为与对齐精细的无缺陷图像相关的低秩部分和包含感兴趣缺陷的显式误差组件。我们的算法是全自动的,不需要训练,只需要简单的预处理/后处理程序,并且参数很少。秩最小化公式只需要一个线性相关的模板图像,并且模板引导的方法减轻了小缺陷的常见假设,使我们的系统非常通用。我们在具有不同外观缺陷的现实世界数据集上定性和定量地展示了我们的新方法的性能。
{"title":"Direct Matrix Factorization and Alignment Refinement: Application to Defect Detection","authors":"Zhen Qin, P. Beek, Xu Chen","doi":"10.1109/CRV.2014.26","DOIUrl":"https://doi.org/10.1109/CRV.2014.26","url":null,"abstract":"Defect detection approaches based on template differencing require precise alignment of the input and template image, however, such alignment is easily affected by the presence of defects. Often, non-trivial pre/post-processing steps and/or manual parameter tuning are needed to remove false alarms, complicating the system and hampering automation. In this work, we explicitly address alignment and defect extraction jointly, and provide a general iterative algorithm to improve both their performance to pixel-wise accuracy. We achieve this by utilizing and extending the robust rank minimization and alignment method of [12]. We propose an effective and efficient optimization algorithm to decompose a template-guided image matrix into a low-rank part relating to alignment-refined defect-free images and an explicit error component containing the defects of interest. Our algorithm is fully automatic, training-free, only needs trivial pre/post-processing procedures, and has few parameters. The rank minimization formulation only requires a linearly correlated template image, and a template-guided approach relieves the common assumption of small defects, making our system very general. We demonstrate the performance of our novel approach qualitatively and quantitatively on a real-world data-set with defects of varying appearance.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114339417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Generalized Exposure Fusion Weights Estimation 广义曝光融合权重估计
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.8
Mohammed Elamine Moumene, R. Nourine, D. Ziou
Only a small part of the large intensities interval found in high dynamic range scenes can be captured with usual image sensors. This is why delivered images may contain under or overexposed pixels. A popular approach to overcome this problem is to take several images using different exposure parameters, and then fuse them into one single image. This exposure fusion is mostly performed as a weighted average between the corresponding pixels. The challenge is to find weights that produce best fused image quality and in a minimum amount of operations to meet real time requirements. In this paper we present a supervised learning method to estimate generalized exposure fusion weights and we demonstrate how they can be used to fuse any exposures very fast. Subjective and objective comparisons with some relevant works are conducted to prove the effectiveness of the proposed method.
通常的图像传感器只能捕捉到高动态范围场景中大强度间隔的一小部分。这就是为什么交付的图像可能包含曝光不足或曝光过度的像素。克服这个问题的一种流行方法是使用不同的曝光参数拍摄几张图像,然后将它们融合成一张图像。这种曝光融合主要是作为相应像素之间的加权平均来执行的。挑战在于找到产生最佳融合图像质量的权重,并以最少的操作满足实时要求。在本文中,我们提出了一种监督学习方法来估计广义暴露融合权值,并演示了如何使用它们来快速融合任何暴露。通过与相关文献的主客观对比,证明了所提方法的有效性。
{"title":"Generalized Exposure Fusion Weights Estimation","authors":"Mohammed Elamine Moumene, R. Nourine, D. Ziou","doi":"10.1109/CRV.2014.8","DOIUrl":"https://doi.org/10.1109/CRV.2014.8","url":null,"abstract":"Only a small part of the large intensities interval found in high dynamic range scenes can be captured with usual image sensors. This is why delivered images may contain under or overexposed pixels. A popular approach to overcome this problem is to take several images using different exposure parameters, and then fuse them into one single image. This exposure fusion is mostly performed as a weighted average between the corresponding pixels. The challenge is to find weights that produce best fused image quality and in a minimum amount of operations to meet real time requirements. In this paper we present a supervised learning method to estimate generalized exposure fusion weights and we demonstrate how they can be used to fuse any exposures very fast. Subjective and objective comparisons with some relevant works are conducted to prove the effectiveness of the proposed method.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129114664","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
A More Robust Feature Correspondence for more Accurate Image Recognition 一种更鲁棒的特征对应关系,用于更准确的图像识别
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.32
Shady Y. El-Mashad, A. Shoukry
In this paper, a novel algorithm for finding the optimal correspondence between two sets of image features has been introduced. The proposed algorithm pays attention not only to the similarity between features but also to the spatial layout of every matched feature and its neighbors. Unlike related methods that use geometrical relations between the neighboring features, the proposed method employees topology that survives against different types of deformations like scaling and rotation, resulting in more robust matching. The features are expressed as an undirected graph where every node represents a local feature and every edge represents adjacency between them. The topology of the resulting graph can be considered as a robust global feature of the represented object. The matching process is modeled as a graph matching problem, which in turn is formulated as a variation of the quadratic assignment problem. In this variation, a number of parameters are used to control the significance of global vs. local features to tune the performance and customize the model. The experimental results show a significant improvement in the number of correct matches using the proposed method compared to different methods.
本文提出了一种寻找两组图像特征之间最优对应关系的新算法。该算法不仅关注特征之间的相似性,而且关注每个匹配特征及其相邻特征的空间布局。与使用相邻特征之间的几何关系的相关方法不同,该方法使用的拓扑结构可以抵抗不同类型的变形(如缩放和旋转),从而产生更鲁棒的匹配。特征表示为无向图,其中每个节点表示一个局部特征,每个边表示它们之间的邻接关系。结果图的拓扑结构可以被认为是所表示对象的鲁棒全局特征。匹配过程被建模为一个图匹配问题,而图匹配问题又被表述为二次分配问题的一个变体。在这种变体中,使用许多参数来控制全局特征与局部特征的重要性,以调整性能并自定义模型。实验结果表明,与其他方法相比,该方法在匹配正确率上有显著提高。
{"title":"A More Robust Feature Correspondence for more Accurate Image Recognition","authors":"Shady Y. El-Mashad, A. Shoukry","doi":"10.1109/CRV.2014.32","DOIUrl":"https://doi.org/10.1109/CRV.2014.32","url":null,"abstract":"In this paper, a novel algorithm for finding the optimal correspondence between two sets of image features has been introduced. The proposed algorithm pays attention not only to the similarity between features but also to the spatial layout of every matched feature and its neighbors. Unlike related methods that use geometrical relations between the neighboring features, the proposed method employees topology that survives against different types of deformations like scaling and rotation, resulting in more robust matching. The features are expressed as an undirected graph where every node represents a local feature and every edge represents adjacency between them. The topology of the resulting graph can be considered as a robust global feature of the represented object. The matching process is modeled as a graph matching problem, which in turn is formulated as a variation of the quadratic assignment problem. In this variation, a number of parameters are used to control the significance of global vs. local features to tune the performance and customize the model. The experimental results show a significant improvement in the number of correct matches using the proposed method compared to different methods.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128143857","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2014 Canadian Conference on Computer and Robot Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1