首页 > 最新文献

The 2nd Canadian Conference on Computer and Robot Vision (CRV'05)最新文献

英文 中文
3D reconstruction of environments for planetary exploration 行星探测环境的三维重建
Pub Date : 2005-05-09 DOI: 10.1109/CRV.2005.3
S. Gemme, J. Bakambu, Ioannis M. Rekleitis
In this paper we present our approach to 3D surface reconstruction from large sparse range data sets. In space robotics constructing an accurate model of the environment is very important for a variety of reasons. In particular, the constructed model can be used for: safe tele-operation, path planning, planetary exploration and mapping of points of interest. Our approach is based on acquiring range scans from different view-points with overlapping regions, merge them together into a single data set, and fit a triangular mesh on the merged data points. We demonstrate the effectiveness of our approach in a path planning scenario and also by creating the accessibility map for a portion of the Mars Yard located in the Canadian Space Agency.
在本文中,我们提出了一种基于大稀疏范围数据集的三维表面重建方法。在空间机器人中,由于各种原因,建立一个精确的环境模型是非常重要的。构建的模型可用于安全远程操作、路径规划、行星探测和兴趣点测绘。我们的方法是基于从具有重叠区域的不同视点获取距离扫描,将它们合并到一个单一的数据集,并在合并的数据点上拟合三角形网格。我们在路径规划场景中展示了我们方法的有效性,并为位于加拿大航天局的火星场的一部分创建了可达性地图。
{"title":"3D reconstruction of environments for planetary exploration","authors":"S. Gemme, J. Bakambu, Ioannis M. Rekleitis","doi":"10.1109/CRV.2005.3","DOIUrl":"https://doi.org/10.1109/CRV.2005.3","url":null,"abstract":"In this paper we present our approach to 3D surface reconstruction from large sparse range data sets. In space robotics constructing an accurate model of the environment is very important for a variety of reasons. In particular, the constructed model can be used for: safe tele-operation, path planning, planetary exploration and mapping of points of interest. Our approach is based on acquiring range scans from different view-points with overlapping regions, merge them together into a single data set, and fit a triangular mesh on the merged data points. We demonstrate the effectiveness of our approach in a path planning scenario and also by creating the accessibility map for a portion of the Mars Yard located in the Canadian Space Agency.","PeriodicalId":307318,"journal":{"name":"The 2nd Canadian Conference on Computer and Robot Vision (CRV'05)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127700157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Identifying precursory cancer lesions using temporal texture analysis 利用时间纹理分析识别癌前病变
Pub Date : 2005-05-09 DOI: 10.1109/CRV.2005.48
Aldrin Barreto-Flores, L. A. Robles, Rosa Maria Morales Tepalt, J. Aragon
This paper describes a method for the temporal analysis of texture in colposcopy. The objective is to find temporal texture patterns in order to detect precursory cancer lesions analyzing colposcopy video frames. Preprocessing of the frames is necessary in order to deal with patient movement and non uniform illumination. We use a stabilization algorithm based in a homography and to eliminate incorrect transformations between frames. Illumination correction is done using a local pixel transformation based in the mean around a small window. Temporal reaction after acetic acid application in the cervix is evaluated through the use of a co-occurrence matrix in different regions of the cervix. The reaction is plotted and analyzed through time. Different patterns for normal and abnormal regions are found by this temporal texture analysis showing the possibility to detect important lesions. The proposed method uses standard colposcopy equipment and it was tested using sequences obtained from different patients.
本文介绍了一种阴道镜下纹理的时域分析方法。目的是找到时间纹理模式,以便检测前期癌症病变分析阴道镜视频帧。帧的预处理是必要的,以处理病人的运动和不均匀的照明。我们使用了一种基于单应性的稳定算法来消除帧间不正确的变换。照明校正是使用基于小窗口周围平均值的局部像素变换来完成的。通过在子宫颈不同区域使用共现矩阵来评估醋酸应用于子宫颈后的时间反应。反应被绘制出来并随时间进行分析。正常和异常区域的不同模式是通过这种时间纹理分析发现的,显示了检测重要病变的可能性。所提出的方法使用标准阴道镜检查设备,并使用从不同患者获得的序列进行测试。
{"title":"Identifying precursory cancer lesions using temporal texture analysis","authors":"Aldrin Barreto-Flores, L. A. Robles, Rosa Maria Morales Tepalt, J. Aragon","doi":"10.1109/CRV.2005.48","DOIUrl":"https://doi.org/10.1109/CRV.2005.48","url":null,"abstract":"This paper describes a method for the temporal analysis of texture in colposcopy. The objective is to find temporal texture patterns in order to detect precursory cancer lesions analyzing colposcopy video frames. Preprocessing of the frames is necessary in order to deal with patient movement and non uniform illumination. We use a stabilization algorithm based in a homography and to eliminate incorrect transformations between frames. Illumination correction is done using a local pixel transformation based in the mean around a small window. Temporal reaction after acetic acid application in the cervix is evaluated through the use of a co-occurrence matrix in different regions of the cervix. The reaction is plotted and analyzed through time. Different patterns for normal and abnormal regions are found by this temporal texture analysis showing the possibility to detect important lesions. The proposed method uses standard colposcopy equipment and it was tested using sequences obtained from different patients.","PeriodicalId":307318,"journal":{"name":"The 2nd Canadian Conference on Computer and Robot Vision (CRV'05)","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123217028","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Entropy-based image merging 基于熵的图像合并
Pub Date : 2005-05-09 DOI: 10.1109/CRV.2005.38
A. German, M. Jenkin, Y. Lespérance
Spacecraft docking using vision is a challenging task. Not least among the problems encountered is the need to visually localize the docking target. Here we consider the task of adapting the local illumination to assist in this docking. An online approach is developed that combines images obtained under different exposure and lighting conditions into a single image upon which docking decisions can be made. This method is designed to be used within an intelligent controller that automatically adjusts lighting and image acquisition in order to obtain the "best" possible composite view of the target for further image processing.
利用视觉进行航天器对接是一项具有挑战性的任务。遇到的最重要的问题是需要可视化地定位对接目标。在这里,我们考虑的任务是调整局部照明来协助这种对接。开发了一种在线方法,将在不同曝光和光照条件下获得的图像合并为单个图像,从而可以做出对接决策。该方法被设计用于智能控制器中,该控制器可以自动调整照明和图像采集,以便获得目标的“最佳”复合视图,以进行进一步的图像处理。
{"title":"Entropy-based image merging","authors":"A. German, M. Jenkin, Y. Lespérance","doi":"10.1109/CRV.2005.38","DOIUrl":"https://doi.org/10.1109/CRV.2005.38","url":null,"abstract":"Spacecraft docking using vision is a challenging task. Not least among the problems encountered is the need to visually localize the docking target. Here we consider the task of adapting the local illumination to assist in this docking. An online approach is developed that combines images obtained under different exposure and lighting conditions into a single image upon which docking decisions can be made. This method is designed to be used within an intelligent controller that automatically adjusts lighting and image acquisition in order to obtain the \"best\" possible composite view of the target for further image processing.","PeriodicalId":307318,"journal":{"name":"The 2nd Canadian Conference on Computer and Robot Vision (CRV'05)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123227887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Detection of occlusion edges from the derivatives of weather degraded images 天气退化图像导数的遮挡边缘检测
Pub Date : 2005-05-09 DOI: 10.1109/CRV.2005.35
Daniel Lévesque, F. Deschênes
Degradation of images of outdoor scenes caused by varying conditions of visibility can be exploited in order to get information on the scene. We propose two new methods for detecting occlusion edges between textured areas of a scene. These methods are based on the use of partial derivatives of two images acquired under different conditions of visibility. They were validated on images of both synthetic and real scenes.
利用能见度条件的变化引起的室外场景图像的退化,可以获得现场的信息。我们提出了两种检测场景纹理区域间遮挡边缘的新方法。这些方法是基于使用在不同能见度条件下获得的两幅图像的偏导数。他们在合成和真实场景的图像上进行了验证。
{"title":"Detection of occlusion edges from the derivatives of weather degraded images","authors":"Daniel Lévesque, F. Deschênes","doi":"10.1109/CRV.2005.35","DOIUrl":"https://doi.org/10.1109/CRV.2005.35","url":null,"abstract":"Degradation of images of outdoor scenes caused by varying conditions of visibility can be exploited in order to get information on the scene. We propose two new methods for detecting occlusion edges between textured areas of a scene. These methods are based on the use of partial derivatives of two images acquired under different conditions of visibility. They were validated on images of both synthetic and real scenes.","PeriodicalId":307318,"journal":{"name":"The 2nd Canadian Conference on Computer and Robot Vision (CRV'05)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121754010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Collaborative exploration for a group of self-interested robots 一组自利机器人的协同探索
Pub Date : 2005-05-09 DOI: 10.1109/CRV.2005.25
M. Schukat, Declan O'Beirne
This paper presents and new approach to robot exploration and mapping using a team of cooperative robots. This approach aims to exploit the increase in sensor data that multiple robots offer to improve efficiency, accuracy and detail in maps created, and also the lower cost in employing a group of inexpensive robots. The exploration technique involves covering an area as efficiently as possible while cooperating to estimate each other's positions and orientations. The ability to observe objects of interest from a number of viewpoints and combine this data means that cooperative robots can localize objects and estimate their shape in cluttered real world scenes. Robots in the system act as social agents, and are motivated to cooperate by a desire to increase their own utility. Within this society, robots form coalitions to complete tasks that arise which require input from multiple robots. The coalitions involve the adoption of certain roles or behaviors on the part of the different robots to carry out these tasks.
本文提出了一种利用协作机器人团队进行机器人探索和测绘的新方法。这种方法旨在利用多个机器人提供的传感器数据的增加,以提高创建地图的效率、准确性和细节,以及使用一组廉价机器人的较低成本。勘探技术包括尽可能有效地覆盖一个区域,同时合作估计彼此的位置和方向。从多个角度观察感兴趣的物体并结合这些数据的能力意味着协作机器人可以在混乱的现实世界场景中定位物体并估计其形状。系统中的机器人扮演着社会代理人的角色,它们的合作动机是想要增加自身的效用。在这个社会中,机器人组成联盟来完成需要多个机器人输入的任务。联盟涉及不同机器人采用特定的角色或行为来执行这些任务。
{"title":"Collaborative exploration for a group of self-interested robots","authors":"M. Schukat, Declan O'Beirne","doi":"10.1109/CRV.2005.25","DOIUrl":"https://doi.org/10.1109/CRV.2005.25","url":null,"abstract":"This paper presents and new approach to robot exploration and mapping using a team of cooperative robots. This approach aims to exploit the increase in sensor data that multiple robots offer to improve efficiency, accuracy and detail in maps created, and also the lower cost in employing a group of inexpensive robots. The exploration technique involves covering an area as efficiently as possible while cooperating to estimate each other's positions and orientations. The ability to observe objects of interest from a number of viewpoints and combine this data means that cooperative robots can localize objects and estimate their shape in cluttered real world scenes. Robots in the system act as social agents, and are motivated to cooperate by a desire to increase their own utility. Within this society, robots form coalitions to complete tasks that arise which require input from multiple robots. The coalitions involve the adoption of certain roles or behaviors on the part of the different robots to carry out these tasks.","PeriodicalId":307318,"journal":{"name":"The 2nd Canadian Conference on Computer and Robot Vision (CRV'05)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129580150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Video with ground-truth for validation of visual registration, tracking and navigation algorithms 视频与地面真实验证视觉注册,跟踪和导航算法
Pub Date : 2005-05-09 DOI: 10.1109/CRV.2005.86
R. Stolkin, A. Greig, J. Gilby
A fundamental task in computer vision is that of determining the position and orientation of a moving camera relative to an observed object or scene. Many such visual tracking algorithms have been proposed in the computer vision, artificial intelligence and robotics literature over the past 30 years. Predominantly, these remain un-validated since the ground-truth camera positions and orientations at each frame in a video sequence are not available for comparison with the outputs of the proposed vision systems. A method is presented for generating real visual test data with complete underlying ground-truth. The method enables the production of long video sequences, filmed along complicated six degree of freedom trajectories, featuring a variety of objects, in a variety of different visibility conditions, for which complete ground-truth data is known including the camera position and orientation at every image frame, intrinsic camera calibration data, a lens distortion model and models of the viewed objects.
计算机视觉的一项基本任务是确定移动摄像机相对于观察到的物体或场景的位置和方向。在过去的30年里,计算机视觉、人工智能和机器人文献中已经提出了许多这样的视觉跟踪算法。主要的是,这些仍然没有得到验证,因为视频序列中每帧的真实摄像机位置和方向无法与提议的视觉系统的输出进行比较。提出了一种生成具有完整底层真值的真实视觉测试数据的方法。该方法能够制作长视频序列,沿着复杂的六自由度轨迹拍摄,具有各种物体,在各种不同的可见性条件下,其中完整的地面真实数据已知,包括相机在每个图像帧的位置和方向,固有的相机校准数据,镜头失真模型和被观察物体的模型。
{"title":"Video with ground-truth for validation of visual registration, tracking and navigation algorithms","authors":"R. Stolkin, A. Greig, J. Gilby","doi":"10.1109/CRV.2005.86","DOIUrl":"https://doi.org/10.1109/CRV.2005.86","url":null,"abstract":"A fundamental task in computer vision is that of determining the position and orientation of a moving camera relative to an observed object or scene. Many such visual tracking algorithms have been proposed in the computer vision, artificial intelligence and robotics literature over the past 30 years. Predominantly, these remain un-validated since the ground-truth camera positions and orientations at each frame in a video sequence are not available for comparison with the outputs of the proposed vision systems. A method is presented for generating real visual test data with complete underlying ground-truth. The method enables the production of long video sequences, filmed along complicated six degree of freedom trajectories, featuring a variety of objects, in a variety of different visibility conditions, for which complete ground-truth data is known including the camera position and orientation at every image frame, intrinsic camera calibration data, a lens distortion model and models of the viewed objects.","PeriodicalId":307318,"journal":{"name":"The 2nd Canadian Conference on Computer and Robot Vision (CRV'05)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127117830","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Seeing the trees before the forest [natural object detection] 先见树后见林[自然物体探测]
Pub Date : 2005-05-09 DOI: 10.1109/CRV.2005.71
Daniel C. Asmar, J. Zelek, Samer M. Abdallah
In this paper, we propose an algorithm that detects and locates natural objects in an outdoor environment using local descriptors. Interest points inside images are detected with a difference of Gaussian (DoG) filter and are then represented using scale invariant local descriptors. Our algorithm learns objects in a weakly supervised manner by clustering similar descriptors together and using those clusters as object classifiers. The intent is to identify stable objects to be used as landmarks for simultaneous localization and mapping (SLAM) of robots. The robot milieu is first identified using a fast environment recognition algorithm and then landmarks are suggested for SLAM that are appropriate for that environment. In our experiments we test our theory on the detection of trees that belong to the plantae pinophyta (pine family). Initial results show that out of 200 test images, our classification yields 85 correct positives, 15 false negatives, 73 correct negatives and 27 false positives.
在本文中,我们提出了一种利用局部描述符检测和定位室外环境中自然物体的算法。用高斯差分(DoG)滤波检测图像内部的兴趣点,然后用尺度不变局部描述符表示。我们的算法通过将相似的描述符聚在一起并使用这些聚类作为对象分类器,以弱监督的方式学习对象。目的是识别稳定的物体,作为机器人同步定位和绘图(SLAM)的地标。首先使用快速环境识别算法识别机器人环境,然后为SLAM建议适合该环境的地标。在我们的实验中,我们测试了我们的理论对属于植物pinophyta(松科)的树木的检测。初步结果表明,在200个测试图像中,我们的分类产生85个正确阳性,15个假阴性,73个正确阴性和27个假阳性。
{"title":"Seeing the trees before the forest [natural object detection]","authors":"Daniel C. Asmar, J. Zelek, Samer M. Abdallah","doi":"10.1109/CRV.2005.71","DOIUrl":"https://doi.org/10.1109/CRV.2005.71","url":null,"abstract":"In this paper, we propose an algorithm that detects and locates natural objects in an outdoor environment using local descriptors. Interest points inside images are detected with a difference of Gaussian (DoG) filter and are then represented using scale invariant local descriptors. Our algorithm learns objects in a weakly supervised manner by clustering similar descriptors together and using those clusters as object classifiers. The intent is to identify stable objects to be used as landmarks for simultaneous localization and mapping (SLAM) of robots. The robot milieu is first identified using a fast environment recognition algorithm and then landmarks are suggested for SLAM that are appropriate for that environment. In our experiments we test our theory on the detection of trees that belong to the plantae pinophyta (pine family). Initial results show that out of 200 test images, our classification yields 85 correct positives, 15 false negatives, 73 correct negatives and 27 false positives.","PeriodicalId":307318,"journal":{"name":"The 2nd Canadian Conference on Computer and Robot Vision (CRV'05)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128406049","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
3D feature tracking using a dynamic structured light system 使用动态结构光系统进行3D特征跟踪
Pub Date : 2005-05-09 DOI: 10.1109/CRV.2005.1
A. Adán, F. Molina, A. Vázquez, Luis Morena
In this paper we present a 3D tracking system that is able to identify several characteristic points on the surface of a moving scene. The system is considered as a preliminary step in the interaction of industrial robots with dynamic scenes. Through a new color structured light technique based on a disordered codeword pattern, 3D coordinates of the object surface are extracted and processed. This method recovers 3D information in such a way that the correspondence problem is easily and robustly solved. After establishing a set of feature points, an inter frame window search algorithm is carried out to solve the tracking problem. A controlled experimental setup has been built in our lab composed by a 2 DOF mobile platform, a light structured sensor and a manipulator robot. The experimentation has been performed on medium spatial resolution and for soft movement specifications giving promising results.
在本文中,我们提出了一种能够识别运动场景表面上的几个特征点的三维跟踪系统。该系统被认为是工业机器人与动态场景交互的第一步。通过一种新的基于无序码字模式的彩色结构光技术,对物体表面的三维坐标进行提取和处理。该方法恢复了三维信息,使得对应问题易于鲁棒解决。在建立一组特征点后,采用帧间窗口搜索算法解决跟踪问题。在实验室建立了由2自由度移动平台、轻结构传感器和机械臂机器人组成的可控实验装置。在中等空间分辨率和软运动规格下进行了实验,取得了良好的效果。
{"title":"3D feature tracking using a dynamic structured light system","authors":"A. Adán, F. Molina, A. Vázquez, Luis Morena","doi":"10.1109/CRV.2005.1","DOIUrl":"https://doi.org/10.1109/CRV.2005.1","url":null,"abstract":"In this paper we present a 3D tracking system that is able to identify several characteristic points on the surface of a moving scene. The system is considered as a preliminary step in the interaction of industrial robots with dynamic scenes. Through a new color structured light technique based on a disordered codeword pattern, 3D coordinates of the object surface are extracted and processed. This method recovers 3D information in such a way that the correspondence problem is easily and robustly solved. After establishing a set of feature points, an inter frame window search algorithm is carried out to solve the tracking problem. A controlled experimental setup has been built in our lab composed by a 2 DOF mobile platform, a light structured sensor and a manipulator robot. The experimentation has been performed on medium spatial resolution and for soft movement specifications giving promising results.","PeriodicalId":307318,"journal":{"name":"The 2nd Canadian Conference on Computer and Robot Vision (CRV'05)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128838181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Photometric stereo via locality sensitive high-dimension hashing 通过局部敏感高维哈希的光度立体
Pub Date : 2005-05-09 DOI: 10.1109/CRV.2005.61
Lin Zhong, J. Little
In this paper, we extend the new photometric stereo method of Hertzmenn and Seitz that uses many images of an object together with a calibration object. For each point in the registered collection of images, we have a large number of brightness values. Photometric stereo finds a similar collection of brightness values from the calibration object and overdetermines the surface normal. With a large number of images, finding similar brightnesses becomes costly search in high dimensions. To speed up the search, we apply locality sensitive high dimensional hashing (LSH) to compute the irregular target object's surface orientation. The experimental results of a simplified photometric stereo experiment show consistent results in surface orientation. LSH can be implemented very efficiently and offers the possibility of practical photometric stereo computation with a large number of images.
在本文中,我们扩展了赫茨门和塞茨的新的光度立体方法,该方法将一个物体的多幅图像与一个标定物体一起使用。对于配准图像集合中的每个点,我们都有大量的亮度值。光度立体从校准对象中找到类似的亮度值集合,并过度确定表面法线。对于大量的图像,在高维上寻找相似亮度的搜索成本很高。为了提高搜索速度,我们采用局部敏感高维哈希(LSH)计算不规则目标物体的表面方向。简化的光度立体实验结果在表面取向上与实验结果一致。LSH可以非常有效地实现,并为大量图像的实际光度立体计算提供了可能。
{"title":"Photometric stereo via locality sensitive high-dimension hashing","authors":"Lin Zhong, J. Little","doi":"10.1109/CRV.2005.61","DOIUrl":"https://doi.org/10.1109/CRV.2005.61","url":null,"abstract":"In this paper, we extend the new photometric stereo method of Hertzmenn and Seitz that uses many images of an object together with a calibration object. For each point in the registered collection of images, we have a large number of brightness values. Photometric stereo finds a similar collection of brightness values from the calibration object and overdetermines the surface normal. With a large number of images, finding similar brightnesses becomes costly search in high dimensions. To speed up the search, we apply locality sensitive high dimensional hashing (LSH) to compute the irregular target object's surface orientation. The experimental results of a simplified photometric stereo experiment show consistent results in surface orientation. LSH can be implemented very efficiently and offers the possibility of practical photometric stereo computation with a large number of images.","PeriodicalId":307318,"journal":{"name":"The 2nd Canadian Conference on Computer and Robot Vision (CRV'05)","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115507909","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Video-based framework for face recognition in video 基于视频的人脸识别框架
Pub Date : 2005-05-09 DOI: 10.1109/CRV.2005.87
D. Gorodnichy
This paper presents a number of new views and techniques claimed to be very important for the problem of face recognition in video (FRiV). First, a clear differentiation is made between photographic facial data and video-acquired facial data as being two different modalities: one providing hard biometrics, the other providing softer biometrics. Second, faces which have the resolution of at least 12 pixels between the eyes are shown to be recognizable by computers just as they are by humans. As a way to deal with low resolution and quality of each individual video frame, the paper offers to use the neuro-associative principle employed by human brain, according to which both memorization and recognition of data are done based on a flow of frames rather than on one frame: synaptic plasticity provides a way to memorize from a sequence, while the collective decision making over time is very suitable for recognition of a sequence. As a benchmark for FRiV approaches, the paper introduces the IIT-NRC video-based database of faces which consists of pairs of low-resolution video clips of unconstrained facial motions. The recognition rate of over 95%, which we achieve on this database, as well as the results obtained on real-time annotation of people on TV allow us to believe that the proposed framework brings us closer to the ultimate benchmark for the FRiV approaches, which is "if you are able to recognize a person, so should the computer".
本文提出了一些新的观点和技术,声称对视频中的人脸识别问题非常重要。首先,照片面部数据和视频面部数据作为两种不同的模式进行了明确的区分:一种提供硬生物识别,另一种提供软生物识别。其次,两眼之间至少有12像素分辨率的人脸被证明可以被计算机识别,就像人类一样。为了解决单个视频帧的低分辨率和低质量问题,本文提出使用人脑的神经联想原理,根据该原理,数据的记忆和识别都是基于帧流而不是一帧完成的;突触可塑性提供了一种从序列中记忆的方法,而随着时间的推移的集体决策非常适合于序列的识别。作为FRiV方法的基准,本文介绍了基于IIT-NRC视频的人脸数据库,该数据库由对无约束面部运动的低分辨率视频片段组成。我们在这个数据库上达到95%以上的识别率,以及在电视上的人物实时注释上获得的结果,让我们相信所提出的框架使我们更接近FRiV方法的最终基准,即“如果你能够识别一个人,那么计算机也应该”。
{"title":"Video-based framework for face recognition in video","authors":"D. Gorodnichy","doi":"10.1109/CRV.2005.87","DOIUrl":"https://doi.org/10.1109/CRV.2005.87","url":null,"abstract":"This paper presents a number of new views and techniques claimed to be very important for the problem of face recognition in video (FRiV). First, a clear differentiation is made between photographic facial data and video-acquired facial data as being two different modalities: one providing hard biometrics, the other providing softer biometrics. Second, faces which have the resolution of at least 12 pixels between the eyes are shown to be recognizable by computers just as they are by humans. As a way to deal with low resolution and quality of each individual video frame, the paper offers to use the neuro-associative principle employed by human brain, according to which both memorization and recognition of data are done based on a flow of frames rather than on one frame: synaptic plasticity provides a way to memorize from a sequence, while the collective decision making over time is very suitable for recognition of a sequence. As a benchmark for FRiV approaches, the paper introduces the IIT-NRC video-based database of faces which consists of pairs of low-resolution video clips of unconstrained facial motions. The recognition rate of over 95%, which we achieve on this database, as well as the results obtained on real-time annotation of people on TV allow us to believe that the proposed framework brings us closer to the ultimate benchmark for the FRiV approaches, which is \"if you are able to recognize a person, so should the computer\".","PeriodicalId":307318,"journal":{"name":"The 2nd Canadian Conference on Computer and Robot Vision (CRV'05)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115730460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 120
期刊
The 2nd Canadian Conference on Computer and Robot Vision (CRV'05)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1