首页 > 最新文献

2012 Ninth Conference on Computer and Robot Vision最新文献

英文 中文
Non-Lambertian Model-based Facial Shape Recovery from Single Image Under Unknown General Illumination 未知光照下基于非lambertian模型的单幅图像人脸形状恢复
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.40
S. Elhabian, Eslam A. Mostafa, H. Rara, A. Farag
Through depth perception, humans have the ability to determine distances based on a single 2D image projected on their retina, where shape-from-shading (SFS) provides a mean to mimic such a phenomenon. The goal of this paper is to recover 3D facial shape from a single image of unknown general illumination, while relaxing the non-realistic assumption of Lambert Ian reflectance. Prior shape, albedo and reflectance models from real data, which are metric in nature, are incorporated into the shape recovery framework. Adopting a frequency-space based representation of the image irradiance equation, we propose an appearance model, termed as Harmonic Projection Images, which accounts explicitly for different human skin types as well as complex illumination conditions. Assuming skin reflectance obeys Torrance-Sparrow model, we prove analytically that it can be represented by at most 5th order harmonic basis whose closed form is provided. The recovery framework is a non-iterative approach which incorporates regression-like algorithm in the minimization process. Our experiments on synthetic and real images illustrate the robustness of our appearance model vis-a-vis illumination variation.
通过深度感知,人类有能力根据投射在视网膜上的单个2D图像来确定距离,其中形状-阴影(SFS)提供了一种模拟这种现象的方法。本文的目标是从未知一般照明的单幅图像中恢复三维面部形状,同时放松兰伯特伊恩反射率的非现实假设。将实际数据的先验形状、反照率和反射率模型(本质上是度量的)纳入形状恢复框架。采用基于频率空间的图像辐照度方程表示,我们提出了一种称为谐波投影图像的外观模型,该模型明确地考虑了不同的人体皮肤类型以及复杂的照明条件。假设皮肤反射率服从Torrance-Sparrow模型,我们解析地证明了它最多可以用5阶调和基表示,并给出了其闭合形式。恢复框架是一种非迭代方法,在最小化过程中引入了类回归算法。我们在合成和真实图像上的实验说明了我们的外观模型相对于光照变化的鲁棒性。
{"title":"Non-Lambertian Model-based Facial Shape Recovery from Single Image Under Unknown General Illumination","authors":"S. Elhabian, Eslam A. Mostafa, H. Rara, A. Farag","doi":"10.1109/CRV.2012.40","DOIUrl":"https://doi.org/10.1109/CRV.2012.40","url":null,"abstract":"Through depth perception, humans have the ability to determine distances based on a single 2D image projected on their retina, where shape-from-shading (SFS) provides a mean to mimic such a phenomenon. The goal of this paper is to recover 3D facial shape from a single image of unknown general illumination, while relaxing the non-realistic assumption of Lambert Ian reflectance. Prior shape, albedo and reflectance models from real data, which are metric in nature, are incorporated into the shape recovery framework. Adopting a frequency-space based representation of the image irradiance equation, we propose an appearance model, termed as Harmonic Projection Images, which accounts explicitly for different human skin types as well as complex illumination conditions. Assuming skin reflectance obeys Torrance-Sparrow model, we prove analytically that it can be represented by at most 5th order harmonic basis whose closed form is provided. The recovery framework is a non-iterative approach which incorporates regression-like algorithm in the minimization process. Our experiments on synthetic and real images illustrate the robustness of our appearance model vis-a-vis illumination variation.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"285 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123034288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Extracting High-Level Intuitive Features (HLIF) for Classifying Skin Lesions Using Standard Camera Images 利用标准相机图像提取高级直观特征(HLIF)进行皮肤病变分类
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.59
R. Amelard, A. Wong, David A Clausi
High-level intuitive features (HLIF) that measure asymmetry of skin lesion images obtained using standard cameras are presented. These features can be used to help dermatologists objectively diagnose lesions as cancerous (melanoma) or benign with intuitive rationale. Existing work defines large sets of low-level statistical features for analysing skin lesions. The proposed HLIFs are designed such that smaller sets of HLIFs can capture more deterministic information than large sets of low-level features. Analytical reasoning is given for each feature to show how it aptly describes asymmetry. Promising experimental results show that classification using the proposed HLIF set, although only one-tenth the size of the existing state-of-the-art low-level feature set, labels the data with better or comparable success. The best classification is obtained by combining the low-level feature set with the HLIF set.
提出了用标准相机测量皮肤病变图像不对称性的高级直观特征(HLIF)。这些特征可以用来帮助皮肤科医生客观地诊断病变是癌性(黑色素瘤)还是良性的直觉原理。现有的工作定义了大量用于分析皮肤病变的低水平统计特征。所提出的HLIFs的设计使得较小的HLIFs集可以捕获比大的低级特征集更多的确定性信息。给出了每个特征的分析推理,以说明它如何恰当地描述不对称。有希望的实验结果表明,使用提出的HLIF集进行分类,尽管只有现有最先进的低级特征集的十分之一大小,但标记数据的效果更好或相当成功。将低级特征集与HLIF集相结合,得到最佳分类。
{"title":"Extracting High-Level Intuitive Features (HLIF) for Classifying Skin Lesions Using Standard Camera Images","authors":"R. Amelard, A. Wong, David A Clausi","doi":"10.1109/CRV.2012.59","DOIUrl":"https://doi.org/10.1109/CRV.2012.59","url":null,"abstract":"High-level intuitive features (HLIF) that measure asymmetry of skin lesion images obtained using standard cameras are presented. These features can be used to help dermatologists objectively diagnose lesions as cancerous (melanoma) or benign with intuitive rationale. Existing work defines large sets of low-level statistical features for analysing skin lesions. The proposed HLIFs are designed such that smaller sets of HLIFs can capture more deterministic information than large sets of low-level features. Analytical reasoning is given for each feature to show how it aptly describes asymmetry. Promising experimental results show that classification using the proposed HLIF set, although only one-tenth the size of the existing state-of-the-art low-level feature set, labels the data with better or comparable success. The best classification is obtained by combining the low-level feature set with the HLIF set.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121048852","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Multi-Robot Repeated Area Coverage: Performance Optimization Under Various Visual Ranges 多机器人重复区域覆盖:不同视距下的性能优化
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.46
Pooyan Fazli, Alireza Davoodi, Alan K. Mackworth
We address the problem of repeated coverage of a target area, of any polygonal shape, by a team of robots having a limited visual range. Three distributed Cluster-based algorithms, and a method called Cyclic Coverage are introduced for the problem. The goal is to evaluate the performance of the repeated coverage algorithms under the effect of changes in the robots' visual range. A comprehensive set of performance metrics are considered, including the distance the robots travel, the frequency of visiting points in the target area, and the degree of balance in workload distribution among the robots. The Cyclic Coverage approach, used as a benchmark to compare the algorithms, produces optimal or near-optimal solutions for the single robot case under some criteria. The results show that the identity of the optimal repeated coverage algorithm depends on the metric and the robots' visual range.
我们解决的问题,重复覆盖的目标区域,任何多边形形状,由一队机器人具有有限的视觉范围。针对该问题,介绍了三种分布式聚类算法和一种循环覆盖方法。目的是评估重复覆盖算法在机器人视觉范围变化影响下的性能。考虑了一套综合的性能指标,包括机器人移动的距离、目标区域访问点的频率以及机器人之间工作负载分配的平衡程度。循环覆盖方法作为比较算法的基准,在某些标准下对单个机器人情况产生最优或近最优解。结果表明,最优重复覆盖算法的一致性取决于度量和机器人的视觉距离。
{"title":"Multi-Robot Repeated Area Coverage: Performance Optimization Under Various Visual Ranges","authors":"Pooyan Fazli, Alireza Davoodi, Alan K. Mackworth","doi":"10.1109/CRV.2012.46","DOIUrl":"https://doi.org/10.1109/CRV.2012.46","url":null,"abstract":"We address the problem of repeated coverage of a target area, of any polygonal shape, by a team of robots having a limited visual range. Three distributed Cluster-based algorithms, and a method called Cyclic Coverage are introduced for the problem. The goal is to evaluate the performance of the repeated coverage algorithms under the effect of changes in the robots' visual range. A comprehensive set of performance metrics are considered, including the distance the robots travel, the frequency of visiting points in the target area, and the degree of balance in workload distribution among the robots. The Cyclic Coverage approach, used as a benchmark to compare the algorithms, produces optimal or near-optimal solutions for the single robot case under some criteria. The results show that the identity of the optimal repeated coverage algorithm depends on the metric and the robots' visual range.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128672603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
In Situ Motion Capture of Speed Skating: Escaping the Treadmill 速度滑冰的现场动作捕捉:逃离跑步机
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.68
J. Boyd, Andrew Godbout, C. Thornton
The advent of the Kinect depth imager has opened the door to motion capture applications that would have been much more costly with previous technologies. In part, the Kinect achieves this by focusing on a very specific application domain, thus narrowing the requirement for the motion capture system. Specifically, Kinect motion capture works best within a small physical space while the camera is stationary. We seek to extend Kinect motion capture for use in athletic training - speed skating in particular - by placing the Kinect on a mobile, robotic platform to capture motion in situ. Athletes move over large distances, so the mobile platform addresses the limited viewing area of the Kinect. As the platform moves, we must also account for the now dynamic background against which the athlete performs. The result is a novel, visually-guided robotic platform that follows athletes, allowing us to capture motion and images that would not be possible with a treadmill. We describe the system in detail and give examples of the system capturing the motion of a speed skater at typical training speeds.
Kinect深度成像仪的出现为动作捕捉应用打开了大门,而使用以前的技术,动作捕捉应用的成本要高得多。在某种程度上,Kinect通过专注于一个非常特定的应用领域来实现这一目标,从而缩小了对动作捕捉系统的要求。具体来说,当摄像机静止时,Kinect动作捕捉在一个小的物理空间内效果最好。我们试图将Kinect动作捕捉扩展到运动训练中——特别是速度滑冰——通过将Kinect放置在移动的机器人平台上来就地捕捉动作。运动员要移动很远的距离,所以移动平台解决了Kinect有限的观看区域。随着平台的移动,我们还必须考虑运动员表演时所处的动态背景。结果是一个新颖的、视觉引导的机器人平台,它可以跟随运动员,让我们捕捉到运动和图像,这在跑步机上是不可能实现的。我们详细描述了该系统,并给出了该系统在典型训练速度下捕捉速滑运动员运动的例子。
{"title":"In Situ Motion Capture of Speed Skating: Escaping the Treadmill","authors":"J. Boyd, Andrew Godbout, C. Thornton","doi":"10.1109/CRV.2012.68","DOIUrl":"https://doi.org/10.1109/CRV.2012.68","url":null,"abstract":"The advent of the Kinect depth imager has opened the door to motion capture applications that would have been much more costly with previous technologies. In part, the Kinect achieves this by focusing on a very specific application domain, thus narrowing the requirement for the motion capture system. Specifically, Kinect motion capture works best within a small physical space while the camera is stationary. We seek to extend Kinect motion capture for use in athletic training - speed skating in particular - by placing the Kinect on a mobile, robotic platform to capture motion in situ. Athletes move over large distances, so the mobile platform addresses the limited viewing area of the Kinect. As the platform moves, we must also account for the now dynamic background against which the athlete performs. The result is a novel, visually-guided robotic platform that follows athletes, allowing us to capture motion and images that would not be possible with a treadmill. We describe the system in detail and give examples of the system capturing the motion of a speed skater at typical training speeds.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"583 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134118009","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Adaptive RGB-D Localization 自适应RGB-D定位
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.11
M. Paton, J. Kosecka
The advent of RGB-D cameras which provide synchronized range and video data creates new opportunities for exploiting both sensing modalities for various robotic applications. This paper exploits the strengths of vision and range measurements and develops a novel robust algorithm for localization using RGB-D cameras. We show how correspondences established by matching visual SIFT features can effectively initialize the generalized ICP algorithm as well as demonstrate situations where such initialization is not viable. We propose an adaptive architecture which computes the pose estimate from the most reliable measurements in a given environment and present thorough evaluation of the resulting algorithm against a dataset of RGB-D benchmarks, demonstrating superior or comparable performance in the absence of the global optimization stage. Lastly we demonstrate the proposed algorithm on a challenging indoor dataset and demonstrate improvements where pose estimation from either pure range sensing or vision techniques perform poorly.
提供同步距离和视频数据的RGB-D摄像机的出现为开发各种机器人应用的传感模式创造了新的机会。本文利用视觉和距离测量的优势,开发了一种新的鲁棒RGB-D相机定位算法。我们展示了通过匹配视觉SIFT特征建立的对应关系如何有效地初始化广义ICP算法,并演示了这种初始化不可行的情况。我们提出了一种自适应架构,该架构根据给定环境中最可靠的测量结果计算姿态估计,并针对RGB-D基准数据集对结果算法进行全面评估,在没有全局优化阶段的情况下展示优越或可比的性能。最后,我们在一个具有挑战性的室内数据集上演示了所提出的算法,并演示了纯距离传感或视觉技术的姿态估计性能较差的改进。
{"title":"Adaptive RGB-D Localization","authors":"M. Paton, J. Kosecka","doi":"10.1109/CRV.2012.11","DOIUrl":"https://doi.org/10.1109/CRV.2012.11","url":null,"abstract":"The advent of RGB-D cameras which provide synchronized range and video data creates new opportunities for exploiting both sensing modalities for various robotic applications. This paper exploits the strengths of vision and range measurements and develops a novel robust algorithm for localization using RGB-D cameras. We show how correspondences established by matching visual SIFT features can effectively initialize the generalized ICP algorithm as well as demonstrate situations where such initialization is not viable. We propose an adaptive architecture which computes the pose estimate from the most reliable measurements in a given environment and present thorough evaluation of the resulting algorithm against a dataset of RGB-D benchmarks, demonstrating superior or comparable performance in the absence of the global optimization stage. Lastly we demonstrate the proposed algorithm on a challenging indoor dataset and demonstrate improvements where pose estimation from either pure range sensing or vision techniques perform poorly.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134619674","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
3D Registration for Verification of Humanoid Justin's Upper Body Kinematics 验证仿人Justin上半身运动学的3D注册
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.43
Nadia Figueroa, Haider Ali, Florian Schmidt
Humanoid robots such as DLR's Justin are built with light-weight structures and flexible mechanical components. These generate positioning errors at the TCP (Tool-Center-Point) end-pose of the hand. The identification of these errors is essential for object manipulation and path planning. We proposed a verification routine to identify the bounds of the TCP end-pose errors by using the on-board stereo vision system. It involves estimating the pose of 3D point clouds of Justin's hand by using state-of-the-art 3D registration techniques. Partial models of the hand were generated by registering subsets of overlapping 3D point clouds. We proposed a method for the selection of overlapping point clouds of self-occluding objects (Justin's hand). It is based on a statistical analysis of the depth values. We applied an extended metaview registration method to the resulting subset of point clouds. The partial models were evaluated with detailed based surface consistency measures. The TCP end-pose errors estimated by using our method are consistent with ground-truth errors.
像DLR的Justin这样的人形机器人是由轻质结构和灵活的机械部件制成的。这些会在手的TCP(工具中心点)末端位姿产生定位错误。这些误差的识别对于物体操作和路径规划至关重要。提出了一种利用车载立体视觉系统识别TCP端位姿误差边界的验证程序。它涉及到估计贾斯汀的手的3D点云的姿势,使用最先进的3D注册技术。手的部分模型是通过注册重叠的3D点云子集生成的。提出了一种自遮挡物体(Justin’s hand)重叠点云的选择方法。它是基于深度值的统计分析。我们将扩展的元视图配准方法应用于得到的点云子集。对部分模型进行了基于表面一致性测量的详细评估。该方法估计的TCP端位姿误差与真值误差基本一致。
{"title":"3D Registration for Verification of Humanoid Justin's Upper Body Kinematics","authors":"Nadia Figueroa, Haider Ali, Florian Schmidt","doi":"10.1109/CRV.2012.43","DOIUrl":"https://doi.org/10.1109/CRV.2012.43","url":null,"abstract":"Humanoid robots such as DLR's Justin are built with light-weight structures and flexible mechanical components. These generate positioning errors at the TCP (Tool-Center-Point) end-pose of the hand. The identification of these errors is essential for object manipulation and path planning. We proposed a verification routine to identify the bounds of the TCP end-pose errors by using the on-board stereo vision system. It involves estimating the pose of 3D point clouds of Justin's hand by using state-of-the-art 3D registration techniques. Partial models of the hand were generated by registering subsets of overlapping 3D point clouds. We proposed a method for the selection of overlapping point clouds of self-occluding objects (Justin's hand). It is based on a statistical analysis of the depth values. We applied an extended metaview registration method to the resulting subset of point clouds. The partial models were evaluated with detailed based surface consistency measures. The TCP end-pose errors estimated by using our method are consistent with ground-truth errors.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134495623","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Dynamic Weighting of Facial Features for Automatic Pose-Invariant Face Recognition 自动姿态不变人脸识别中人脸特征的动态加权
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.61
Eslam A. Mostafa, A. Farag
This paper proposes an automatic pose-invariant face recognition system. In our approach, we consider the texture information around the facial features to compute the similarity measure between the probe and gallery images. The weight of each facial feature is dynamically estimated based on its robustness to the pose of the captured image. An approach to extract the 9 facial features used to initialize the Active shape model is proposed. The approach is not dependent on the texture around the facial feature only but incorporates the information obtained about the facial feature relations. Our face recognition system is tested on common datasets in pose evaluation CMU-PIE and FERET. The results show out-performance of the state of the art automatic face recognition systems.
提出了一种自动姿态不变人脸识别系统。在我们的方法中,我们考虑面部特征周围的纹理信息来计算探针和画廊图像之间的相似性度量。基于每个面部特征对捕获图像姿态的鲁棒性,动态估计其权重。提出了一种用于初始化主动形状模型的9个面部特征的提取方法。该方法不仅依赖于面部特征周围的纹理,而且结合了面部特征关系的信息。我们的人脸识别系统在姿态评估、CMU-PIE和FERET等常用数据集上进行了测试。结果表明,该方法优于目前最先进的自动人脸识别系统。
{"title":"Dynamic Weighting of Facial Features for Automatic Pose-Invariant Face Recognition","authors":"Eslam A. Mostafa, A. Farag","doi":"10.1109/CRV.2012.61","DOIUrl":"https://doi.org/10.1109/CRV.2012.61","url":null,"abstract":"This paper proposes an automatic pose-invariant face recognition system. In our approach, we consider the texture information around the facial features to compute the similarity measure between the probe and gallery images. The weight of each facial feature is dynamically estimated based on its robustness to the pose of the captured image. An approach to extract the 9 facial features used to initialize the Active shape model is proposed. The approach is not dependent on the texture around the facial feature only but incorporates the information obtained about the facial feature relations. Our face recognition system is tested on common datasets in pose evaluation CMU-PIE and FERET. The results show out-performance of the state of the art automatic face recognition systems.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"215 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121201558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Optical Flow at Occlusion 遮挡光流
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.34
Jieyu Zhang, J. Barron
We implement and quantitatively/qualitatively evaluate two optical flow methods that model occlusion. The Yuan et al. method [1] improves on the Horn and Schunck optical flow method at occlusion boundaries by using a dynamic coefficient (the Lagrange multiplier α) at each pixel that weighs the smoothness constraint relative to the optical flow constraint, by adopting a modified scheme to calculate average velocities and by using a “compensating” iterative algorithm to achieve higher computational efficiency. The Niu et al. method [2] is based on a modified version of the Lucas and Kanade optical flow method, that selects local intensity neighbourhoods, spatially and temporally, based on pixels that are on different sides of an occlusion boundary and then corrects any erroneous flow computed at occlusion boundaries. We present quantitative results for sinusoidal sequence with a known occlusion boundary. We also present qualitative evaluation of the methods on the Hamburg Taxi sequence and and the Trees sequence.
我们实现和定量/定性评估两种光流方法建模遮挡。Yuan等人的方法[1]改进了Horn和Schunck遮挡边界处的光流方法,在每个像素处使用动态系数(拉格朗日乘子α)来衡量相对于光流约束的平滑性约束,采用改进的方案来计算平均速度,并使用“补偿”迭代算法来获得更高的计算效率。Niu等人的方法[2]是基于Lucas和Kanade光流方法的改进版本,该方法基于位于遮挡边界不同侧面的像素,在空间和时间上选择局部强度邻域,然后校正在遮挡边界计算的任何错误流。我们给出了具有已知遮挡边界的正弦序列的定量结果。我们还对汉堡出租车序列和树序列的方法进行了定性评价。
{"title":"Optical Flow at Occlusion","authors":"Jieyu Zhang, J. Barron","doi":"10.1109/CRV.2012.34","DOIUrl":"https://doi.org/10.1109/CRV.2012.34","url":null,"abstract":"We implement and quantitatively/qualitatively evaluate two optical flow methods that model occlusion. The Yuan et al. method [1] improves on the Horn and Schunck optical flow method at occlusion boundaries by using a dynamic coefficient (the Lagrange multiplier α) at each pixel that weighs the smoothness constraint relative to the optical flow constraint, by adopting a modified scheme to calculate average velocities and by using a “compensating” iterative algorithm to achieve higher computational efficiency. The Niu et al. method [2] is based on a modified version of the Lucas and Kanade optical flow method, that selects local intensity neighbourhoods, spatially and temporally, based on pixels that are on different sides of an occlusion boundary and then corrects any erroneous flow computed at occlusion boundaries. We present quantitative results for sinusoidal sequence with a known occlusion boundary. We also present qualitative evaluation of the methods on the Hamburg Taxi sequence and and the Trees sequence.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129486840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Multi-Scale Saliency-Guided Compressive Sensing Approach to Efficient Robotic Laser Range Measurements 机器人激光距离有效测量的多尺度显著性导向压缩感知方法
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.8
S. Schwartz, A. Wong, David A Clausi
Improving laser range data acquisition speed is important for many robotic applications such as mapping and localization. One approach to reducing acquisition time is to acquire laser range data through a dynamically small subset of measurement locations. The reconstruction can then be performed based on the concept of compressed sensing (CS), where a sparse signal representation allows for signal reconstruction at sub-Nyquist measurements. Motivated by this, a novel multi-scale saliency-guided CS-based algorithm is proposed for an efficient robotic laser range data acquisition for robotic vision. The proposed system samples the objects of interest through an optimized probability density function derived based on multi-scale saliency rather than the uniform random distribution used in traditional CS systems. Experimental results with laser range data from indoor and outdoor environments show that the proposed approach requires less than half the samples needed by existing CS-based approaches while maintaining the same reconstruction performance. In addition, the proposed method offers significant improvement in reconstruction SNR compared to current CS-based approaches.
提高激光距离数据采集速度对许多机器人应用,如测绘和定位是重要的。减少采集时间的一种方法是通过测量位置的动态小子集来获取激光距离数据。然后可以基于压缩感知(CS)的概念进行重建,其中稀疏的信号表示允许在亚奈奎斯特测量下进行信号重建。基于此,提出了一种基于多尺度显著性引导的基于cs的机器人激光距离数据高效采集算法。该系统通过基于多尺度显著性的优化概率密度函数对感兴趣的对象进行采样,而不是传统CS系统中使用的均匀随机分布。室内和室外环境的激光距离数据实验结果表明,该方法在保持相同重建性能的同时,所需样本数量不到现有基于cs的方法的一半。此外,与目前基于cs的方法相比,该方法在重建信噪比方面有显著提高。
{"title":"Multi-Scale Saliency-Guided Compressive Sensing Approach to Efficient Robotic Laser Range Measurements","authors":"S. Schwartz, A. Wong, David A Clausi","doi":"10.1109/CRV.2012.8","DOIUrl":"https://doi.org/10.1109/CRV.2012.8","url":null,"abstract":"Improving laser range data acquisition speed is important for many robotic applications such as mapping and localization. One approach to reducing acquisition time is to acquire laser range data through a dynamically small subset of measurement locations. The reconstruction can then be performed based on the concept of compressed sensing (CS), where a sparse signal representation allows for signal reconstruction at sub-Nyquist measurements. Motivated by this, a novel multi-scale saliency-guided CS-based algorithm is proposed for an efficient robotic laser range data acquisition for robotic vision. The proposed system samples the objects of interest through an optimized probability density function derived based on multi-scale saliency rather than the uniform random distribution used in traditional CS systems. Experimental results with laser range data from indoor and outdoor environments show that the proposed approach requires less than half the samples needed by existing CS-based approaches while maintaining the same reconstruction performance. In addition, the proposed method offers significant improvement in reconstruction SNR compared to current CS-based approaches.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131135361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Improved Edge Representation via Early Recurrent Inhibition 通过早期复发性抑制改善边缘表征
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.13
Xun Shi, John K. Tsotsos
This paper describes a biologically motivated computational model, termed as early recurrent inhibition, to improve edge representation. The computation borrows the idea from the primate visual system that visual features are calculated in the two main visual pathways with different speeds and thus one can positively affect the other via early recurrent mechanisms. Based on the collected results, we conclude such a recurrent processing from area MT to the ventral layers of the primary visual area (V1) may be at play, and hypothesize that one effect of this recurrent mechanism is that V1 responses to high-spatial frequency edges are suppressed by signals sent from MT, leading to a cleaner edge representation. The operation is modeled as a weighted multiplicative inhibition process. Depending on the weighting methods, two types of inhibition are investigated, namely isotropic and anisotropic inhibition. To evaluate the inhibited edge representation, our model is attached to a contour operator to generate binary contour maps. Using real images, we quantitatively compared contours calculated by our work with those by a well-known biologically motivated model. Results clearly demonstrate that early recurrent inhibition has a positive and consistent influence on edge detection.
本文描述了一种生物驱动的计算模型,称为早期复发抑制,以改善边缘表示。这种计算借鉴了灵长类视觉系统的思想,即视觉特征是在两个主要的视觉路径中以不同的速度计算的,因此一个可以通过早期循环机制积极影响另一个。根据收集到的结果,我们得出结论,从MT区域到主要视觉区域(V1)的腹侧层的这种循环处理可能起作用,并假设这种循环机制的一个影响是V1对高空间频率边缘的响应被MT发送的信号抑制,从而导致更清晰的边缘表示。该操作被建模为加权乘法抑制过程。根据加权方法,研究了两种类型的抑制,即各向同性和各向异性抑制。为了评估被抑制的边缘表示,我们的模型附加了一个轮廓算子来生成二进制轮廓图。使用真实的图像,我们定量地比较了我们的工作计算出的轮廓与那些众所周知的生物动机模型。结果清楚地表明,早期复发性抑制对边缘检测具有积极和一致的影响。
{"title":"Improved Edge Representation via Early Recurrent Inhibition","authors":"Xun Shi, John K. Tsotsos","doi":"10.1109/CRV.2012.13","DOIUrl":"https://doi.org/10.1109/CRV.2012.13","url":null,"abstract":"This paper describes a biologically motivated computational model, termed as early recurrent inhibition, to improve edge representation. The computation borrows the idea from the primate visual system that visual features are calculated in the two main visual pathways with different speeds and thus one can positively affect the other via early recurrent mechanisms. Based on the collected results, we conclude such a recurrent processing from area MT to the ventral layers of the primary visual area (V1) may be at play, and hypothesize that one effect of this recurrent mechanism is that V1 responses to high-spatial frequency edges are suppressed by signals sent from MT, leading to a cleaner edge representation. The operation is modeled as a weighted multiplicative inhibition process. Depending on the weighting methods, two types of inhibition are investigated, namely isotropic and anisotropic inhibition. To evaluate the inhibited edge representation, our model is attached to a contour operator to generate binary contour maps. Using real images, we quantitatively compared contours calculated by our work with those by a well-known biologically motivated model. Results clearly demonstrate that early recurrent inhibition has a positive and consistent influence on edge detection.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132762135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2012 Ninth Conference on Computer and Robot Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1