首页 > 最新文献

The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)最新文献

英文 中文
Object detection and tracking using iterative division and correlograms 使用迭代分割和相关图的目标检测和跟踪
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.52
R. Bourezak, Guillaume-Alexandre Bilodeau
This paper presents algorithms for detecting moving objects, tracking them and interpreting their relationships. The algorithms are based on color and texture analysis in HSV space for real time processing. Our goal is to study human interaction by tracking people and objects for surveillance applications. The object detection algorithm is based on color histograms and iteratively divided interest regions for motion detection. The tracking algorithm is based on correlograms which combine spectral and spatial information to match detected objects in consecutive frames.
本文提出了检测运动物体、跟踪运动物体和解释运动物体之间关系的算法。该算法基于HSV空间的颜色和纹理分析进行实时处理。我们的目标是通过跟踪监视应用中的人和物体来研究人类互动。目标检测算法基于颜色直方图和迭代划分的兴趣区域进行运动检测。该算法基于相关图,结合光谱信息和空间信息来匹配连续帧中的检测目标。
{"title":"Object detection and tracking using iterative division and correlograms","authors":"R. Bourezak, Guillaume-Alexandre Bilodeau","doi":"10.1109/CRV.2006.52","DOIUrl":"https://doi.org/10.1109/CRV.2006.52","url":null,"abstract":"This paper presents algorithms for detecting moving objects, tracking them and interpreting their relationships. The algorithms are based on color and texture analysis in HSV space for real time processing. Our goal is to study human interaction by tracking people and objects for surveillance applications. The object detection algorithm is based on color histograms and iteratively divided interest regions for motion detection. The tracking algorithm is based on correlograms which combine spectral and spatial information to match detected objects in consecutive frames.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115345616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
3D Terrain Modeling for Rover Localization and Navigation 面向漫游车定位与导航的三维地形建模
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.2
J. Bakambu, P. Allard, E. Dupuis
This paper presents the problem of constructing a 3D terrain model for localization and navigation of planetary rover. We presented our approach to 3D terrain reconstruction from large sparse range data sets. In space robotics applications, an accurate and up-to-date model of the environment is very important for a variety of reasons. In particular, the model can be used for safe tele-operation, path planning and mapping points of interest. We propose an on-line terrain modeling using data provided by an on-board high resolution, accurate, 3D range sensor. Our approach is based on on-line acquisition of range scans from different view-points with overlapping regions, merge them together into a single point cloud, and then fit an irregular triangular mesh on the merged data. The outdoor experimental results demonstrate the effectiveness of the reconstructed terrain model for rover localization, path planning and motion execution scenario.
提出了行星漫游者定位导航所需的三维地形模型的构建问题。我们提出了一种基于大范围稀疏数据集的三维地形重建方法。在空间机器人应用中,由于各种原因,准确和最新的环境模型非常重要。特别是,该模型可用于安全远程操作,路径规划和绘制兴趣点。我们提出了一个在线地形建模利用数据提供的机载高分辨率,精确的,三维距离传感器。我们的方法是基于在线获取具有重叠区域的不同视点的距离扫描,将它们合并成单个点云,然后在合并的数据上拟合不规则三角形网格。室外实验结果验证了重构地形模型在漫游车定位、路径规划和运动执行场景中的有效性。
{"title":"3D Terrain Modeling for Rover Localization and Navigation","authors":"J. Bakambu, P. Allard, E. Dupuis","doi":"10.1109/CRV.2006.2","DOIUrl":"https://doi.org/10.1109/CRV.2006.2","url":null,"abstract":"This paper presents the problem of constructing a 3D terrain model for localization and navigation of planetary rover. We presented our approach to 3D terrain reconstruction from large sparse range data sets. In space robotics applications, an accurate and up-to-date model of the environment is very important for a variety of reasons. In particular, the model can be used for safe tele-operation, path planning and mapping points of interest. We propose an on-line terrain modeling using data provided by an on-board high resolution, accurate, 3D range sensor. Our approach is based on on-line acquisition of range scans from different view-points with overlapping regions, merge them together into a single point cloud, and then fit an irregular triangular mesh on the merged data. The outdoor experimental results demonstrate the effectiveness of the reconstructed terrain model for rover localization, path planning and motion execution scenario.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"93 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121121998","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Design and analysis of a framework for real-time vision-based SLAM using Rao-Blackwellised particle filters 基于rao - blackwell化粒子滤波的实时视觉SLAM框架设计与分析
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.25
Robert Sim, P. Elinas, Matt Griffin, Alex Shyr, J. Little
This paper addresses the problem of simultaneous localization and mapping (SLAM) using vision-based sensing. We present and analyse an implementation of a Rao- Blackwellised particle filter (RBPF) that uses stereo vision to localize a camera and 3D landmarks as the camera moves through an unknown environment. Our implementation is robust, can operate in real-time, and can operate without odometric or inertial measurements. Furthermore, our approach supports a 6-degree-of-freedom pose representation, vision-based ego-motion estimation, adaptive resampling, monocular operation, and a selection of odometry-based, observation-based, and mixture (combining local and global pose estimation) proposal distributions. This paper also examines the run-time behavior of efficiently designed RBPFs, providing an extensive empirical analysis of the memory and processing characteristics of RBPFs for vision-based SLAM. Finally, we present experimental results demonstrating the accuracy and efficiency of our approach.
本文研究了基于视觉感知的同时定位与制图问题。我们提出并分析了Rao- blackwell化粒子滤波器(RBPF)的实现,该滤波器使用立体视觉来定位相机和3D地标,当相机在未知环境中移动时。我们的实现是鲁棒的,可以实时操作,并且可以在没有里程或惯性测量的情况下操作。此外,我们的方法支持6个自由度的姿态表示、基于视觉的自我运动估计、自适应重采样、单目操作,以及基于里程计、基于观测和混合(结合局部和全局姿态估计)的建议分布选择。本文还研究了高效设计的rbpf的运行时行为,对基于视觉SLAM的rbpf的记忆和处理特征进行了广泛的实证分析。最后,我们给出了实验结果,证明了我们的方法的准确性和效率。
{"title":"Design and analysis of a framework for real-time vision-based SLAM using Rao-Blackwellised particle filters","authors":"Robert Sim, P. Elinas, Matt Griffin, Alex Shyr, J. Little","doi":"10.1109/CRV.2006.25","DOIUrl":"https://doi.org/10.1109/CRV.2006.25","url":null,"abstract":"This paper addresses the problem of simultaneous localization and mapping (SLAM) using vision-based sensing. We present and analyse an implementation of a Rao- Blackwellised particle filter (RBPF) that uses stereo vision to localize a camera and 3D landmarks as the camera moves through an unknown environment. Our implementation is robust, can operate in real-time, and can operate without odometric or inertial measurements. Furthermore, our approach supports a 6-degree-of-freedom pose representation, vision-based ego-motion estimation, adaptive resampling, monocular operation, and a selection of odometry-based, observation-based, and mixture (combining local and global pose estimation) proposal distributions. This paper also examines the run-time behavior of efficiently designed RBPFs, providing an extensive empirical analysis of the memory and processing characteristics of RBPFs for vision-based SLAM. Finally, we present experimental results demonstrating the accuracy and efficiency of our approach.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127132991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 82
Photometric Stereo with Nearby Planar Distributed Illuminants 具有邻近平面分布光源的光度立体
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.55
James J. Clark
This paper considers the problem of shape-from-shading using nearby planar distributed illuminants. It is shown that a rectangular planar nearby distributed uniform isotropic illuminant shining on a small Lambertian surface patch is equivalent to a single isotropic point light source at infinity. A closed-form solution is given for the equivalent point light source direction in terms of the illuminant corner locations. Equivalent point light sources can be obtained for multiple rectangular illuminants allowing standard photometric stereo algorithms to be used. An extension is given to the case of a rectangular planar illuminant with arbitrary radiance distribution. It is shown that a Walsh function approximation to the arbitrary illuminant distribution leads to an efficient computation of the equivalent point light source directions. A search technique employing a solution consistency measure is presented to handle the case of unknown depths. Applications of the theory presented in this paper include visual user-interfaces using shape-from-shading algorithms making use of the illumination from computer monitors, or movie screens.
本文研究了利用邻近平面分布光源进行形状渐变的问题。结果表明,一个矩形平面附近分布的均匀各向同性光源照射在一个小的朗伯曲面斑块上,在无穷远处相当于一个单一的各向同性点光源。给出了等效点光源方向在光源角位置上的封闭解。等效点光源可以获得多个矩形光源,允许使用标准的光度立体算法。对具有任意亮度分布的矩形平面光源进行了推广。结果表明,用Walsh函数逼近任意光源分布可以有效地计算等效点光源方向。针对深度未知的情况,提出了一种采用解一致性度量的搜索技术。本文中提出的理论的应用包括使用形状-从阴影算法的视觉用户界面,利用来自计算机显示器或电影屏幕的照明。
{"title":"Photometric Stereo with Nearby Planar Distributed Illuminants","authors":"James J. Clark","doi":"10.1109/CRV.2006.55","DOIUrl":"https://doi.org/10.1109/CRV.2006.55","url":null,"abstract":"This paper considers the problem of shape-from-shading using nearby planar distributed illuminants. It is shown that a rectangular planar nearby distributed uniform isotropic illuminant shining on a small Lambertian surface patch is equivalent to a single isotropic point light source at infinity. A closed-form solution is given for the equivalent point light source direction in terms of the illuminant corner locations. Equivalent point light sources can be obtained for multiple rectangular illuminants allowing standard photometric stereo algorithms to be used. An extension is given to the case of a rectangular planar illuminant with arbitrary radiance distribution. It is shown that a Walsh function approximation to the arbitrary illuminant distribution leads to an efficient computation of the equivalent point light source directions. A search technique employing a solution consistency measure is presented to handle the case of unknown depths. Applications of the theory presented in this paper include visual user-interfaces using shape-from-shading algorithms making use of the illumination from computer monitors, or movie screens.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"104 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114373636","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Detection and tracking of pianist hands and fingers 钢琴家的手和手指的检测和跟踪
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.26
D. Gorodnichy, A. Yogeswaran
Current MIDI recording and transmitting technology allows teachers to teach piano playing remotely (or off-line): a teacher plays a MIDI-keyboard at one place and a student observes the played piano keys on another MIDI-keyboard at another place. What this technology does not allow is to see how the piano keys are played, namely: which hand and finger was used to play a key. In this paper we present a video recognition tool that makes it possible to provide this information. A video-camera is mounted on top of the piano keyboard and video recognition techniques are then used to calibrate piano image with MIDI sound, then to detect and track pianist hands and then to annotate the fingers that play the piano. The result of the obtained video annotation of piano playing can then be shown on a computer screen for further perusal by a piano teacher or a student.
目前的MIDI录制和传输技术允许教师远程(或离线)教授钢琴演奏:教师在一个地方弹奏MIDI键盘,学生在另一个地方观察另一个MIDI键盘上弹奏的钢琴键。这项技术不允许看到钢琴琴键是如何弹奏的,即:哪只手和哪根手指是用来弹奏琴键的。在本文中,我们提出了一个视频识别工具,使提供这些信息成为可能。摄像机安装在钢琴键盘上,然后使用视频识别技术来校准带有MIDI声音的钢琴图像,然后检测和跟踪钢琴家的手,然后对弹钢琴的手指进行注释。然后,获得的钢琴演奏视频注释的结果可以显示在计算机屏幕上,供钢琴老师或学生进一步阅读。
{"title":"Detection and tracking of pianist hands and fingers","authors":"D. Gorodnichy, A. Yogeswaran","doi":"10.1109/CRV.2006.26","DOIUrl":"https://doi.org/10.1109/CRV.2006.26","url":null,"abstract":"Current MIDI recording and transmitting technology allows teachers to teach piano playing remotely (or off-line): a teacher plays a MIDI-keyboard at one place and a student observes the played piano keys on another MIDI-keyboard at another place. What this technology does not allow is to see how the piano keys are played, namely: which hand and finger was used to play a key. In this paper we present a video recognition tool that makes it possible to provide this information. A video-camera is mounted on top of the piano keyboard and video recognition techniques are then used to calibrate piano image with MIDI sound, then to detect and track pianist hands and then to annotate the fingers that play the piano. The result of the obtained video annotation of piano playing can then be shown on a computer screen for further perusal by a piano teacher or a student.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"160 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124480899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 50
Evaluation of Colour Image Segmentation Hierarchies 彩色图像分割层次的评价
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.31
D. MacDonald, J. Lang, M. McAllister
Image segmentation is a key low-level vision process for many tasks in image understanding and applications in image processing. The groups of related pixels produced by a segmentation are expected to reveal some information about the objects in a scene. Many equally valid segmentations exist for most images depending on the context of the segmentation, in particular, the levelof- detail. Segmentation hierarchies can represent different level-of-detail simultaneously. In this paper, we employ the Earth Mover’s Distance (EMD) as a segmentation criterion in a hierarchical colour segmentation algorithm. The use of EMD is motivated by its success in content-based image retrieval based on colour signatures. We also develop a novel evaluation method which looks at the stability of segmentation hierarchies in stereo images. This evaluation method allows us to compare the EMD segmentation criterion with variants and with criterions based on mean colour.
图像分割是图像理解和图像处理中重要的底层视觉过程。分割产生的相关像素组有望揭示场景中物体的一些信息。对于大多数图像,存在许多同样有效的分割,这取决于分割的上下文,特别是细节水平。分割层次结构可以同时表示不同的细节层次。在本文中,我们采用了大地移动距离(EMD)作为分层颜色分割算法的分割标准。EMD在基于颜色签名的基于内容的图像检索中取得了成功,这促使了它的使用。我们还开发了一种新的评估方法,该方法着眼于立体图像中分割层次的稳定性。这种评估方法允许我们将EMD分割标准与变体和基于平均颜色的标准进行比较。
{"title":"Evaluation of Colour Image Segmentation Hierarchies","authors":"D. MacDonald, J. Lang, M. McAllister","doi":"10.1109/CRV.2006.31","DOIUrl":"https://doi.org/10.1109/CRV.2006.31","url":null,"abstract":"Image segmentation is a key low-level vision process for many tasks in image understanding and applications in image processing. The groups of related pixels produced by a segmentation are expected to reveal some information about the objects in a scene. Many equally valid segmentations exist for most images depending on the context of the segmentation, in particular, the levelof- detail. Segmentation hierarchies can represent different level-of-detail simultaneously. In this paper, we employ the Earth Mover’s Distance (EMD) as a segmentation criterion in a hierarchical colour segmentation algorithm. The use of EMD is motivated by its success in content-based image retrieval based on colour signatures. We also develop a novel evaluation method which looks at the stability of segmentation hierarchies in stereo images. This evaluation method allows us to compare the EMD segmentation criterion with variants and with criterions based on mean colour.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134035375","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Shape-Based Object Segmentation with Simultaneous Intensity Adjustment 同时调整强度的基于形状的目标分割
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.64
Sarawut Tae-O-Sot, S. Jitapunkul, S. Auethavekiat
Most segmentation algorithms are based on the assumption of intensity homogeneity within an object. However, in many applications, the object of interest contains more than one homogenous region. Even when the object’s shape is known, such object is not effectively extracted. In this paper, we propose a segmentation process for the objects containing 2 homogenous regions. Our method is based on the level set method. We construct the shape model from the set of manually extracted objects. The parameters that represent the shape model are coefficients of PCA basis. Instead of defining a new cost-function based on heterogeneity assumption, we repeatedly form a homogenous region inside the evolving curve and evolve the curve by the level set method. Our experiment on medical images indicated that our method effectively segmented object with one and two homogenous regions.
大多数分割算法都是基于物体内部强度均匀性的假设。然而,在许多应用程序中,感兴趣的对象包含多个同质区域。即使知道了物体的形状,也不能有效地提取出来。本文提出了一种包含2个同质区域的目标分割方法。我们的方法是基于水平集方法。我们从人工提取的对象集合中构建形状模型。表示形状模型的参数是主成分分析基的系数。我们没有基于异质性假设定义新的成本函数,而是在演化曲线内部反复形成一个齐次区域,并通过水平集方法对曲线进行演化。在医学图像上的实验表明,该方法可以有效地分割出一个均匀区域和两个均匀区域。
{"title":"Shape-Based Object Segmentation with Simultaneous Intensity Adjustment","authors":"Sarawut Tae-O-Sot, S. Jitapunkul, S. Auethavekiat","doi":"10.1109/CRV.2006.64","DOIUrl":"https://doi.org/10.1109/CRV.2006.64","url":null,"abstract":"Most segmentation algorithms are based on the assumption of intensity homogeneity within an object. However, in many applications, the object of interest contains more than one homogenous region. Even when the object’s shape is known, such object is not effectively extracted. In this paper, we propose a segmentation process for the objects containing 2 homogenous regions. Our method is based on the level set method. We construct the shape model from the set of manually extracted objects. The parameters that represent the shape model are coefficients of PCA basis. Instead of defining a new cost-function based on heterogeneity assumption, we repeatedly form a homogenous region inside the evolving curve and evolve the curve by the level set method. Our experiment on medical images indicated that our method effectively segmented object with one and two homogenous regions.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130695464","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Confidence Based updation of Motion Conspicuity in Dynamic Scenes 动态场景中基于置信度的运动显著性更新
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.24
V. Singh, Subhransu Maji, A. Mukerjee
Computational models of visual attention result in considerable data compression by eliminating processing on regions likely to be devoid of meaningful content. While saliency maps in static images is indexed on image region (pixels), psychovisual data indicates that in dynamic scenes human attention is object driven and localized motion is a significant determiner of object conspicuity. We have introduced a confidence map, which indicates the uncertainty in the position of the moving objects incorporating the exponential loss of information as we move away from the fovea. We improve the model further using a computational model of visual attention based on perceptual grouping of objects with motion and computation of a motion saliency map based on localized motion conspicuity of the objects. Behaviors exhibited in the system include attentive focus on moving wholes, shifting focus in multiple object motion, focus on objects moving contrary to the majority motion. We also present experimental data contrasting the model with human gaze tracking in a simple visual task.
视觉注意的计算模型通过消除对可能缺乏有意义内容的区域的处理,导致相当大的数据压缩。虽然静态图像中的显著性映射是在图像区域(像素)上索引的,但心理视觉数据表明,在动态场景中,人类的注意力是物体驱动的,局部运动是物体显著性的重要决定因素。我们引入了一个置信度图,它显示了移动物体位置的不确定性,包括当我们离开中央凹时信息的指数损失。我们使用基于运动对象感知分组的视觉注意计算模型和基于物体局部运动显著性计算的运动显著性图进一步改进了模型。该系统表现出的行为包括对运动整体的关注、对多物体运动的关注、对与多数运动相反的物体的关注。我们还提供了在一个简单的视觉任务中将该模型与人类注视跟踪进行对比的实验数据。
{"title":"Confidence Based updation of Motion Conspicuity in Dynamic Scenes","authors":"V. Singh, Subhransu Maji, A. Mukerjee","doi":"10.1109/CRV.2006.24","DOIUrl":"https://doi.org/10.1109/CRV.2006.24","url":null,"abstract":"Computational models of visual attention result in considerable data compression by eliminating processing on regions likely to be devoid of meaningful content. While saliency maps in static images is indexed on image region (pixels), psychovisual data indicates that in dynamic scenes human attention is object driven and localized motion is a significant determiner of object conspicuity. We have introduced a confidence map, which indicates the uncertainty in the position of the moving objects incorporating the exponential loss of information as we move away from the fovea. We improve the model further using a computational model of visual attention based on perceptual grouping of objects with motion and computation of a motion saliency map based on localized motion conspicuity of the objects. Behaviors exhibited in the system include attentive focus on moving wholes, shifting focus in multiple object motion, focus on objects moving contrary to the majority motion. We also present experimental data contrasting the model with human gaze tracking in a simple visual task.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132907242","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Toward a Realistic Interpretation of Blue-spill for Blue-screen Matting 对蓝屏抠图中蓝溢现象的现实解释
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.77
Jonathan Dupont, F. Deschênes
This paper proposes a realistic interpretation of the bluespill problem that affects blue-screen matting in digital image composition. This phenomenon consists in the contamination of the foreground object color by the background color. Based on the notion of mutual illumination introduced by Forsyth and Zisserman [5], we interpret the bluespill phenomenon and propose an extended composition equation explicitly taking this into consideration. Experiments confirm the accuracy of the proposed method. Comparisons are made with the classical composition equation.
本文对数字图像合成中影响蓝屏抠图的蓝溢问题提出了一种现实的解释。这种现象是由于前景物体的颜色受到背景颜色的污染。基于Forsyth和Zisserman[5]引入的相互照明的概念,我们解释了蓝溢现象,并明确地提出了一个考虑到这一点的扩展成分方程。实验验证了该方法的准确性。并与经典成分方程进行了比较。
{"title":"Toward a Realistic Interpretation of Blue-spill for Blue-screen Matting","authors":"Jonathan Dupont, F. Deschênes","doi":"10.1109/CRV.2006.77","DOIUrl":"https://doi.org/10.1109/CRV.2006.77","url":null,"abstract":"This paper proposes a realistic interpretation of the bluespill problem that affects blue-screen matting in digital image composition. This phenomenon consists in the contamination of the foreground object color by the background color. Based on the notion of mutual illumination introduced by Forsyth and Zisserman [5], we interpret the bluespill phenomenon and propose an extended composition equation explicitly taking this into consideration. Experiments confirm the accuracy of the proposed method. Comparisons are made with the classical composition equation.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133207885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Image Classification and Retrieval using Correlation 基于相关性的图像分类与检索
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.40
Imran Ahmad, M. T. Ibrahim
Image retrieval methods aim to retrieve relevant images from an image database that are similar to the query image. The ability to effectively retrieve non-alphanumeric data is a complex issue. The problem becomes even more difficult due to the high dimension of the variable space associated with the images. Image classification is a very active and promising research domain in the area of image management and retrieval. In this paper, we propose a new image classification and retrieval scheme that automatically selects the discriminating features. Our method consists of two phases: (i) classification of images on the basis of maximum cross correlation and (ii) retrieval of images from the database against a given query image. The proposed retrieval algorithm recursively searches similar images on the basis of their correlation against a given query image from a set of registered images in the database. The algorithm is very efficient, provided that the mean images of all of the classes are computed and available in advance. The proposed method classifies the images on the basis of maximum correlation so that the images with more similarities and, hence, exhibiting maximum correlation with each other are grouped in the same class and, are retrieved accordingly.
图像检索方法的目的是从图像数据库中检索与查询图像相似的相关图像。有效检索非字母数字数据的能力是一个复杂的问题。由于与图像相关的可变空间的高维,这个问题变得更加困难。在图像管理和检索领域,图像分类是一个非常活跃和有前途的研究领域。本文提出了一种自动选择识别特征的图像分类检索方法。我们的方法包括两个阶段:(i)基于最大相互关联的图像分类和(ii)根据给定的查询图像从数据库中检索图像。本文提出的检索算法基于与数据库中一组注册图像之间的相关性,递归地搜索相似图像。该算法是非常有效的,前提是所有类别的平均图像都是预先计算出来的。该方法基于最大相关性对图像进行分类,将相似度较高且相互间相关性最大的图像归为一类,并进行检索。
{"title":"Image Classification and Retrieval using Correlation","authors":"Imran Ahmad, M. T. Ibrahim","doi":"10.1109/CRV.2006.40","DOIUrl":"https://doi.org/10.1109/CRV.2006.40","url":null,"abstract":"Image retrieval methods aim to retrieve relevant images from an image database that are similar to the query image. The ability to effectively retrieve non-alphanumeric data is a complex issue. The problem becomes even more difficult due to the high dimension of the variable space associated with the images. Image classification is a very active and promising research domain in the area of image management and retrieval. In this paper, we propose a new image classification and retrieval scheme that automatically selects the discriminating features. Our method consists of two phases: (i) classification of images on the basis of maximum cross correlation and (ii) retrieval of images from the database against a given query image. The proposed retrieval algorithm recursively searches similar images on the basis of their correlation against a given query image from a set of registered images in the database. The algorithm is very efficient, provided that the mean images of all of the classes are computed and available in advance. The proposed method classifies the images on the basis of maximum correlation so that the images with more similarities and, hence, exhibiting maximum correlation with each other are grouped in the same class and, are retrieved accordingly.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114361122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1