首页 > 最新文献

The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)最新文献

英文 中文
Handling Occlusions in Real-time Augmented Reality : Dealing with Movable Real and Virtual Objects 在实时增强现实中处理遮挡:处理可移动的真实和虚拟物体
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.38
P. Fortin, P. Hébert
Realistic rendering in real-time augmented reality applications leads one to consider physical interactions between real and virtual worlds. One of these interactions is mutual occlusions in the rendered viewpoint. This paper presents two approaches for handling occlusions when the real objects can be displaced or deformed. The first approach is model-based. It is suited for a static viewpoint and relies only on a tracked bounding volume model within which the object’s silhouette is carved. The second approach is depthbased and makes it possible to change the viewpoint by exploiting a handheld stereo camera. Both approaches are devised to minimize the effect of real object tracking errors in the rendered viewpoint.
实时增强现实应用中的逼真渲染导致人们考虑真实世界和虚拟世界之间的物理交互。其中一种交互作用是渲染视点中的相互遮挡。本文提出了两种处理真实物体发生位移或变形时遮挡的方法。第一种方法是基于模型的。它适合于静态视点,并且只依赖于物体轮廓雕刻的跟踪边界体模型。第二种方法是基于深度的,可以利用手持立体摄像机来改变视点。这两种方法都是为了最小化渲染视点中真实目标跟踪误差的影响。
{"title":"Handling Occlusions in Real-time Augmented Reality : Dealing with Movable Real and Virtual Objects","authors":"P. Fortin, P. Hébert","doi":"10.1109/CRV.2006.38","DOIUrl":"https://doi.org/10.1109/CRV.2006.38","url":null,"abstract":"Realistic rendering in real-time augmented reality applications leads one to consider physical interactions between real and virtual worlds. One of these interactions is mutual occlusions in the rendered viewpoint. This paper presents two approaches for handling occlusions when the real objects can be displaced or deformed. The first approach is model-based. It is suited for a static viewpoint and relies only on a tracked bounding volume model within which the object’s silhouette is carved. The second approach is depthbased and makes it possible to change the viewpoint by exploiting a handheld stereo camera. Both approaches are devised to minimize the effect of real object tracking errors in the rendered viewpoint.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122390923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Hierarchical Region Mean-Based Image Segmentation 基于层次区域均值的图像分割
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.39
S. Wesolkowski, P. Fieguth
Gibbs Random Fields (GRFs), which produce elegant models, but which have very poor computational speed have been widely applied to image segmentation. In contrast to block-based hierarchies usually constructed for GRFs, the irregular region-based approach is a more natural model in segmenting real images. In this paper, we show that the fineto- coarse region-based hierarchical regions framework for the well-known Potts model can be extended to non-edge based interactions. By deliberately oversegmenting at the finer scale, the method proceeds conservatively by avoiding the construction of regions which straddle a region boundary by computing region mean differences. This demonstrates the hierarchical method is able to model region interactions through new generalizations at higher levels in the hierarchy which represent regions. Promising results are presented.
吉布斯随机场(Gibbs Random Fields, GRFs)产生的模型简洁,但其计算速度较差,已被广泛应用于图像分割中。相对于通常为grf构建的基于块的层次结构,基于不规则区域的方法是一种更自然的真实图像分割模型。在本文中,我们证明了著名的Potts模型的基于精细到粗糙区域的分层区域框架可以扩展到非边缘的相互作用。通过在更精细的尺度上故意进行过分割,该方法通过计算区域均值差来避免构建跨越区域边界的区域,从而保守地进行。这证明了分层方法能够通过在表示区域的层次结构的更高级别上的新泛化来建模区域交互。提出了有希望的结果。
{"title":"Hierarchical Region Mean-Based Image Segmentation","authors":"S. Wesolkowski, P. Fieguth","doi":"10.1109/CRV.2006.39","DOIUrl":"https://doi.org/10.1109/CRV.2006.39","url":null,"abstract":"Gibbs Random Fields (GRFs), which produce elegant models, but which have very poor computational speed have been widely applied to image segmentation. In contrast to block-based hierarchies usually constructed for GRFs, the irregular region-based approach is a more natural model in segmenting real images. In this paper, we show that the fineto- coarse region-based hierarchical regions framework for the well-known Potts model can be extended to non-edge based interactions. By deliberately oversegmenting at the finer scale, the method proceeds conservatively by avoiding the construction of regions which straddle a region boundary by computing region mean differences. This demonstrates the hierarchical method is able to model region interactions through new generalizations at higher levels in the hierarchy which represent regions. Promising results are presented.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134345539","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Simultaneous Tracking and Action Recognition using the PCA-HOG Descriptor 使用PCA-HOG描述符的同步跟踪和动作识别
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.66
Wei-Lwun Lu, J. Little
This paper presents a template-based algorithm to track and recognize athlete’s actions in an integrated system using only visual information. Conventional template-based action recognition systems usually consider action recognition and tracking as two independent problems, and solve them separately. In contrast, our algorithm emphasizes that tracking and action recognition can be tightly coupled into a single framework, where tracking assists action recognition and vise versa. Moreover, this paper proposes to represent the athletes by the PCA-HOG descriptor, which can be computed by first transforming the athletes to the grids of Histograms of Oriented Gradient (HOG) descriptor and then project it to a linear subspace by Principal Component Analysis (PCA). The exploitation of the PCA-HOG descriptor not only helps the tracker to be robust under illumination, pose, and view-point changes, but also implicitly centers the figure in the tracking region, which makes action recognition possible. Empirical results in hockey and soccer sequences show the effectiveness of this algorithm.
本文提出了一种基于模板的算法,仅利用视觉信息在集成系统中跟踪和识别运动员的动作。传统的基于模板的动作识别系统通常将动作识别和动作跟踪视为两个独立的问题,分别进行解决。相比之下,我们的算法强调跟踪和动作识别可以紧密耦合到一个框架中,其中跟踪辅助动作识别,反之亦然。此外,本文提出了用PCA-HOG描述符来表示运动员,该描述符可以通过将运动员转换为定向梯度直方图(HOG)描述符的网格,然后通过主成分分析(PCA)将其投影到线性子空间来计算。利用PCA-HOG描述符不仅可以帮助跟踪器在光照、姿态和视点变化下保持鲁棒性,而且可以隐式地将图像集中在跟踪区域,从而使动作识别成为可能。曲棍球和足球序列的实验结果表明了该算法的有效性。
{"title":"Simultaneous Tracking and Action Recognition using the PCA-HOG Descriptor","authors":"Wei-Lwun Lu, J. Little","doi":"10.1109/CRV.2006.66","DOIUrl":"https://doi.org/10.1109/CRV.2006.66","url":null,"abstract":"This paper presents a template-based algorithm to track and recognize athlete’s actions in an integrated system using only visual information. Conventional template-based action recognition systems usually consider action recognition and tracking as two independent problems, and solve them separately. In contrast, our algorithm emphasizes that tracking and action recognition can be tightly coupled into a single framework, where tracking assists action recognition and vise versa. Moreover, this paper proposes to represent the athletes by the PCA-HOG descriptor, which can be computed by first transforming the athletes to the grids of Histograms of Oriented Gradient (HOG) descriptor and then project it to a linear subspace by Principal Component Analysis (PCA). The exploitation of the PCA-HOG descriptor not only helps the tracker to be robust under illumination, pose, and view-point changes, but also implicitly centers the figure in the tracking region, which makes action recognition possible. Empirical results in hockey and soccer sequences show the effectiveness of this algorithm.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129356215","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 175
Urban Position Estimation from One Dimensional Visual Cues 基于一维视觉线索的城市位置估计
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.81
Derek Johns, G. Dudek
We consider the problem of vision-based position estimation in urban environments. In particular, we are interested in position estimation from visual cues, but using only limited computational resources. Our particular solution to this problem is based on representing the variability of the "horizon" of the cityscape when seen from within the city; that is, the outlines of the rooftops of adjacent buildings. By encoding the image using only such a one-dimensional contour, we obtain an image encoding that is exceedingly compact. This, in turn, allows us to both efficiently transmit this representation to a remote "recognition engine" as well as allowing for an efficient storage and matching process. We outline our approach and representation, and provide experimental data supporting its feasibility.
我们研究了城市环境中基于视觉的位置估计问题。特别是,我们感兴趣的是从视觉线索的位置估计,但只使用有限的计算资源。我们对这个问题的特殊解决方案是基于从城市内部看到的城市景观的“地平线”的可变性;也就是相邻建筑屋顶的轮廓。通过仅使用这样的一维轮廓编码图像,我们获得了非常紧凑的图像编码。反过来,这允许我们既有效地将这种表示传输到远程“识别引擎”,也允许有效的存储和匹配过程。我们概述了我们的方法和表示,并提供了实验数据支持其可行性。
{"title":"Urban Position Estimation from One Dimensional Visual Cues","authors":"Derek Johns, G. Dudek","doi":"10.1109/CRV.2006.81","DOIUrl":"https://doi.org/10.1109/CRV.2006.81","url":null,"abstract":"We consider the problem of vision-based position estimation in urban environments. In particular, we are interested in position estimation from visual cues, but using only limited computational resources. Our particular solution to this problem is based on representing the variability of the \"horizon\" of the cityscape when seen from within the city; that is, the outlines of the rooftops of adjacent buildings. By encoding the image using only such a one-dimensional contour, we obtain an image encoding that is exceedingly compact. This, in turn, allows us to both efficiently transmit this representation to a remote \"recognition engine\" as well as allowing for an efficient storage and matching process. We outline our approach and representation, and provide experimental data supporting its feasibility.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130872701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
A feature-based tracking algorithm for vehicles in intersections 基于特征的交叉口车辆跟踪算法
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.3
N. Saunier, T. Sayed
Intelligent Transportation Systems need methods to automatically monitor the road traffic, and especially track vehicles. Most research has concentrated on highways. Traffic in intersections is more variable, with multiple entrance and exit regions. This paper describes an extension to intersections of the feature-tracking algorithm described in [1]. Vehicle features are rarely tracked from their entrance in the field of view to their exit. Our algorithm can accommodate the problem caused by the disruption of feature tracks. It is evaluated on video sequences recorded on four different intersections.
智能交通系统需要自动监控道路交通,特别是跟踪车辆的方法。大多数研究都集中在高速公路上。十字路口的交通变化较大,有多个出入口区域。本文描述了[1]中描述的特征跟踪算法对交点的扩展。车辆特征很少被跟踪,从他们的入口在视野中,他们的出口。我们的算法可以适应特征轨迹中断引起的问题。对在四个不同路口记录的视频序列进行了评估。
{"title":"A feature-based tracking algorithm for vehicles in intersections","authors":"N. Saunier, T. Sayed","doi":"10.1109/CRV.2006.3","DOIUrl":"https://doi.org/10.1109/CRV.2006.3","url":null,"abstract":"Intelligent Transportation Systems need methods to automatically monitor the road traffic, and especially track vehicles. Most research has concentrated on highways. Traffic in intersections is more variable, with multiple entrance and exit regions. This paper describes an extension to intersections of the feature-tracking algorithm described in [1]. Vehicle features are rarely tracked from their entrance in the field of view to their exit. Our algorithm can accommodate the problem caused by the disruption of feature tracks. It is evaluated on video sequences recorded on four different intersections.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130221703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 250
Expression-Invariant Face Recognition with Expression Classification 基于表情分类的表情不变人脸识别
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.34
Xiaoxing Li, Greg Mori, Hao Zhang
Face recognition is one of the most intensively studied topics in computer vision and pattern recognition. Facial expression, which changes face geometry, usually has an adverse effect on the performance of a face recognition system. On the other hand, face geometry is a useful cue for recognition. Taking these into account, we utilize the idea of separating geometry and texture information in a face image and model the two types of information by projecting them into separate PCA spaces which are specially designed to capture the distinctive features among different individuals. Subsequently, the texture and geometry attributes are re-combined to form a classifier which is capable of recognizing faces with different expressions. Finally, by studying face geometry, we are able to determine which type of facial expression has been carried out, thus build an expression classifier. Numerical validations of the proposed method are given.
人脸识别是计算机视觉和模式识别领域研究最深入的课题之一。面部表情改变了人脸的几何形状,通常会对人脸识别系统的性能产生不利影响。另一方面,面部几何形状是一个有用的识别线索。考虑到这些,我们利用分离人脸图像中的几何和纹理信息的思想,并通过将两种类型的信息投影到单独的PCA空间中来建模,这些PCA空间专门用于捕获不同个体之间的独特特征。然后,将纹理和几何属性重新组合,形成能够识别不同表情人脸的分类器。最后,通过对面部几何的研究,我们可以确定哪种类型的面部表情被执行,从而建立一个表情分类器。给出了该方法的数值验证。
{"title":"Expression-Invariant Face Recognition with Expression Classification","authors":"Xiaoxing Li, Greg Mori, Hao Zhang","doi":"10.1109/CRV.2006.34","DOIUrl":"https://doi.org/10.1109/CRV.2006.34","url":null,"abstract":"Face recognition is one of the most intensively studied topics in computer vision and pattern recognition. Facial expression, which changes face geometry, usually has an adverse effect on the performance of a face recognition system. On the other hand, face geometry is a useful cue for recognition. Taking these into account, we utilize the idea of separating geometry and texture information in a face image and model the two types of information by projecting them into separate PCA spaces which are specially designed to capture the distinctive features among different individuals. Subsequently, the texture and geometry attributes are re-combined to form a classifier which is capable of recognizing faces with different expressions. Finally, by studying face geometry, we are able to determine which type of facial expression has been carried out, thus build an expression classifier. Numerical validations of the proposed method are given.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115928207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 80
Stability Improvement of Vision Algorithms 视觉算法稳定性的改进
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.69
K. Shahid, G. Okouneva, D. McTavish, J. Karpynczyk
This paper presents and demonstrates an automated generic approach to improving the accuracy and stability of iterative pose estimation in computer vision applications. The class of problem involves the use of calibrated CCD camera video imagery to compute the pose of a slowly moving object based on an arrangement of visual targets on the surface of the object. The basis of stereo-vision algorithms is to minimize a re-projection error cost function. The proposed method estimates the optimal target locations within the area of interest. The optimal target configuration delivers the minimal condition number of the linear system associated with the iterative algorithm. The method is demonstrated for the case when targets are located within a 3D domain. Two pose estimation algorithms are compared: single camera and two-camera algorithms. A better accuracy in pose estimation can be achieved with a single camera algorithm with optimized target locations. Also, this method can be applied to perform optimization of target locations attached to a 2D surface.
本文提出并演示了一种提高计算机视觉应用中迭代姿态估计精度和稳定性的自动化通用方法。这类问题涉及使用经过校准的CCD摄像机视频图像,根据物体表面的视觉目标排列来计算缓慢移动物体的姿态。立体视觉算法的基础是最小化重投影误差代价函数。该方法在感兴趣的区域内估计最优目标位置。最优目标配置提供了与迭代算法相关的线性系统的最小条件数。针对目标位于三维区域内的情况,对该方法进行了验证。比较了单摄像头和双摄像头两种姿态估计算法。利用优化目标位置的单摄像机算法可以获得更好的姿态估计精度。此外,该方法还可用于二维曲面上附着目标位置的优化。
{"title":"Stability Improvement of Vision Algorithms","authors":"K. Shahid, G. Okouneva, D. McTavish, J. Karpynczyk","doi":"10.1109/CRV.2006.69","DOIUrl":"https://doi.org/10.1109/CRV.2006.69","url":null,"abstract":"This paper presents and demonstrates an automated generic approach to improving the accuracy and stability of iterative pose estimation in computer vision applications. The class of problem involves the use of calibrated CCD camera video imagery to compute the pose of a slowly moving object based on an arrangement of visual targets on the surface of the object. The basis of stereo-vision algorithms is to minimize a re-projection error cost function. The proposed method estimates the optimal target locations within the area of interest. The optimal target configuration delivers the minimal condition number of the linear system associated with the iterative algorithm. The method is demonstrated for the case when targets are located within a 3D domain. Two pose estimation algorithms are compared: single camera and two-camera algorithms. A better accuracy in pose estimation can be achieved with a single camera algorithm with optimized target locations. Also, this method can be applied to perform optimization of target locations attached to a 2D surface.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116630435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Epipolar Geometry for the Rectification of Cubic Panoramas 三次全景校正的极极几何
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.29
Florian Kangni, R. Laganière
Image panoramas are of importance for virtual navigation in remote or synthetic environments. To process these panoramas, different representations have been proposed; this paper presents a study of cubic panoramas. Standard projective geometry concepts are adapted to cubic panoramas to derive the notions of fundamental matrix, essential matrix and the equivalent of stereo rectification. Methods and results are presented which could be very helpful in obtaining solutions to disparity estimation, pose estimation and view interpolation problems in the context of cubic panoramas.
图像全景对于远程或合成环境下的虚拟导航具有重要意义。为了处理这些全景图,提出了不同的表示方法;本文对立方全景图进行了研究。将标准的射影几何概念应用于三次全景图,导出了基本矩阵、本质矩阵和立体校正等效的概念。本文提出的方法和结果对三次全景图视差估计、位姿估计和视点插值问题的求解有很大帮助。
{"title":"Epipolar Geometry for the Rectification of Cubic Panoramas","authors":"Florian Kangni, R. Laganière","doi":"10.1109/CRV.2006.29","DOIUrl":"https://doi.org/10.1109/CRV.2006.29","url":null,"abstract":"Image panoramas are of importance for virtual navigation in remote or synthetic environments. To process these panoramas, different representations have been proposed; this paper presents a study of cubic panoramas. Standard projective geometry concepts are adapted to cubic panoramas to derive the notions of fundamental matrix, essential matrix and the equivalent of stereo rectification. Methods and results are presented which could be very helpful in obtaining solutions to disparity estimation, pose estimation and view interpolation problems in the context of cubic panoramas.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128596801","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Image Thresholding Using Ant Colony Optimization 基于蚁群优化的图像阈值分割
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.42
Alice R. Malisia, H. Tizhoosh
This study is an investigation of the application of ant colony optimization to image thresholding. This paper presents an approach where one ant is assigned to each pixel of an image and then moves around the image seeking low grayscale regions. Experimental results demonstrate that the proposed ant-based method performs better than other two established thresholding algorithms. Further work must be conducted to optimize the algorithm parameters, improve the analysis of the pheromone data and reduce computation time. However, the study indicates that an ant-based approach has the potential of becoming an established image thresholding technique.
本文主要研究蚁群算法在图像阈值分割中的应用。本文提出了一种方法,在图像的每个像素上分配一只蚂蚁,然后在图像周围移动,寻找低灰度区域。实验结果表明,本文提出的基于蚁群的阈值算法比其他两种已有的阈值算法性能更好。进一步优化算法参数,提高信息素数据的分析能力,减少计算时间。然而,该研究表明,基于蚂蚁的方法有可能成为一种成熟的图像阈值分割技术。
{"title":"Image Thresholding Using Ant Colony Optimization","authors":"Alice R. Malisia, H. Tizhoosh","doi":"10.1109/CRV.2006.42","DOIUrl":"https://doi.org/10.1109/CRV.2006.42","url":null,"abstract":"This study is an investigation of the application of ant colony optimization to image thresholding. This paper presents an approach where one ant is assigned to each pixel of an image and then moves around the image seeking low grayscale regions. Experimental results demonstrate that the proposed ant-based method performs better than other two established thresholding algorithms. Further work must be conducted to optimize the algorithm parameters, improve the analysis of the pheromone data and reduce computation time. However, the study indicates that an ant-based approach has the potential of becoming an established image thresholding technique.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134216572","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 50
Interpreting Camera Operations in the Context of Content-based Video Indexing and Retrieval 在基于内容的视频索引和检索的背景下解读摄像机操作
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.44
Wei Pan, F. Deschênes
In this work, we intend to go one step further to overcome the difficulty that lies in the gap between low-level media features (e.g. colors, texture, motion, etc.) and high-level concepts to perform a reliable content-based indexing and retrieval. More especially, our work proposes a new way to establish a connection between both geometric and radiometric deformations and the characterization of them in terms of camera operations. Based on both the apparent motion and the defocus blur (low-level features), we estimate extrinsic and intrinsic camera parameter changes, and then deduce 3D camera operations (i.e. mid-level features), such as panning/tracking, tilting/booming, zooming/ dollying and rolling, as well as focus changes. Finally, camera operations are recorded into an index which is then used for video retrieval. Experiments confirm that the proposed mid-level features can be accurately deduced from low-level features and that they can be used for indexing and retrieval purpose.
在这项工作中,我们打算进一步克服底层媒体特征(如颜色、纹理、运动等)与高层概念之间的差距,以执行可靠的基于内容的索引和检索。更特别的是,我们的工作提出了一种新的方法来建立几何和辐射变形之间的联系,并在相机操作方面对它们进行表征。基于表观运动和离焦模糊(低级特征),我们估计了相机的外在和内在参数变化,然后推断出3D相机操作(即中级特征),如平移/跟踪,倾斜/蓬勃发展,变焦/平移和滚动,以及焦点变化。最后,摄像机操作被记录到索引中,然后用于视频检索。实验证明,所提出的中级特征可以准确地从低级特征中推导出来,并可用于索引和检索目的。
{"title":"Interpreting Camera Operations in the Context of Content-based Video Indexing and Retrieval","authors":"Wei Pan, F. Deschênes","doi":"10.1109/CRV.2006.44","DOIUrl":"https://doi.org/10.1109/CRV.2006.44","url":null,"abstract":"In this work, we intend to go one step further to overcome the difficulty that lies in the gap between low-level media features (e.g. colors, texture, motion, etc.) and high-level concepts to perform a reliable content-based indexing and retrieval. More especially, our work proposes a new way to establish a connection between both geometric and radiometric deformations and the characterization of them in terms of camera operations. Based on both the apparent motion and the defocus blur (low-level features), we estimate extrinsic and intrinsic camera parameter changes, and then deduce 3D camera operations (i.e. mid-level features), such as panning/tracking, tilting/booming, zooming/ dollying and rolling, as well as focus changes. Finally, camera operations are recorded into an index which is then used for video retrieval. Experiments confirm that the proposed mid-level features can be accurately deduced from low-level features and that they can be used for indexing and retrieval purpose.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121301536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1